|
52129
|
1126
|
36
|
2026-04-20T06:40:48.348406+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667248348_m1.jpg...
|
iTerm2
|
EU (ssh)
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
kibana | {"type":"log", kibana | {"type":"log","@timestamp":"2026-04-18T13:26:35Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:35Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:37Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:38Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:38Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:40Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:41Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:41Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:42Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:49Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"kibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:35Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:35Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:35Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:37Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:38Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:38Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:38Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:40Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:42Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:45Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:48Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:49Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"reporting\",\"esqueue\",\"queue-worker\",\"error\"],\"pid\":7,\"message\":\"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\\n at process._tickCallback (internal/process/next_tick.js:61:11)\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:55Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:58Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:00Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:03Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:10Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work\nWARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion \n[+] Running 2/2\n ✔ Container redis Running 0.0s \n ✔ Container docker_lamp_1 Running 0.0s \nAttaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis\ndocker_lamp_1 | \ndocker_lamp_1 | Illuminate\\Database\\QueryException \ndocker_lamp_1 | \ndocker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)\ndocker_lamp_1 | \ndocker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838\ndocker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)\ndocker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class\ndocker_lamp_1 | 836▕ : QueryException::class;\ndocker_lamp_1 | 837▕ \ndocker_lamp_1 | ➜ 838▕ throw new $exceptionType(\ndocker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),\ndocker_lamp_1 | 840▕ $query,\ndocker_lamp_1 | 841▕ $this->prepareBindings($bindings),\ndocker_lamp_1 | 842▕ $e,\ndocker_lamp_1 | \ndocker_lamp_1 | +42 vendor frames \ndocker_lamp_1 | \ndocker_lamp_1 | 43 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \nmariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\nblackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.\nblackfire-1 | usage blackfire-agent [options]\nblackfire-1 | --collector=\"https://blackfire.io\": Sets the URL of Blackfire's data collector\nblackfire-1 | --config=\"/etc/blackfire/agent\": Sets the path to the configuration file\nblackfire-1 | -d: Prints the current configuration\nblackfire-1 | --http-proxy=\"\": Sets the HTTP proxy to use\nblackfire-1 | --https-proxy=\"\": Sets the HTTPS proxy to use\nblackfire-1 | --log-file=\"stderr\": Sets the path of the log file. Use stderr to log to stderr\nblackfire-1 | --log-level=\"1\": log verbosity level (4: debug, 3: info, 2: warning, 1: error)\nblackfire-1 | --register: Helps you with registering the agent\nblackfire-1 | --server-id=\"\": Sets the server id used to authenticate with Blackfire API\nblackfire-1 | --server-token=\"\": Sets the server token used to authenticate with Blackfire API. It is unsafe to set this from the command line\nblackfire-1 | --socket=\"unix:///var/run/blackfire/agent.sock\": Sets the socket the agent should read traces from. Possible value can be a unix socket or a TCP address. ie: unix:///var/run/blackfire/agent.sock or tcp://127.0.0.1:8307\nblackfire-1 | --test: Tests the configuration\nblackfire-1 | --timeout=\"15s\": Sets the Blackfire connection timeout\nblackfire-1 | -v: Prints the version number\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"no configuration paths supplied\"\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"using configuration at default config path\" path=/home/ngrok/.ngrok2/ngrok.yml\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"open config file\" path=/home/ngrok/.ngrok2/ngrok.yml err=nil\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"starting web service\" obj=web addr=0.0.0.0:4040\njiminny_ext-1 exited with code 0\nblackfire-1 exited with code 1\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"tunnel session started\" obj=tunnels.session\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"client session established\" obj=csess id=7d0b19e3215d\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=\"command_line (http)\" addr=http://lamp:3080 url=http://lukask.ngrok.io\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=command_line addr=http://lamp:3080 url=https://lukask.ngrok.io\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Warn] [Entrypoint]: /sys/fs/cgroup///memory.pressure not writable, functionality unavailable to MariaDB\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\ndatadog-1 | [s6-init] making user provided files available at /var/run/s6/etc...exited 0.\nngrok | t=2026-04-20T06:38:55+0000 lvl=info msg=\"update available\" obj=updater\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade information missing, assuming required\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade (mariadb-upgrade or creating healthcheck users) required, but skipped due to $MARIADB_AUTO_UPGRADE setting\nmariadb-1 | 2026-04-20 6:38:55 0 [Note] Starting MariaDB 11.4.5-MariaDB-ubu2404 source revision 0771110266ff5c04216af4bf1243c65f8c67ccf4 server_uid geAM8SSFn8P6Gz+0WMyyW27EUEM= as process 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Compressed tables use zlib 1.3\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Number of transaction pools: 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using ARMv8 crc32 + pmull instructions\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using liburing\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Completed initialization of buffer pool\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes)\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: End of log at LSN=7614257762\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Opened 3 undo tablespaces\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: 128 rollback segments in 3 undo tablespaces are active.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ...\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: log sequence number 7614257762; transaction id 11481082\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'FEEDBACK' is disabled.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'wsrep-provider' is disabled.\nmariadb-1 | 2026-04-20 6:38:58 0 [Note] InnoDB: Buffer pool(s) load completed at 260420 6:38:58\ndatadog-1 | [s6-init] ensuring user provided files have correct perms...exited 0.\ndatadog-1 | [fix-attrs.d] applying ownership & permissions fixes...\ndatadog-1 | [fix-attrs.d] done.\ndatadog-1 | [cont-init.d] executing container initialization scripts...\ndatadog-1 | [cont-init.d] 01-check-apikey.sh: executing... \ndatadog-1 | \ndatadog-1 | ==================================================================================\ndatadog-1 | You must set an DD_API_KEY environment variable to run the Datadog Agent container\ndatadog-1 | ==================================================================================\ndatadog-1 | \ndatadog-1 | [cont-init.d] 01-check-apikey.sh: exited 1.\ndatadog-1 exited with code 1\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '0.0.0.0'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '::'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: Event Scheduler: Loaded 0 events\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: ready for connections.\nmariadb-1 | Version: '11.4.5-MariaDB-ubu2404' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution\ndocker_lamp_1 | \ndocker_lamp_1 | INFO Configuration cached successfully. \ndocker_lamp_1 | \ndocker_lamp_1 | cp: cannot stat '/home/jiminny/bootstrap/cache/config-new.php': No such file or directory\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + [[ false == \\f\\a\\l\\s\\e ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + [[ -f /home/jiminny/storage/oauth-private.key ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + setup_local_environment\ndocker_lamp_1 | + storage_permissions_workaround\ndocker_lamp_1 | + [[ false != \\t\\r\\u\\e ]]\ndocker_lamp_1 | + ln -sf /home/jiminny/storage /home/jiminny/jiminny_storage\ndocker_lamp_1 | + return\ndocker_lamp_1 | + chmod 660 /home/jiminny/storage/oauth-private.key\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/views\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/cache\ndocker_lamp_1 | + chmod 775 /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/cache\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,413Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"version[7.10.2], pid[7], build[default/docker/747e1cc71def077253878a59143c1f785afa92b9/2021-01-13T04:42:47.157277Z], OS[Linux/6.12.54-linuxkit/aarch64], JVM[AdoptOpenJDK/OpenJDK 64-Bit Server VM/15.0.1/15.0.1+9]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM home [/usr/share/elasticsearch/jdk], using bundled JDK [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Djava.locale.providers=SPI,COMPAT, -Xms1g, -Xmx1g, -XX:+UseG1GC, -XX:G1ReservePercent=25, -XX:InitiatingHeapOccupancyPercent=30, -Djava.io.tmpdir=/tmp/elasticsearch-11712280666606670653, -XX:+HeapDumpOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -Des.cgroups.hierarchy.override=/, -Xms700m, -Xmx700m, -XX:MaxDirectMemorySize=367001600, -Des.path.home=/usr/share/elasticsearch, -Des.path.conf=/usr/share/elasticsearch/config, -Des.distribution.flavor=default, -Des.distribution.type=docker, -Des.bundled_jdk=true]\" }\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chmod g+w '{}' ';'\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,926Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [aggs-matrix-stats]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [analysis-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [constant-keyword]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [flattened]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [frozen-indices]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-geoip]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-user-agent]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [kibana]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-expression]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-mustache]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-painless]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-extras]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-version]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [parent-join]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [percolator]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [rank-eval]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [reindex]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repositories-metering-api]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repository-url]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [search-business-rules]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [searchable-snapshots]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [spatial]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transform]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transport-netty4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [unsigned-long]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [vectors]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [wildcard]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-analytics]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async-search]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-autoscaling]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ccr]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-core]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-data-streams]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-deprecation]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-enrich]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-eql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-graph]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-identity-provider]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,936Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ilm]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-logstash]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ml]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-monitoring]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-rollup]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-security]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-sql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-stack]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-voting-only-node]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,938Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-watcher]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,943Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"no plugins loaded\" }\nelasticsearch | {\"type\": \"deprecation\", \"timestamp\": \"2026-04-20T06:39:12,054Z\", \"level\": \"DEPRECATION\", \"component\": \"o.e.d.c.s.Settings\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[node.data] setting was deprecated in Elasticsearch and will be removed in a future release! See the breaking changes documentation for the next major version.\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using [1] data paths, mounts [[/usr/share/elasticsearch/data (/dev/vda1)]], net usable_space [14.5gb], net total_space [58.3gb], types [ext4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"heap size [700mb], compressed ordinary object pointers [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,356Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"node name [e802ad473a4f], node ID [e2ZKzgw4Q4aCf2w5ljWr1A], cluster name [docker-cluster], roles [transform, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:18,805Z\", \"level\": \"INFO\", \"component\": \"o.e.x.m.p.l.CppLogMessageHandler\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[controller/213] [Main.cc@114] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,521Z\", \"level\": \"INFO\", \"component\": \"o.e.t.NettyAllocator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,695Z\", \"level\": \"INFO\", \"component\": \"o.e.d.DiscoveryModule\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using discovery type [single-node] and seed hosts providers [settings]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:22,462Z\", \"level\": \"WARN\", \"component\": \"o.e.g.DanglingIndicesState\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,302Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"initialized\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,303Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"starting ...\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,565Z\", \"level\": \"INFO\", \"component\": \"o.e.t.TransportService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9300}, bound_addresses {[::]:9300}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,081Z\", \"level\": \"INFO\", \"component\": \"o.e.c.c.Coordinator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,252Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.MasterService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,401Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.ClusterApplierService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.h.AbstractHttpServerTransport\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9200}, bound_addresses {[::]:9200}\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"started\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,162Z\", \"level\": \"INFO\", \"component\": \"o.e.l.LicenseService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,169Z\", \"level\": \"INFO\", \"component\": \"o.e.g.GatewayService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"recovered [15] indices into cluster_state\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:29,975Z\", \"level\": \"INFO\", \"component\": \"o.e.c.r.a.AllocationService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:02Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"auditTrail\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"visTypeXy\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"warning\",\"config\",\"deprecation\"],\"pid\":7,\"message\":\"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\\\"\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\ndocker_lamp_1 | + create_dot_env_local_file\ndocker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak\ndocker_lamp_1 | + create_dot_env\ndocker_lamp_1 | + [[ -f /home/jiminny/.env ]]\ndocker_lamp_1 | + return\ndocker_lamp_1 | + declare DB_ADMIN_PASSWORD\ndocker_lamp_1 | + declare DB_ADMIN_USERNAME\ndocker_lamp_1 | + declare DB_DEV_PASSWORD\ndocker_lamp_1 | + declare DB_DEV_USERNAME\ndocker_lamp_1 | + declare DB_ROOT_PASSWORD\ndocker_lamp_1 | + declare DB_ROOT_USERNAME\ndocker_lamp_1 | + declare DB_WEB_PASSWORD\ndocker_lamp_1 | + declare DB_WEB_USERNAME\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_PASSWORD='dgyt$rTe21-d'\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_USERNAME=root\ndocker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + readonly DB_ADMIN_PASSWORD\ndocker_lamp_1 | + readonly DB_ADMIN_USERNAME\ndocker_lamp_1 | + readonly DB_DEV_PASSWORD\ndocker_lamp_1 | + readonly DB_DEV_USERNAME\ndocker_lamp_1 | + readonly DB_ROOT_PASSWORD\ndocker_lamp_1 | + readonly DB_ROOT_USERNAME\ndocker_lamp_1 | + readonly DB_WEB_PASSWORD\ndocker_lamp_1 | + readonly DB_WEB_USERNAME\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=dgyt$rTe21-d~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.root\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml\"}\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Session cookies will be transmitted over insecure connections. This is not recommended.\"}\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local\ndocker_lamp_1 | + echo ''\ndocker_lamp_1 | + echo 'DB_ADMIN_PASSWORD=dgyt$rTe21-d'\ndocker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | + echo DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | + echo DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | + echo DB_ROOT_USERNAME=root\ndocker_lamp_1 | + echo DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + declare IP_ADDRESS\ndocker_lamp_1 | ++ curl ifconfig.co/\ndocker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current\ndocker_lamp_1 | Dload Upload Total Spent Left Speed\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"encryptedSavedObjects\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"ingestManager\"],\"pid\":7,\"message\":\"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\n100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224\ndocker_lamp_1 | + IP_ADDRESS=212.5.153.87\ndocker_lamp_1 | + readonly IP_ADDRESS\ndocker_lamp_1 | + sed -i -E 's~X-Forwarded-For \"\\$remote_addr\"~X-Forwarded-For \"IP_ADDRESS\"~g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + sed -i -E s~IP_ADDRESS~212.5.153.87~g /etc/nginx/nginx.conf\ndocker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \\{([^#]*)# QA Only End \\}//g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + storage_init\ndocker_lamp_1 | + false\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]\ndocker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1\ndocker_lamp_1 | ++ wc -l\ndocker_lamp_1 | + [[ 11 -lt 7 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]\ndocker_lamp_1 | + sleep 5\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Found 'server.host: \\\"0\\\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' is being automatically to the configuration. You can change the setting to 'server.host: 0.0.0.0' or add 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' in kibana.yml to prevent this message.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"actions\",\"actions\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"alerts\",\"plugins\",\"alerting\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\"],\"pid\":7,\"message\":\"config sourced from: production cluster\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Starting saved objects migrations\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\n\n\nv View in Docker Desktop o View Config w Enable Watch","depth":4,"value":"kibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:35Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:35Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:35Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:37Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:38Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:38Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:38Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:40Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:42Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:45Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:48Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:49Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"reporting\",\"esqueue\",\"queue-worker\",\"error\"],\"pid\":7,\"message\":\"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\\n at process._tickCallback (internal/process/next_tick.js:61:11)\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:55Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:58Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:00Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:03Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:10Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work\nWARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion \n[+] Running 2/2\n ✔ Container redis Running 0.0s \n ✔ Container docker_lamp_1 Running 0.0s \nAttaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis\ndocker_lamp_1 | \ndocker_lamp_1 | Illuminate\\Database\\QueryException \ndocker_lamp_1 | \ndocker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)\ndocker_lamp_1 | \ndocker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838\ndocker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)\ndocker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class\ndocker_lamp_1 | 836▕ : QueryException::class;\ndocker_lamp_1 | 837▕ \ndocker_lamp_1 | ➜ 838▕ throw new $exceptionType(\ndocker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),\ndocker_lamp_1 | 840▕ $query,\ndocker_lamp_1 | 841▕ $this->prepareBindings($bindings),\ndocker_lamp_1 | 842▕ $e,\ndocker_lamp_1 | \ndocker_lamp_1 | +42 vendor frames \ndocker_lamp_1 | \ndocker_lamp_1 | 43 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \nmariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\nblackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.\nblackfire-1 | usage blackfire-agent [options]\nblackfire-1 | --collector=\"https://blackfire.io\": Sets the URL of Blackfire's data collector\nblackfire-1 | --config=\"/etc/blackfire/agent\": Sets the path to the configuration file\nblackfire-1 | -d: Prints the current configuration\nblackfire-1 | --http-proxy=\"\": Sets the HTTP proxy to use\nblackfire-1 | --https-proxy=\"\": Sets the HTTPS proxy to use\nblackfire-1 | --log-file=\"stderr\": Sets the path of the log file. Use stderr to log to stderr\nblackfire-1 | --log-level=\"1\": log verbosity level (4: debug, 3: info, 2: warning, 1: error)\nblackfire-1 | --register: Helps you with registering the agent\nblackfire-1 | --server-id=\"\": Sets the server id used to authenticate with Blackfire API\nblackfire-1 | --server-token=\"\": Sets the server token used to authenticate with Blackfire API. It is unsafe to set this from the command line\nblackfire-1 | --socket=\"unix:///var/run/blackfire/agent.sock\": Sets the socket the agent should read traces from. Possible value can be a unix socket or a TCP address. ie: unix:///var/run/blackfire/agent.sock or tcp://127.0.0.1:8307\nblackfire-1 | --test: Tests the configuration\nblackfire-1 | --timeout=\"15s\": Sets the Blackfire connection timeout\nblackfire-1 | -v: Prints the version number\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"no configuration paths supplied\"\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"using configuration at default config path\" path=/home/ngrok/.ngrok2/ngrok.yml\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"open config file\" path=/home/ngrok/.ngrok2/ngrok.yml err=nil\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"starting web service\" obj=web addr=0.0.0.0:4040\njiminny_ext-1 exited with code 0\nblackfire-1 exited with code 1\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"tunnel session started\" obj=tunnels.session\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"client session established\" obj=csess id=7d0b19e3215d\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=\"command_line (http)\" addr=http://lamp:3080 url=http://lukask.ngrok.io\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=command_line addr=http://lamp:3080 url=https://lukask.ngrok.io\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Warn] [Entrypoint]: /sys/fs/cgroup///memory.pressure not writable, functionality unavailable to MariaDB\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\ndatadog-1 | [s6-init] making user provided files available at /var/run/s6/etc...exited 0.\nngrok | t=2026-04-20T06:38:55+0000 lvl=info msg=\"update available\" obj=updater\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade information missing, assuming required\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade (mariadb-upgrade or creating healthcheck users) required, but skipped due to $MARIADB_AUTO_UPGRADE setting\nmariadb-1 | 2026-04-20 6:38:55 0 [Note] Starting MariaDB 11.4.5-MariaDB-ubu2404 source revision 0771110266ff5c04216af4bf1243c65f8c67ccf4 server_uid geAM8SSFn8P6Gz+0WMyyW27EUEM= as process 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Compressed tables use zlib 1.3\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Number of transaction pools: 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using ARMv8 crc32 + pmull instructions\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using liburing\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Completed initialization of buffer pool\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes)\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: End of log at LSN=7614257762\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Opened 3 undo tablespaces\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: 128 rollback segments in 3 undo tablespaces are active.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ...\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: log sequence number 7614257762; transaction id 11481082\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'FEEDBACK' is disabled.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'wsrep-provider' is disabled.\nmariadb-1 | 2026-04-20 6:38:58 0 [Note] InnoDB: Buffer pool(s) load completed at 260420 6:38:58\ndatadog-1 | [s6-init] ensuring user provided files have correct perms...exited 0.\ndatadog-1 | [fix-attrs.d] applying ownership & permissions fixes...\ndatadog-1 | [fix-attrs.d] done.\ndatadog-1 | [cont-init.d] executing container initialization scripts...\ndatadog-1 | [cont-init.d] 01-check-apikey.sh: executing... \ndatadog-1 | \ndatadog-1 | ==================================================================================\ndatadog-1 | You must set an DD_API_KEY environment variable to run the Datadog Agent container\ndatadog-1 | ==================================================================================\ndatadog-1 | \ndatadog-1 | [cont-init.d] 01-check-apikey.sh: exited 1.\ndatadog-1 exited with code 1\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '0.0.0.0'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '::'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: Event Scheduler: Loaded 0 events\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: ready for connections.\nmariadb-1 | Version: '11.4.5-MariaDB-ubu2404' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution\ndocker_lamp_1 | \ndocker_lamp_1 | INFO Configuration cached successfully. \ndocker_lamp_1 | \ndocker_lamp_1 | cp: cannot stat '/home/jiminny/bootstrap/cache/config-new.php': No such file or directory\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + [[ false == \\f\\a\\l\\s\\e ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + [[ -f /home/jiminny/storage/oauth-private.key ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + setup_local_environment\ndocker_lamp_1 | + storage_permissions_workaround\ndocker_lamp_1 | + [[ false != \\t\\r\\u\\e ]]\ndocker_lamp_1 | + ln -sf /home/jiminny/storage /home/jiminny/jiminny_storage\ndocker_lamp_1 | + return\ndocker_lamp_1 | + chmod 660 /home/jiminny/storage/oauth-private.key\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/views\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/cache\ndocker_lamp_1 | + chmod 775 /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/cache\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,413Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"version[7.10.2], pid[7], build[default/docker/747e1cc71def077253878a59143c1f785afa92b9/2021-01-13T04:42:47.157277Z], OS[Linux/6.12.54-linuxkit/aarch64], JVM[AdoptOpenJDK/OpenJDK 64-Bit Server VM/15.0.1/15.0.1+9]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM home [/usr/share/elasticsearch/jdk], using bundled JDK [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Djava.locale.providers=SPI,COMPAT, -Xms1g, -Xmx1g, -XX:+UseG1GC, -XX:G1ReservePercent=25, -XX:InitiatingHeapOccupancyPercent=30, -Djava.io.tmpdir=/tmp/elasticsearch-11712280666606670653, -XX:+HeapDumpOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -Des.cgroups.hierarchy.override=/, -Xms700m, -Xmx700m, -XX:MaxDirectMemorySize=367001600, -Des.path.home=/usr/share/elasticsearch, -Des.path.conf=/usr/share/elasticsearch/config, -Des.distribution.flavor=default, -Des.distribution.type=docker, -Des.bundled_jdk=true]\" }\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chmod g+w '{}' ';'\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,926Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [aggs-matrix-stats]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [analysis-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [constant-keyword]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [flattened]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [frozen-indices]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-geoip]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-user-agent]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [kibana]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-expression]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-mustache]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-painless]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-extras]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-version]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [parent-join]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [percolator]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [rank-eval]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [reindex]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repositories-metering-api]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repository-url]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [search-business-rules]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [searchable-snapshots]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [spatial]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transform]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transport-netty4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [unsigned-long]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [vectors]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [wildcard]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-analytics]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async-search]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-autoscaling]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ccr]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-core]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-data-streams]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-deprecation]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-enrich]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-eql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-graph]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-identity-provider]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,936Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ilm]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-logstash]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ml]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-monitoring]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-rollup]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-security]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-sql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-stack]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-voting-only-node]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,938Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-watcher]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,943Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"no plugins loaded\" }\nelasticsearch | {\"type\": \"deprecation\", \"timestamp\": \"2026-04-20T06:39:12,054Z\", \"level\": \"DEPRECATION\", \"component\": \"o.e.d.c.s.Settings\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[node.data] setting was deprecated in Elasticsearch and will be removed in a future release! See the breaking changes documentation for the next major version.\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using [1] data paths, mounts [[/usr/share/elasticsearch/data (/dev/vda1)]], net usable_space [14.5gb], net total_space [58.3gb], types [ext4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"heap size [700mb], compressed ordinary object pointers [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,356Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"node name [e802ad473a4f], node ID [e2ZKzgw4Q4aCf2w5ljWr1A], cluster name [docker-cluster], roles [transform, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:18,805Z\", \"level\": \"INFO\", \"component\": \"o.e.x.m.p.l.CppLogMessageHandler\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[controller/213] [Main.cc@114] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,521Z\", \"level\": \"INFO\", \"component\": \"o.e.t.NettyAllocator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,695Z\", \"level\": \"INFO\", \"component\": \"o.e.d.DiscoveryModule\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using discovery type [single-node] and seed hosts providers [settings]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:22,462Z\", \"level\": \"WARN\", \"component\": \"o.e.g.DanglingIndicesState\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,302Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"initialized\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,303Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"starting ...\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,565Z\", \"level\": \"INFO\", \"component\": \"o.e.t.TransportService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9300}, bound_addresses {[::]:9300}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,081Z\", \"level\": \"INFO\", \"component\": \"o.e.c.c.Coordinator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,252Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.MasterService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,401Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.ClusterApplierService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.h.AbstractHttpServerTransport\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9200}, bound_addresses {[::]:9200}\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"started\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,162Z\", \"level\": \"INFO\", \"component\": \"o.e.l.LicenseService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,169Z\", \"level\": \"INFO\", \"component\": \"o.e.g.GatewayService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"recovered [15] indices into cluster_state\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:29,975Z\", \"level\": \"INFO\", \"component\": \"o.e.c.r.a.AllocationService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:02Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"auditTrail\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"visTypeXy\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"warning\",\"config\",\"deprecation\"],\"pid\":7,\"message\":\"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\\\"\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\ndocker_lamp_1 | + create_dot_env_local_file\ndocker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak\ndocker_lamp_1 | + create_dot_env\ndocker_lamp_1 | + [[ -f /home/jiminny/.env ]]\ndocker_lamp_1 | + return\ndocker_lamp_1 | + declare DB_ADMIN_PASSWORD\ndocker_lamp_1 | + declare DB_ADMIN_USERNAME\ndocker_lamp_1 | + declare DB_DEV_PASSWORD\ndocker_lamp_1 | + declare DB_DEV_USERNAME\ndocker_lamp_1 | + declare DB_ROOT_PASSWORD\ndocker_lamp_1 | + declare DB_ROOT_USERNAME\ndocker_lamp_1 | + declare DB_WEB_PASSWORD\ndocker_lamp_1 | + declare DB_WEB_USERNAME\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_PASSWORD='dgyt$rTe21-d'\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_USERNAME=root\ndocker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + readonly DB_ADMIN_PASSWORD\ndocker_lamp_1 | + readonly DB_ADMIN_USERNAME\ndocker_lamp_1 | + readonly DB_DEV_PASSWORD\ndocker_lamp_1 | + readonly DB_DEV_USERNAME\ndocker_lamp_1 | + readonly DB_ROOT_PASSWORD\ndocker_lamp_1 | + readonly DB_ROOT_USERNAME\ndocker_lamp_1 | + readonly DB_WEB_PASSWORD\ndocker_lamp_1 | + readonly DB_WEB_USERNAME\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=dgyt$rTe21-d~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.root\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml\"}\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Session cookies will be transmitted over insecure connections. This is not recommended.\"}\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local\ndocker_lamp_1 | + echo ''\ndocker_lamp_1 | + echo 'DB_ADMIN_PASSWORD=dgyt$rTe21-d'\ndocker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | + echo DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | + echo DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | + echo DB_ROOT_USERNAME=root\ndocker_lamp_1 | + echo DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + declare IP_ADDRESS\ndocker_lamp_1 | ++ curl ifconfig.co/\ndocker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current\ndocker_lamp_1 | Dload Upload Total Spent Left Speed\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"encryptedSavedObjects\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"ingestManager\"],\"pid\":7,\"message\":\"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\n100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224\ndocker_lamp_1 | + IP_ADDRESS=212.5.153.87\ndocker_lamp_1 | + readonly IP_ADDRESS\ndocker_lamp_1 | + sed -i -E 's~X-Forwarded-For \"\\$remote_addr\"~X-Forwarded-For \"IP_ADDRESS\"~g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + sed -i -E s~IP_ADDRESS~212.5.153.87~g /etc/nginx/nginx.conf\ndocker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \\{([^#]*)# QA Only End \\}//g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + storage_init\ndocker_lamp_1 | + false\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]\ndocker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1\ndocker_lamp_1 | ++ wc -l\ndocker_lamp_1 | + [[ 11 -lt 7 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]\ndocker_lamp_1 | + sleep 5\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Found 'server.host: \\\"0\\\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' is being automatically to the configuration. You can change the setting to 'server.host: 0.0.0.0' or add 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' in kibana.yml to prevent this message.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"actions\",\"actions\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"alerts\",\"plugins\",\"alerting\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\"],\"pid\":7,\"message\":\"config sourced from: production cluster\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Starting saved objects migrations\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\n\n\nv View in Docker Desktop o View Config w Enable Watch","is_focused":true},{"role":"AXButton","text":"Menu","depth":3,"bounds":{"left":0.48402777,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥1 DOCKER (docker-compose)","depth":3,"bounds":{"left":0.01875,"top":0.09,"width":0.46180555,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥2 PROD (ssh)","depth":4,"bounds":{"left":0.51805556,"top":0.09,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.24944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥3 EU (ssh)","depth":4,"bounds":{"left":0.51805556,"top":0.25,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.4422222,"width":0.50069445,"height":0.12222222},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.41944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥4 STAGE (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.42,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.5922222,"width":0.50069445,"height":0.12222222},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.5694444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥5 QA (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.57,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:48:07 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.74222225,"width":0.50069445,"height":0.12222222},"value":"Last login: Thu Apr 16 15:48:07 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.71944445,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥6 FE (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.72,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:48:07 on ttys005\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.87222224,"width":0.50069445,"height":0.12777776},"value":"Last login: Thu Apr 16 15:48:07 on ttys005\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.86944443,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥7 EXT (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.87,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.140625,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.14479166,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.28125,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.28541666,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.421875,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.42604166,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.5625,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.56666666,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"✳ Review screenpipe usage and Boosteroid integration (claude)","depth":2,"bounds":{"left":0.7027778,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.70694447,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.84305555,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.8472222,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.9548611,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"EU (ssh)","depth":1,"bounds":{"left":0.47916666,"top":0.033333335,"width":0.04097222,"height":0.017777778},"role_description":"text"}]...
|
9162762734871117331
|
-8918215769196404978
|
visual_change
|
accessibility
|
NULL
|
kibana | {"type":"log", kibana | {"type":"log","@timestamp":"2026-04-18T13:26:35Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:35Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:37Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:38Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:38Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:40Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:41Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:41Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:42Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:49Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ...
|
52128
|
|
52153
|
1126
|
49
|
2026-04-20T06:42:02.839965+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667322839_m1.jpg...
|
iTerm2
|
EU (ssh)
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
asticsearch","data"],"pid":7,& asticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:41Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:42Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:49Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_ROOT_USERNAME=root
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local
docker_lamp_1 | + declare IP_ADDRESS
docker_lamp_1 | ++ curl ifconfig.co/
docker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current
docker_lamp_1 | Dload Upload Total Spent Left Speed
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","encryptedSavedObjects","config"],"pid":7,"message":"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","ingestManager"],"pid":7,"message":"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224
docker_lamp_1 | + IP_ADDRESS=[IP_ADDRESS]
docker_lamp_1 | + readonly IP_ADDRESS
docker_lamp_1 | + sed -i -E 's~X-Forwarded-For "\$remote_addr"~X-Forwarded-For "IP_ADDRESS"~g' /etc/nginx/nginx.conf
docker_lamp_1 | + sed -i -E s~IP_ADDRESS~[IP_ADDRESS]~g /etc/nginx/nginx.conf
docker_lamp_...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"asticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:42Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:45Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:48Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:49Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"reporting\",\"esqueue\",\"queue-worker\",\"error\"],\"pid\":7,\"message\":\"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\\n at process._tickCallback (internal/process/next_tick.js:61:11)\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:55Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:58Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:00Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:03Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:10Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work\nWARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion \n[+] Running 2/2\n ✔ Container redis Running 0.0s \n ✔ Container docker_lamp_1 Running 0.0s \nAttaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis\ndocker_lamp_1 | \ndocker_lamp_1 | Illuminate\\Database\\QueryException \ndocker_lamp_1 | \ndocker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)\ndocker_lamp_1 | \ndocker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838\ndocker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)\ndocker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class\ndocker_lamp_1 | 836▕ : QueryException::class;\ndocker_lamp_1 | 837▕ \ndocker_lamp_1 | ➜ 838▕ throw new $exceptionType(\ndocker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),\ndocker_lamp_1 | 840▕ $query,\ndocker_lamp_1 | 841▕ $this->prepareBindings($bindings),\ndocker_lamp_1 | 842▕ $e,\ndocker_lamp_1 | \ndocker_lamp_1 | +42 vendor frames \ndocker_lamp_1 | \ndocker_lamp_1 | 43 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \nmariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\nblackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.\nblackfire-1 | usage blackfire-agent [options]\nblackfire-1 | --collector=\"https://blackfire.io\": Sets the URL of Blackfire's data collector\nblackfire-1 | --config=\"/etc/blackfire/agent\": Sets the path to the configuration file\nblackfire-1 | -d: Prints the current configuration\nblackfire-1 | --http-proxy=\"\": Sets the HTTP proxy to use\nblackfire-1 | --https-proxy=\"\": Sets the HTTPS proxy to use\nblackfire-1 | --log-file=\"stderr\": Sets the path of the log file. Use stderr to log to stderr\nblackfire-1 | --log-level=\"1\": log verbosity level (4: debug, 3: info, 2: warning, 1: error)\nblackfire-1 | --register: Helps you with registering the agent\nblackfire-1 | --server-id=\"\": Sets the server id used to authenticate with Blackfire API\nblackfire-1 | --server-token=\"\": Sets the server token used to authenticate with Blackfire API. It is unsafe to set this from the command line\nblackfire-1 | --socket=\"unix:///var/run/blackfire/agent.sock\": Sets the socket the agent should read traces from. Possible value can be a unix socket or a TCP address. ie: unix:///var/run/blackfire/agent.sock or tcp://127.0.0.1:8307\nblackfire-1 | --test: Tests the configuration\nblackfire-1 | --timeout=\"15s\": Sets the Blackfire connection timeout\nblackfire-1 | -v: Prints the version number\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"no configuration paths supplied\"\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"using configuration at default config path\" path=/home/ngrok/.ngrok2/ngrok.yml\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"open config file\" path=/home/ngrok/.ngrok2/ngrok.yml err=nil\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"starting web service\" obj=web addr=0.0.0.0:4040\njiminny_ext-1 exited with code 0\nblackfire-1 exited with code 1\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"tunnel session started\" obj=tunnels.session\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"client session established\" obj=csess id=7d0b19e3215d\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=\"command_line (http)\" addr=http://lamp:3080 url=http://lukask.ngrok.io\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=command_line addr=http://lamp:3080 url=https://lukask.ngrok.io\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Warn] [Entrypoint]: /sys/fs/cgroup///memory.pressure not writable, functionality unavailable to MariaDB\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\ndatadog-1 | [s6-init] making user provided files available at /var/run/s6/etc...exited 0.\nngrok | t=2026-04-20T06:38:55+0000 lvl=info msg=\"update available\" obj=updater\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade information missing, assuming required\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade (mariadb-upgrade or creating healthcheck users) required, but skipped due to $MARIADB_AUTO_UPGRADE setting\nmariadb-1 | 2026-04-20 6:38:55 0 [Note] Starting MariaDB 11.4.5-MariaDB-ubu2404 source revision 0771110266ff5c04216af4bf1243c65f8c67ccf4 server_uid geAM8SSFn8P6Gz+0WMyyW27EUEM= as process 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Compressed tables use zlib 1.3\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Number of transaction pools: 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using ARMv8 crc32 + pmull instructions\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using liburing\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Completed initialization of buffer pool\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes)\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: End of log at LSN=7614257762\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Opened 3 undo tablespaces\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: 128 rollback segments in 3 undo tablespaces are active.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ...\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: log sequence number 7614257762; transaction id 11481082\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'FEEDBACK' is disabled.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'wsrep-provider' is disabled.\nmariadb-1 | 2026-04-20 6:38:58 0 [Note] InnoDB: Buffer pool(s) load completed at 260420 6:38:58\ndatadog-1 | [s6-init] ensuring user provided files have correct perms...exited 0.\ndatadog-1 | [fix-attrs.d] applying ownership & permissions fixes...\ndatadog-1 | [fix-attrs.d] done.\ndatadog-1 | [cont-init.d] executing container initialization scripts...\ndatadog-1 | [cont-init.d] 01-check-apikey.sh: executing... \ndatadog-1 | \ndatadog-1 | ==================================================================================\ndatadog-1 | You must set an DD_API_KEY environment variable to run the Datadog Agent container\ndatadog-1 | ==================================================================================\ndatadog-1 | \ndatadog-1 | [cont-init.d] 01-check-apikey.sh: exited 1.\ndatadog-1 exited with code 1\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '0.0.0.0'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '::'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: Event Scheduler: Loaded 0 events\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: ready for connections.\nmariadb-1 | Version: '11.4.5-MariaDB-ubu2404' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution\ndocker_lamp_1 | \ndocker_lamp_1 | INFO Configuration cached successfully. \ndocker_lamp_1 | \ndocker_lamp_1 | cp: cannot stat '/home/jiminny/bootstrap/cache/config-new.php': No such file or directory\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + [[ false == \\f\\a\\l\\s\\e ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + [[ -f /home/jiminny/storage/oauth-private.key ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + setup_local_environment\ndocker_lamp_1 | + storage_permissions_workaround\ndocker_lamp_1 | + [[ false != \\t\\r\\u\\e ]]\ndocker_lamp_1 | + ln -sf /home/jiminny/storage /home/jiminny/jiminny_storage\ndocker_lamp_1 | + return\ndocker_lamp_1 | + chmod 660 /home/jiminny/storage/oauth-private.key\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/views\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/cache\ndocker_lamp_1 | + chmod 775 /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/cache\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,413Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"version[7.10.2], pid[7], build[default/docker/747e1cc71def077253878a59143c1f785afa92b9/2021-01-13T04:42:47.157277Z], OS[Linux/6.12.54-linuxkit/aarch64], JVM[AdoptOpenJDK/OpenJDK 64-Bit Server VM/15.0.1/15.0.1+9]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM home [/usr/share/elasticsearch/jdk], using bundled JDK [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Djava.locale.providers=SPI,COMPAT, -Xms1g, -Xmx1g, -XX:+UseG1GC, -XX:G1ReservePercent=25, -XX:InitiatingHeapOccupancyPercent=30, -Djava.io.tmpdir=/tmp/elasticsearch-11712280666606670653, -XX:+HeapDumpOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -Des.cgroups.hierarchy.override=/, -Xms700m, -Xmx700m, -XX:MaxDirectMemorySize=367001600, -Des.path.home=/usr/share/elasticsearch, -Des.path.conf=/usr/share/elasticsearch/config, -Des.distribution.flavor=default, -Des.distribution.type=docker, -Des.bundled_jdk=true]\" }\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chmod g+w '{}' ';'\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,926Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [aggs-matrix-stats]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [analysis-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [constant-keyword]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [flattened]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [frozen-indices]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-geoip]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-user-agent]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [kibana]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-expression]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-mustache]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-painless]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-extras]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-version]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [parent-join]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [percolator]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [rank-eval]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [reindex]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repositories-metering-api]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repository-url]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [search-business-rules]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [searchable-snapshots]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [spatial]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transform]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transport-netty4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [unsigned-long]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [vectors]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [wildcard]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-analytics]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async-search]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-autoscaling]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ccr]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-core]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-data-streams]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-deprecation]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-enrich]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-eql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-graph]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-identity-provider]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,936Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ilm]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-logstash]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ml]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-monitoring]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-rollup]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-security]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-sql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-stack]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-voting-only-node]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,938Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-watcher]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,943Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"no plugins loaded\" }\nelasticsearch | {\"type\": \"deprecation\", \"timestamp\": \"2026-04-20T06:39:12,054Z\", \"level\": \"DEPRECATION\", \"component\": \"o.e.d.c.s.Settings\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[node.data] setting was deprecated in Elasticsearch and will be removed in a future release! See the breaking changes documentation for the next major version.\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using [1] data paths, mounts [[/usr/share/elasticsearch/data (/dev/vda1)]], net usable_space [14.5gb], net total_space [58.3gb], types [ext4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"heap size [700mb], compressed ordinary object pointers [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,356Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"node name [e802ad473a4f], node ID [e2ZKzgw4Q4aCf2w5ljWr1A], cluster name [docker-cluster], roles [transform, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:18,805Z\", \"level\": \"INFO\", \"component\": \"o.e.x.m.p.l.CppLogMessageHandler\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[controller/213] [Main.cc@114] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,521Z\", \"level\": \"INFO\", \"component\": \"o.e.t.NettyAllocator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,695Z\", \"level\": \"INFO\", \"component\": \"o.e.d.DiscoveryModule\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using discovery type [single-node] and seed hosts providers [settings]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:22,462Z\", \"level\": \"WARN\", \"component\": \"o.e.g.DanglingIndicesState\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,302Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"initialized\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,303Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"starting ...\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,565Z\", \"level\": \"INFO\", \"component\": \"o.e.t.TransportService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9300}, bound_addresses {[::]:9300}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,081Z\", \"level\": \"INFO\", \"component\": \"o.e.c.c.Coordinator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,252Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.MasterService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,401Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.ClusterApplierService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.h.AbstractHttpServerTransport\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9200}, bound_addresses {[::]:9200}\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"started\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,162Z\", \"level\": \"INFO\", \"component\": \"o.e.l.LicenseService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,169Z\", \"level\": \"INFO\", \"component\": \"o.e.g.GatewayService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"recovered [15] indices into cluster_state\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:29,975Z\", \"level\": \"INFO\", \"component\": \"o.e.c.r.a.AllocationService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:02Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"auditTrail\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"visTypeXy\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"warning\",\"config\",\"deprecation\"],\"pid\":7,\"message\":\"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\\\"\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\ndocker_lamp_1 | + create_dot_env_local_file\ndocker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak\ndocker_lamp_1 | + create_dot_env\ndocker_lamp_1 | + [[ -f /home/jiminny/.env ]]\ndocker_lamp_1 | + return\ndocker_lamp_1 | + declare DB_ADMIN_PASSWORD\ndocker_lamp_1 | + declare DB_ADMIN_USERNAME\ndocker_lamp_1 | + declare DB_DEV_PASSWORD\ndocker_lamp_1 | + declare DB_DEV_USERNAME\ndocker_lamp_1 | + declare DB_ROOT_PASSWORD\ndocker_lamp_1 | + declare DB_ROOT_USERNAME\ndocker_lamp_1 | + declare DB_WEB_PASSWORD\ndocker_lamp_1 | + declare DB_WEB_USERNAME\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_PASSWORD='dgyt$rTe21-d'\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_USERNAME=root\ndocker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + readonly DB_ADMIN_PASSWORD\ndocker_lamp_1 | + readonly DB_ADMIN_USERNAME\ndocker_lamp_1 | + readonly DB_DEV_PASSWORD\ndocker_lamp_1 | + readonly DB_DEV_USERNAME\ndocker_lamp_1 | + readonly DB_ROOT_PASSWORD\ndocker_lamp_1 | + readonly DB_ROOT_USERNAME\ndocker_lamp_1 | + readonly DB_WEB_PASSWORD\ndocker_lamp_1 | + readonly DB_WEB_USERNAME\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=dgyt$rTe21-d~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.root\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml\"}\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Session cookies will be transmitted over insecure connections. This is not recommended.\"}\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local\ndocker_lamp_1 | + echo ''\ndocker_lamp_1 | + echo 'DB_ADMIN_PASSWORD=dgyt$rTe21-d'\ndocker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | + echo DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | + echo DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | + echo DB_ROOT_USERNAME=root\ndocker_lamp_1 | + echo DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + declare IP_ADDRESS\ndocker_lamp_1 | ++ curl ifconfig.co/\ndocker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current\ndocker_lamp_1 | Dload Upload Total Spent Left Speed\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"encryptedSavedObjects\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"ingestManager\"],\"pid\":7,\"message\":\"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\n100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224\ndocker_lamp_1 | + IP_ADDRESS=212.5.153.87\ndocker_lamp_1 | + readonly IP_ADDRESS\ndocker_lamp_1 | + sed -i -E 's~X-Forwarded-For \"\\$remote_addr\"~X-Forwarded-For \"IP_ADDRESS\"~g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + sed -i -E s~IP_ADDRESS~212.5.153.87~g /etc/nginx/nginx.conf\ndocker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \\{([^#]*)# QA Only End \\}//g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + storage_init\ndocker_lamp_1 | + false\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]\ndocker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1\ndocker_lamp_1 | ++ wc -l\ndocker_lamp_1 | + [[ 11 -lt 7 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]\ndocker_lamp_1 | + sleep 5\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Found 'server.host: \\\"0\\\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' is being automatically to the configuration. You can change the setting to 'server.host: 0.0.0.0' or add 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' in kibana.yml to prevent this message.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"actions\",\"actions\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"alerts\",\"plugins\",\"alerting\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\"],\"pid\":7,\"message\":\"config sourced from: production cluster\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Starting saved objects migrations\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\n\n\nv View in Docker Desktop o View Config w Enable Watch","depth":4,"value":"asticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:42Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:45Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:48Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:49Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"reporting\",\"esqueue\",\"queue-worker\",\"error\"],\"pid\":7,\"message\":\"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\\n at process._tickCallback (internal/process/next_tick.js:61:11)\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:55Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:58Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:00Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:03Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:10Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work\nWARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion \n[+] Running 2/2\n ✔ Container redis Running 0.0s \n ✔ Container docker_lamp_1 Running 0.0s \nAttaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis\ndocker_lamp_1 | \ndocker_lamp_1 | Illuminate\\Database\\QueryException \ndocker_lamp_1 | \ndocker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)\ndocker_lamp_1 | \ndocker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838\ndocker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)\ndocker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class\ndocker_lamp_1 | 836▕ : QueryException::class;\ndocker_lamp_1 | 837▕ \ndocker_lamp_1 | ➜ 838▕ throw new $exceptionType(\ndocker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),\ndocker_lamp_1 | 840▕ $query,\ndocker_lamp_1 | 841▕ $this->prepareBindings($bindings),\ndocker_lamp_1 | 842▕ $e,\ndocker_lamp_1 | \ndocker_lamp_1 | +42 vendor frames \ndocker_lamp_1 | \ndocker_lamp_1 | 43 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \nmariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\nblackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.\nblackfire-1 | usage blackfire-agent [options]\nblackfire-1 | --collector=\"https://blackfire.io\": Sets the URL of Blackfire's data collector\nblackfire-1 | --config=\"/etc/blackfire/agent\": Sets the path to the configuration file\nblackfire-1 | -d: Prints the current configuration\nblackfire-1 | --http-proxy=\"\": Sets the HTTP proxy to use\nblackfire-1 | --https-proxy=\"\": Sets the HTTPS proxy to use\nblackfire-1 | --log-file=\"stderr\": Sets the path of the log file. Use stderr to log to stderr\nblackfire-1 | --log-level=\"1\": log verbosity level (4: debug, 3: info, 2: warning, 1: error)\nblackfire-1 | --register: Helps you with registering the agent\nblackfire-1 | --server-id=\"\": Sets the server id used to authenticate with Blackfire API\nblackfire-1 | --server-token=\"\": Sets the server token used to authenticate with Blackfire API. It is unsafe to set this from the command line\nblackfire-1 | --socket=\"unix:///var/run/blackfire/agent.sock\": Sets the socket the agent should read traces from. Possible value can be a unix socket or a TCP address. ie: unix:///var/run/blackfire/agent.sock or tcp://127.0.0.1:8307\nblackfire-1 | --test: Tests the configuration\nblackfire-1 | --timeout=\"15s\": Sets the Blackfire connection timeout\nblackfire-1 | -v: Prints the version number\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"no configuration paths supplied\"\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"using configuration at default config path\" path=/home/ngrok/.ngrok2/ngrok.yml\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"open config file\" path=/home/ngrok/.ngrok2/ngrok.yml err=nil\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"starting web service\" obj=web addr=0.0.0.0:4040\njiminny_ext-1 exited with code 0\nblackfire-1 exited with code 1\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"tunnel session started\" obj=tunnels.session\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"client session established\" obj=csess id=7d0b19e3215d\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=\"command_line (http)\" addr=http://lamp:3080 url=http://lukask.ngrok.io\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=command_line addr=http://lamp:3080 url=https://lukask.ngrok.io\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Warn] [Entrypoint]: /sys/fs/cgroup///memory.pressure not writable, functionality unavailable to MariaDB\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\ndatadog-1 | [s6-init] making user provided files available at /var/run/s6/etc...exited 0.\nngrok | t=2026-04-20T06:38:55+0000 lvl=info msg=\"update available\" obj=updater\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade information missing, assuming required\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade (mariadb-upgrade or creating healthcheck users) required, but skipped due to $MARIADB_AUTO_UPGRADE setting\nmariadb-1 | 2026-04-20 6:38:55 0 [Note] Starting MariaDB 11.4.5-MariaDB-ubu2404 source revision 0771110266ff5c04216af4bf1243c65f8c67ccf4 server_uid geAM8SSFn8P6Gz+0WMyyW27EUEM= as process 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Compressed tables use zlib 1.3\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Number of transaction pools: 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using ARMv8 crc32 + pmull instructions\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using liburing\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Completed initialization of buffer pool\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes)\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: End of log at LSN=7614257762\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Opened 3 undo tablespaces\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: 128 rollback segments in 3 undo tablespaces are active.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ...\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: log sequence number 7614257762; transaction id 11481082\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'FEEDBACK' is disabled.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'wsrep-provider' is disabled.\nmariadb-1 | 2026-04-20 6:38:58 0 [Note] InnoDB: Buffer pool(s) load completed at 260420 6:38:58\ndatadog-1 | [s6-init] ensuring user provided files have correct perms...exited 0.\ndatadog-1 | [fix-attrs.d] applying ownership & permissions fixes...\ndatadog-1 | [fix-attrs.d] done.\ndatadog-1 | [cont-init.d] executing container initialization scripts...\ndatadog-1 | [cont-init.d] 01-check-apikey.sh: executing... \ndatadog-1 | \ndatadog-1 | ==================================================================================\ndatadog-1 | You must set an DD_API_KEY environment variable to run the Datadog Agent container\ndatadog-1 | ==================================================================================\ndatadog-1 | \ndatadog-1 | [cont-init.d] 01-check-apikey.sh: exited 1.\ndatadog-1 exited with code 1\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '0.0.0.0'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '::'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: Event Scheduler: Loaded 0 events\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: ready for connections.\nmariadb-1 | Version: '11.4.5-MariaDB-ubu2404' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution\ndocker_lamp_1 | \ndocker_lamp_1 | INFO Configuration cached successfully. \ndocker_lamp_1 | \ndocker_lamp_1 | cp: cannot stat '/home/jiminny/bootstrap/cache/config-new.php': No such file or directory\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + [[ false == \\f\\a\\l\\s\\e ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + [[ -f /home/jiminny/storage/oauth-private.key ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + setup_local_environment\ndocker_lamp_1 | + storage_permissions_workaround\ndocker_lamp_1 | + [[ false != \\t\\r\\u\\e ]]\ndocker_lamp_1 | + ln -sf /home/jiminny/storage /home/jiminny/jiminny_storage\ndocker_lamp_1 | + return\ndocker_lamp_1 | + chmod 660 /home/jiminny/storage/oauth-private.key\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/views\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/cache\ndocker_lamp_1 | + chmod 775 /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/cache\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,413Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"version[7.10.2], pid[7], build[default/docker/747e1cc71def077253878a59143c1f785afa92b9/2021-01-13T04:42:47.157277Z], OS[Linux/6.12.54-linuxkit/aarch64], JVM[AdoptOpenJDK/OpenJDK 64-Bit Server VM/15.0.1/15.0.1+9]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM home [/usr/share/elasticsearch/jdk], using bundled JDK [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Djava.locale.providers=SPI,COMPAT, -Xms1g, -Xmx1g, -XX:+UseG1GC, -XX:G1ReservePercent=25, -XX:InitiatingHeapOccupancyPercent=30, -Djava.io.tmpdir=/tmp/elasticsearch-11712280666606670653, -XX:+HeapDumpOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -Des.cgroups.hierarchy.override=/, -Xms700m, -Xmx700m, -XX:MaxDirectMemorySize=367001600, -Des.path.home=/usr/share/elasticsearch, -Des.path.conf=/usr/share/elasticsearch/config, -Des.distribution.flavor=default, -Des.distribution.type=docker, -Des.bundled_jdk=true]\" }\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chmod g+w '{}' ';'\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,926Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [aggs-matrix-stats]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [analysis-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [constant-keyword]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [flattened]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [frozen-indices]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-geoip]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-user-agent]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [kibana]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-expression]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-mustache]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-painless]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-extras]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-version]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [parent-join]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [percolator]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [rank-eval]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [reindex]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repositories-metering-api]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repository-url]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [search-business-rules]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [searchable-snapshots]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [spatial]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transform]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transport-netty4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [unsigned-long]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [vectors]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [wildcard]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-analytics]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async-search]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-autoscaling]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ccr]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-core]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-data-streams]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-deprecation]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-enrich]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-eql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-graph]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-identity-provider]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,936Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ilm]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-logstash]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ml]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-monitoring]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-rollup]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-security]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-sql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-stack]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-voting-only-node]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,938Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-watcher]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,943Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"no plugins loaded\" }\nelasticsearch | {\"type\": \"deprecation\", \"timestamp\": \"2026-04-20T06:39:12,054Z\", \"level\": \"DEPRECATION\", \"component\": \"o.e.d.c.s.Settings\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[node.data] setting was deprecated in Elasticsearch and will be removed in a future release! See the breaking changes documentation for the next major version.\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using [1] data paths, mounts [[/usr/share/elasticsearch/data (/dev/vda1)]], net usable_space [14.5gb], net total_space [58.3gb], types [ext4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"heap size [700mb], compressed ordinary object pointers [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,356Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"node name [e802ad473a4f], node ID [e2ZKzgw4Q4aCf2w5ljWr1A], cluster name [docker-cluster], roles [transform, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:18,805Z\", \"level\": \"INFO\", \"component\": \"o.e.x.m.p.l.CppLogMessageHandler\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[controller/213] [Main.cc@114] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,521Z\", \"level\": \"INFO\", \"component\": \"o.e.t.NettyAllocator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,695Z\", \"level\": \"INFO\", \"component\": \"o.e.d.DiscoveryModule\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using discovery type [single-node] and seed hosts providers [settings]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:22,462Z\", \"level\": \"WARN\", \"component\": \"o.e.g.DanglingIndicesState\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,302Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"initialized\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,303Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"starting ...\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,565Z\", \"level\": \"INFO\", \"component\": \"o.e.t.TransportService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9300}, bound_addresses {[::]:9300}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,081Z\", \"level\": \"INFO\", \"component\": \"o.e.c.c.Coordinator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,252Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.MasterService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,401Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.ClusterApplierService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.h.AbstractHttpServerTransport\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9200}, bound_addresses {[::]:9200}\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"started\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,162Z\", \"level\": \"INFO\", \"component\": \"o.e.l.LicenseService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,169Z\", \"level\": \"INFO\", \"component\": \"o.e.g.GatewayService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"recovered [15] indices into cluster_state\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:29,975Z\", \"level\": \"INFO\", \"component\": \"o.e.c.r.a.AllocationService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:02Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"auditTrail\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"visTypeXy\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"warning\",\"config\",\"deprecation\"],\"pid\":7,\"message\":\"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\\\"\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\ndocker_lamp_1 | + create_dot_env_local_file\ndocker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak\ndocker_lamp_1 | + create_dot_env\ndocker_lamp_1 | + [[ -f /home/jiminny/.env ]]\ndocker_lamp_1 | + return\ndocker_lamp_1 | + declare DB_ADMIN_PASSWORD\ndocker_lamp_1 | + declare DB_ADMIN_USERNAME\ndocker_lamp_1 | + declare DB_DEV_PASSWORD\ndocker_lamp_1 | + declare DB_DEV_USERNAME\ndocker_lamp_1 | + declare DB_ROOT_PASSWORD\ndocker_lamp_1 | + declare DB_ROOT_USERNAME\ndocker_lamp_1 | + declare DB_WEB_PASSWORD\ndocker_lamp_1 | + declare DB_WEB_USERNAME\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_PASSWORD='dgyt$rTe21-d'\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_USERNAME=root\ndocker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + readonly DB_ADMIN_PASSWORD\ndocker_lamp_1 | + readonly DB_ADMIN_USERNAME\ndocker_lamp_1 | + readonly DB_DEV_PASSWORD\ndocker_lamp_1 | + readonly DB_DEV_USERNAME\ndocker_lamp_1 | + readonly DB_ROOT_PASSWORD\ndocker_lamp_1 | + readonly DB_ROOT_USERNAME\ndocker_lamp_1 | + readonly DB_WEB_PASSWORD\ndocker_lamp_1 | + readonly DB_WEB_USERNAME\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=dgyt$rTe21-d~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.root\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml\"}\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Session cookies will be transmitted over insecure connections. This is not recommended.\"}\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local\ndocker_lamp_1 | + echo ''\ndocker_lamp_1 | + echo 'DB_ADMIN_PASSWORD=dgyt$rTe21-d'\ndocker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | + echo DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | + echo DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | + echo DB_ROOT_USERNAME=root\ndocker_lamp_1 | + echo DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + declare IP_ADDRESS\ndocker_lamp_1 | ++ curl ifconfig.co/\ndocker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current\ndocker_lamp_1 | Dload Upload Total Spent Left Speed\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"encryptedSavedObjects\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"ingestManager\"],\"pid\":7,\"message\":\"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\n100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224\ndocker_lamp_1 | + IP_ADDRESS=212.5.153.87\ndocker_lamp_1 | + readonly IP_ADDRESS\ndocker_lamp_1 | + sed -i -E 's~X-Forwarded-For \"\\$remote_addr\"~X-Forwarded-For \"IP_ADDRESS\"~g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + sed -i -E s~IP_ADDRESS~212.5.153.87~g /etc/nginx/nginx.conf\ndocker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \\{([^#]*)# QA Only End \\}//g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + storage_init\ndocker_lamp_1 | + false\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]\ndocker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1\ndocker_lamp_1 | ++ wc -l\ndocker_lamp_1 | + [[ 11 -lt 7 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]\ndocker_lamp_1 | + sleep 5\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Found 'server.host: \\\"0\\\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' is being automatically to the configuration. You can change the setting to 'server.host: 0.0.0.0' or add 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' in kibana.yml to prevent this message.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"actions\",\"actions\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"alerts\",\"plugins\",\"alerting\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\"],\"pid\":7,\"message\":\"config sourced from: production cluster\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Starting saved objects migrations\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\n\n\nv View in Docker Desktop o View Config w Enable Watch","is_focused":true},{"role":"AXButton","text":"Menu","depth":3,"bounds":{"left":0.48402777,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥1 DOCKER (docker-compose)","depth":3,"bounds":{"left":0.01875,"top":0.09,"width":0.46180555,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥2 PROD (ssh)","depth":4,"bounds":{"left":0.51805556,"top":0.09,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.24944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥3 EU (ssh)","depth":4,"bounds":{"left":0.51805556,"top":0.25,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.4422222,"width":0.50069445,"height":0.12222222},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.41944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥4 STAGE (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.42,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.5922222,"width":0.50069445,"height":0.12222222},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.5694444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥5 QA (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.57,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:48:07 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.74222225,"width":0.50069445,"height":0.12222222},"value":"Last login: Thu Apr 16 15:48:07 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.71944445,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥6 FE (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.72,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:48:07 on ttys005\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.87222224,"width":0.50069445,"height":0.12777776},"value":"Last login: Thu Apr 16 15:48:07 on ttys005\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.86944443,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥7 EXT (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.87,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.140625,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.14479166,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.28125,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.28541666,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.421875,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.42604166,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.5625,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.56666666,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"✳ Review screenpipe usage and Boosteroid integration (claude)","depth":2,"bounds":{"left":0.7027778,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.70694447,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.84305555,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.8472222,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.9548611,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"EU (ssh)","depth":1,"bounds":{"left":0.47916666,"top":0.033333335,"width":0.04097222,"height":0.017777778},"role_description":"text"}]...
|
3578503106199058064
|
-8629985393044693234
|
app_switch
|
accessibility
|
NULL
|
asticsearch","data"],"pid":7,& asticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:41Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:42Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:49Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_ROOT_USERNAME=root
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local
docker_lamp_1 | + declare IP_ADDRESS
docker_lamp_1 | ++ curl ifconfig.co/
docker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current
docker_lamp_1 | Dload Upload Total Spent Left Speed
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","encryptedSavedObjects","config"],"pid":7,"message":"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","ingestManager"],"pid":7,"message":"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224
docker_lamp_1 | + IP_ADDRESS=[IP_ADDRESS]
docker_lamp_1 | + readonly IP_ADDRESS
docker_lamp_1 | + sed -i -E 's~X-Forwarded-For "\$remote_addr"~X-Forwarded-For "IP_ADDRESS"~g' /etc/nginx/nginx.conf
docker_lamp_1 | + sed -i -E s~IP_ADDRESS~[IP_ADDRESS]~g /etc/nginx/nginx.conf
docker_lamp_...
|
NULL
|
|
52154
|
NULL
|
0
|
2026-04-20T06:42:02.840143+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667322840_m2.jpg...
|
iTerm2
|
EU (ssh)
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
asticsearch","data"],"pid":7,& asticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:41Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:42Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:49Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_ROOT_USERNAME=root
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local
docker_lamp_1 | + declare IP_ADDRESS
docker_lamp_1 | ++ curl ifconfig.co/
docker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current
docker_lamp_1 | Dload Upload Total Spent Left Speed
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","encryptedSavedObjects","config"],"pid":7,"message":"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","ingestManager"],"pid":7,"message":"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224
docker_lamp_1 | + IP_ADDRESS=[IP_ADDRESS]
docker_lamp_1 | + readonly IP_ADDRESS
docker_lamp_1 | + sed -i -E 's~X-Forwarded-For "\$remote_addr"~X-Forwarded-For "IP_ADDRESS"~g' /etc/nginx/nginx.conf
docker_lamp_1 | + sed -i -E s~IP_ADDRESS~[IP_ADDRESS]~g /etc/nginx/nginx.conf
docker_lamp_...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"asticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:42Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:45Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:48Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:49Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"reporting\",\"esqueue\",\"queue-worker\",\"error\"],\"pid\":7,\"message\":\"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\\n at process._tickCallback (internal/process/next_tick.js:61:11)\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:55Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:58Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:00Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:03Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:10Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work\nWARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion \n[+] Running 2/2\n ✔ Container redis Running 0.0s \n ✔ Container docker_lamp_1 Running 0.0s \nAttaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis\ndocker_lamp_1 | \ndocker_lamp_1 | Illuminate\\Database\\QueryException \ndocker_lamp_1 | \ndocker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)\ndocker_lamp_1 | \ndocker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838\ndocker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)\ndocker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class\ndocker_lamp_1 | 836▕ : QueryException::class;\ndocker_lamp_1 | 837▕ \ndocker_lamp_1 | ➜ 838▕ throw new $exceptionType(\ndocker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),\ndocker_lamp_1 | 840▕ $query,\ndocker_lamp_1 | 841▕ $this->prepareBindings($bindings),\ndocker_lamp_1 | 842▕ $e,\ndocker_lamp_1 | \ndocker_lamp_1 | +42 vendor frames \ndocker_lamp_1 | \ndocker_lamp_1 | 43 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \nmariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\nblackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.\nblackfire-1 | usage blackfire-agent [options]\nblackfire-1 | --collector=\"https://blackfire.io\": Sets the URL of Blackfire's data collector\nblackfire-1 | --config=\"/etc/blackfire/agent\": Sets the path to the configuration file\nblackfire-1 | -d: Prints the current configuration\nblackfire-1 | --http-proxy=\"\": Sets the HTTP proxy to use\nblackfire-1 | --https-proxy=\"\": Sets the HTTPS proxy to use\nblackfire-1 | --log-file=\"stderr\": Sets the path of the log file. Use stderr to log to stderr\nblackfire-1 | --log-level=\"1\": log verbosity level (4: debug, 3: info, 2: warning, 1: error)\nblackfire-1 | --register: Helps you with registering the agent\nblackfire-1 | --server-id=\"\": Sets the server id used to authenticate with Blackfire API\nblackfire-1 | --server-token=\"\": Sets the server token used to authenticate with Blackfire API. It is unsafe to set this from the command line\nblackfire-1 | --socket=\"unix:///var/run/blackfire/agent.sock\": Sets the socket the agent should read traces from. Possible value can be a unix socket or a TCP address. ie: unix:///var/run/blackfire/agent.sock or tcp://127.0.0.1:8307\nblackfire-1 | --test: Tests the configuration\nblackfire-1 | --timeout=\"15s\": Sets the Blackfire connection timeout\nblackfire-1 | -v: Prints the version number\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"no configuration paths supplied\"\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"using configuration at default config path\" path=/home/ngrok/.ngrok2/ngrok.yml\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"open config file\" path=/home/ngrok/.ngrok2/ngrok.yml err=nil\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"starting web service\" obj=web addr=0.0.0.0:4040\njiminny_ext-1 exited with code 0\nblackfire-1 exited with code 1\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"tunnel session started\" obj=tunnels.session\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"client session established\" obj=csess id=7d0b19e3215d\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=\"command_line (http)\" addr=http://lamp:3080 url=http://lukask.ngrok.io\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=command_line addr=http://lamp:3080 url=https://lukask.ngrok.io\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Warn] [Entrypoint]: /sys/fs/cgroup///memory.pressure not writable, functionality unavailable to MariaDB\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\ndatadog-1 | [s6-init] making user provided files available at /var/run/s6/etc...exited 0.\nngrok | t=2026-04-20T06:38:55+0000 lvl=info msg=\"update available\" obj=updater\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade information missing, assuming required\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade (mariadb-upgrade or creating healthcheck users) required, but skipped due to $MARIADB_AUTO_UPGRADE setting\nmariadb-1 | 2026-04-20 6:38:55 0 [Note] Starting MariaDB 11.4.5-MariaDB-ubu2404 source revision 0771110266ff5c04216af4bf1243c65f8c67ccf4 server_uid geAM8SSFn8P6Gz+0WMyyW27EUEM= as process 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Compressed tables use zlib 1.3\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Number of transaction pools: 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using ARMv8 crc32 + pmull instructions\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using liburing\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Completed initialization of buffer pool\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes)\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: End of log at LSN=7614257762\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Opened 3 undo tablespaces\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: 128 rollback segments in 3 undo tablespaces are active.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ...\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: log sequence number 7614257762; transaction id 11481082\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'FEEDBACK' is disabled.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'wsrep-provider' is disabled.\nmariadb-1 | 2026-04-20 6:38:58 0 [Note] InnoDB: Buffer pool(s) load completed at 260420 6:38:58\ndatadog-1 | [s6-init] ensuring user provided files have correct perms...exited 0.\ndatadog-1 | [fix-attrs.d] applying ownership & permissions fixes...\ndatadog-1 | [fix-attrs.d] done.\ndatadog-1 | [cont-init.d] executing container initialization scripts...\ndatadog-1 | [cont-init.d] 01-check-apikey.sh: executing... \ndatadog-1 | \ndatadog-1 | ==================================================================================\ndatadog-1 | You must set an DD_API_KEY environment variable to run the Datadog Agent container\ndatadog-1 | ==================================================================================\ndatadog-1 | \ndatadog-1 | [cont-init.d] 01-check-apikey.sh: exited 1.\ndatadog-1 exited with code 1\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '0.0.0.0'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '::'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: Event Scheduler: Loaded 0 events\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: ready for connections.\nmariadb-1 | Version: '11.4.5-MariaDB-ubu2404' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution\ndocker_lamp_1 | \ndocker_lamp_1 | INFO Configuration cached successfully. \ndocker_lamp_1 | \ndocker_lamp_1 | cp: cannot stat '/home/jiminny/bootstrap/cache/config-new.php': No such file or directory\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + [[ false == \\f\\a\\l\\s\\e ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + [[ -f /home/jiminny/storage/oauth-private.key ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + setup_local_environment\ndocker_lamp_1 | + storage_permissions_workaround\ndocker_lamp_1 | + [[ false != \\t\\r\\u\\e ]]\ndocker_lamp_1 | + ln -sf /home/jiminny/storage /home/jiminny/jiminny_storage\ndocker_lamp_1 | + return\ndocker_lamp_1 | + chmod 660 /home/jiminny/storage/oauth-private.key\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/views\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/cache\ndocker_lamp_1 | + chmod 775 /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/cache\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,413Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"version[7.10.2], pid[7], build[default/docker/747e1cc71def077253878a59143c1f785afa92b9/2021-01-13T04:42:47.157277Z], OS[Linux/6.12.54-linuxkit/aarch64], JVM[AdoptOpenJDK/OpenJDK 64-Bit Server VM/15.0.1/15.0.1+9]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM home [/usr/share/elasticsearch/jdk], using bundled JDK [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Djava.locale.providers=SPI,COMPAT, -Xms1g, -Xmx1g, -XX:+UseG1GC, -XX:G1ReservePercent=25, -XX:InitiatingHeapOccupancyPercent=30, -Djava.io.tmpdir=/tmp/elasticsearch-11712280666606670653, -XX:+HeapDumpOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -Des.cgroups.hierarchy.override=/, -Xms700m, -Xmx700m, -XX:MaxDirectMemorySize=367001600, -Des.path.home=/usr/share/elasticsearch, -Des.path.conf=/usr/share/elasticsearch/config, -Des.distribution.flavor=default, -Des.distribution.type=docker, -Des.bundled_jdk=true]\" }\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chmod g+w '{}' ';'\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,926Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [aggs-matrix-stats]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [analysis-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [constant-keyword]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [flattened]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [frozen-indices]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-geoip]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-user-agent]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [kibana]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-expression]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-mustache]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-painless]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-extras]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-version]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [parent-join]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [percolator]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [rank-eval]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [reindex]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repositories-metering-api]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repository-url]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [search-business-rules]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [searchable-snapshots]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [spatial]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transform]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transport-netty4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [unsigned-long]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [vectors]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [wildcard]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-analytics]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async-search]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-autoscaling]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ccr]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-core]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-data-streams]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-deprecation]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-enrich]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-eql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-graph]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-identity-provider]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,936Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ilm]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-logstash]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ml]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-monitoring]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-rollup]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-security]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-sql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-stack]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-voting-only-node]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,938Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-watcher]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,943Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"no plugins loaded\" }\nelasticsearch | {\"type\": \"deprecation\", \"timestamp\": \"2026-04-20T06:39:12,054Z\", \"level\": \"DEPRECATION\", \"component\": \"o.e.d.c.s.Settings\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[node.data] setting was deprecated in Elasticsearch and will be removed in a future release! See the breaking changes documentation for the next major version.\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using [1] data paths, mounts [[/usr/share/elasticsearch/data (/dev/vda1)]], net usable_space [14.5gb], net total_space [58.3gb], types [ext4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"heap size [700mb], compressed ordinary object pointers [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,356Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"node name [e802ad473a4f], node ID [e2ZKzgw4Q4aCf2w5ljWr1A], cluster name [docker-cluster], roles [transform, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:18,805Z\", \"level\": \"INFO\", \"component\": \"o.e.x.m.p.l.CppLogMessageHandler\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[controller/213] [Main.cc@114] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,521Z\", \"level\": \"INFO\", \"component\": \"o.e.t.NettyAllocator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,695Z\", \"level\": \"INFO\", \"component\": \"o.e.d.DiscoveryModule\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using discovery type [single-node] and seed hosts providers [settings]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:22,462Z\", \"level\": \"WARN\", \"component\": \"o.e.g.DanglingIndicesState\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,302Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"initialized\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,303Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"starting ...\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,565Z\", \"level\": \"INFO\", \"component\": \"o.e.t.TransportService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9300}, bound_addresses {[::]:9300}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,081Z\", \"level\": \"INFO\", \"component\": \"o.e.c.c.Coordinator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,252Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.MasterService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,401Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.ClusterApplierService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.h.AbstractHttpServerTransport\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9200}, bound_addresses {[::]:9200}\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"started\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,162Z\", \"level\": \"INFO\", \"component\": \"o.e.l.LicenseService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,169Z\", \"level\": \"INFO\", \"component\": \"o.e.g.GatewayService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"recovered [15] indices into cluster_state\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:29,975Z\", \"level\": \"INFO\", \"component\": \"o.e.c.r.a.AllocationService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:02Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"auditTrail\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"visTypeXy\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"warning\",\"config\",\"deprecation\"],\"pid\":7,\"message\":\"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\\\"\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\ndocker_lamp_1 | + create_dot_env_local_file\ndocker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak\ndocker_lamp_1 | + create_dot_env\ndocker_lamp_1 | + [[ -f /home/jiminny/.env ]]\ndocker_lamp_1 | + return\ndocker_lamp_1 | + declare DB_ADMIN_PASSWORD\ndocker_lamp_1 | + declare DB_ADMIN_USERNAME\ndocker_lamp_1 | + declare DB_DEV_PASSWORD\ndocker_lamp_1 | + declare DB_DEV_USERNAME\ndocker_lamp_1 | + declare DB_ROOT_PASSWORD\ndocker_lamp_1 | + declare DB_ROOT_USERNAME\ndocker_lamp_1 | + declare DB_WEB_PASSWORD\ndocker_lamp_1 | + declare DB_WEB_USERNAME\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_PASSWORD='dgyt$rTe21-d'\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_USERNAME=root\ndocker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + readonly DB_ADMIN_PASSWORD\ndocker_lamp_1 | + readonly DB_ADMIN_USERNAME\ndocker_lamp_1 | + readonly DB_DEV_PASSWORD\ndocker_lamp_1 | + readonly DB_DEV_USERNAME\ndocker_lamp_1 | + readonly DB_ROOT_PASSWORD\ndocker_lamp_1 | + readonly DB_ROOT_USERNAME\ndocker_lamp_1 | + readonly DB_WEB_PASSWORD\ndocker_lamp_1 | + readonly DB_WEB_USERNAME\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=dgyt$rTe21-d~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.root\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml\"}\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Session cookies will be transmitted over insecure connections. This is not recommended.\"}\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local\ndocker_lamp_1 | + echo ''\ndocker_lamp_1 | + echo 'DB_ADMIN_PASSWORD=dgyt$rTe21-d'\ndocker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | + echo DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | + echo DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | + echo DB_ROOT_USERNAME=root\ndocker_lamp_1 | + echo DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + declare IP_ADDRESS\ndocker_lamp_1 | ++ curl ifconfig.co/\ndocker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current\ndocker_lamp_1 | Dload Upload Total Spent Left Speed\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"encryptedSavedObjects\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"ingestManager\"],\"pid\":7,\"message\":\"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\n100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224\ndocker_lamp_1 | + IP_ADDRESS=212.5.153.87\ndocker_lamp_1 | + readonly IP_ADDRESS\ndocker_lamp_1 | + sed -i -E 's~X-Forwarded-For \"\\$remote_addr\"~X-Forwarded-For \"IP_ADDRESS\"~g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + sed -i -E s~IP_ADDRESS~212.5.153.87~g /etc/nginx/nginx.conf\ndocker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \\{([^#]*)# QA Only End \\}//g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + storage_init\ndocker_lamp_1 | + false\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]\ndocker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1\ndocker_lamp_1 | ++ wc -l\ndocker_lamp_1 | + [[ 11 -lt 7 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]\ndocker_lamp_1 | + sleep 5\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Found 'server.host: \\\"0\\\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' is being automatically to the configuration. You can change the setting to 'server.host: 0.0.0.0' or add 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' in kibana.yml to prevent this message.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"actions\",\"actions\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"alerts\",\"plugins\",\"alerting\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\"],\"pid\":7,\"message\":\"config sourced from: production cluster\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Starting saved objects migrations\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\n\n\nv View in Docker Desktop o View Config w Enable Watch","depth":4,"value":"asticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:41Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:42Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:45Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:48Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:49Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"reporting\",\"esqueue\",\"queue-worker\",\"error\"],\"pid\":7,\"message\":\"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\\n at process._tickCallback (internal/process/next_tick.js:61:11)\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:55Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:58Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:00Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:03Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:10Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work\nWARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion \n[+] Running 2/2\n ✔ Container redis Running 0.0s \n ✔ Container docker_lamp_1 Running 0.0s \nAttaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis\ndocker_lamp_1 | \ndocker_lamp_1 | Illuminate\\Database\\QueryException \ndocker_lamp_1 | \ndocker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)\ndocker_lamp_1 | \ndocker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838\ndocker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)\ndocker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class\ndocker_lamp_1 | 836▕ : QueryException::class;\ndocker_lamp_1 | 837▕ \ndocker_lamp_1 | ➜ 838▕ throw new $exceptionType(\ndocker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),\ndocker_lamp_1 | 840▕ $query,\ndocker_lamp_1 | 841▕ $this->prepareBindings($bindings),\ndocker_lamp_1 | 842▕ $e,\ndocker_lamp_1 | \ndocker_lamp_1 | +42 vendor frames \ndocker_lamp_1 | \ndocker_lamp_1 | 43 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \nmariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\nblackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.\nblackfire-1 | usage blackfire-agent [options]\nblackfire-1 | --collector=\"https://blackfire.io\": Sets the URL of Blackfire's data collector\nblackfire-1 | --config=\"/etc/blackfire/agent\": Sets the path to the configuration file\nblackfire-1 | -d: Prints the current configuration\nblackfire-1 | --http-proxy=\"\": Sets the HTTP proxy to use\nblackfire-1 | --https-proxy=\"\": Sets the HTTPS proxy to use\nblackfire-1 | --log-file=\"stderr\": Sets the path of the log file. Use stderr to log to stderr\nblackfire-1 | --log-level=\"1\": log verbosity level (4: debug, 3: info, 2: warning, 1: error)\nblackfire-1 | --register: Helps you with registering the agent\nblackfire-1 | --server-id=\"\": Sets the server id used to authenticate with Blackfire API\nblackfire-1 | --server-token=\"\": Sets the server token used to authenticate with Blackfire API. It is unsafe to set this from the command line\nblackfire-1 | --socket=\"unix:///var/run/blackfire/agent.sock\": Sets the socket the agent should read traces from. Possible value can be a unix socket or a TCP address. ie: unix:///var/run/blackfire/agent.sock or tcp://127.0.0.1:8307\nblackfire-1 | --test: Tests the configuration\nblackfire-1 | --timeout=\"15s\": Sets the Blackfire connection timeout\nblackfire-1 | -v: Prints the version number\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"no configuration paths supplied\"\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"using configuration at default config path\" path=/home/ngrok/.ngrok2/ngrok.yml\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"open config file\" path=/home/ngrok/.ngrok2/ngrok.yml err=nil\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"starting web service\" obj=web addr=0.0.0.0:4040\njiminny_ext-1 exited with code 0\nblackfire-1 exited with code 1\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"tunnel session started\" obj=tunnels.session\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"client session established\" obj=csess id=7d0b19e3215d\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=\"command_line (http)\" addr=http://lamp:3080 url=http://lukask.ngrok.io\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=command_line addr=http://lamp:3080 url=https://lukask.ngrok.io\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Warn] [Entrypoint]: /sys/fs/cgroup///memory.pressure not writable, functionality unavailable to MariaDB\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\ndatadog-1 | [s6-init] making user provided files available at /var/run/s6/etc...exited 0.\nngrok | t=2026-04-20T06:38:55+0000 lvl=info msg=\"update available\" obj=updater\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade information missing, assuming required\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade (mariadb-upgrade or creating healthcheck users) required, but skipped due to $MARIADB_AUTO_UPGRADE setting\nmariadb-1 | 2026-04-20 6:38:55 0 [Note] Starting MariaDB 11.4.5-MariaDB-ubu2404 source revision 0771110266ff5c04216af4bf1243c65f8c67ccf4 server_uid geAM8SSFn8P6Gz+0WMyyW27EUEM= as process 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Compressed tables use zlib 1.3\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Number of transaction pools: 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using ARMv8 crc32 + pmull instructions\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using liburing\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Completed initialization of buffer pool\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes)\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: End of log at LSN=7614257762\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Opened 3 undo tablespaces\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: 128 rollback segments in 3 undo tablespaces are active.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ...\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: log sequence number 7614257762; transaction id 11481082\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'FEEDBACK' is disabled.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'wsrep-provider' is disabled.\nmariadb-1 | 2026-04-20 6:38:58 0 [Note] InnoDB: Buffer pool(s) load completed at 260420 6:38:58\ndatadog-1 | [s6-init] ensuring user provided files have correct perms...exited 0.\ndatadog-1 | [fix-attrs.d] applying ownership & permissions fixes...\ndatadog-1 | [fix-attrs.d] done.\ndatadog-1 | [cont-init.d] executing container initialization scripts...\ndatadog-1 | [cont-init.d] 01-check-apikey.sh: executing... \ndatadog-1 | \ndatadog-1 | ==================================================================================\ndatadog-1 | You must set an DD_API_KEY environment variable to run the Datadog Agent container\ndatadog-1 | ==================================================================================\ndatadog-1 | \ndatadog-1 | [cont-init.d] 01-check-apikey.sh: exited 1.\ndatadog-1 exited with code 1\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '0.0.0.0'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '::'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: Event Scheduler: Loaded 0 events\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: ready for connections.\nmariadb-1 | Version: '11.4.5-MariaDB-ubu2404' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution\ndocker_lamp_1 | \ndocker_lamp_1 | INFO Configuration cached successfully. \ndocker_lamp_1 | \ndocker_lamp_1 | cp: cannot stat '/home/jiminny/bootstrap/cache/config-new.php': No such file or directory\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + [[ false == \\f\\a\\l\\s\\e ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + [[ -f /home/jiminny/storage/oauth-private.key ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + setup_local_environment\ndocker_lamp_1 | + storage_permissions_workaround\ndocker_lamp_1 | + [[ false != \\t\\r\\u\\e ]]\ndocker_lamp_1 | + ln -sf /home/jiminny/storage /home/jiminny/jiminny_storage\ndocker_lamp_1 | + return\ndocker_lamp_1 | + chmod 660 /home/jiminny/storage/oauth-private.key\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/views\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/cache\ndocker_lamp_1 | + chmod 775 /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/cache\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,413Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"version[7.10.2], pid[7], build[default/docker/747e1cc71def077253878a59143c1f785afa92b9/2021-01-13T04:42:47.157277Z], OS[Linux/6.12.54-linuxkit/aarch64], JVM[AdoptOpenJDK/OpenJDK 64-Bit Server VM/15.0.1/15.0.1+9]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM home [/usr/share/elasticsearch/jdk], using bundled JDK [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Djava.locale.providers=SPI,COMPAT, -Xms1g, -Xmx1g, -XX:+UseG1GC, -XX:G1ReservePercent=25, -XX:InitiatingHeapOccupancyPercent=30, -Djava.io.tmpdir=/tmp/elasticsearch-11712280666606670653, -XX:+HeapDumpOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -Des.cgroups.hierarchy.override=/, -Xms700m, -Xmx700m, -XX:MaxDirectMemorySize=367001600, -Des.path.home=/usr/share/elasticsearch, -Des.path.conf=/usr/share/elasticsearch/config, -Des.distribution.flavor=default, -Des.distribution.type=docker, -Des.bundled_jdk=true]\" }\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chmod g+w '{}' ';'\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,926Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [aggs-matrix-stats]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [analysis-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [constant-keyword]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [flattened]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [frozen-indices]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-geoip]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-user-agent]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [kibana]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-expression]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-mustache]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-painless]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-extras]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-version]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [parent-join]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [percolator]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [rank-eval]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [reindex]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repositories-metering-api]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repository-url]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [search-business-rules]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [searchable-snapshots]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [spatial]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transform]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transport-netty4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [unsigned-long]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [vectors]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [wildcard]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-analytics]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async-search]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-autoscaling]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ccr]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-core]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-data-streams]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-deprecation]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-enrich]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-eql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-graph]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-identity-provider]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,936Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ilm]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-logstash]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ml]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-monitoring]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-rollup]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-security]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-sql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-stack]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-voting-only-node]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,938Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-watcher]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,943Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"no plugins loaded\" }\nelasticsearch | {\"type\": \"deprecation\", \"timestamp\": \"2026-04-20T06:39:12,054Z\", \"level\": \"DEPRECATION\", \"component\": \"o.e.d.c.s.Settings\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[node.data] setting was deprecated in Elasticsearch and will be removed in a future release! See the breaking changes documentation for the next major version.\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using [1] data paths, mounts [[/usr/share/elasticsearch/data (/dev/vda1)]], net usable_space [14.5gb], net total_space [58.3gb], types [ext4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"heap size [700mb], compressed ordinary object pointers [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,356Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"node name [e802ad473a4f], node ID [e2ZKzgw4Q4aCf2w5ljWr1A], cluster name [docker-cluster], roles [transform, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:18,805Z\", \"level\": \"INFO\", \"component\": \"o.e.x.m.p.l.CppLogMessageHandler\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[controller/213] [Main.cc@114] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,521Z\", \"level\": \"INFO\", \"component\": \"o.e.t.NettyAllocator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,695Z\", \"level\": \"INFO\", \"component\": \"o.e.d.DiscoveryModule\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using discovery type [single-node] and seed hosts providers [settings]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:22,462Z\", \"level\": \"WARN\", \"component\": \"o.e.g.DanglingIndicesState\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,302Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"initialized\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,303Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"starting ...\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,565Z\", \"level\": \"INFO\", \"component\": \"o.e.t.TransportService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9300}, bound_addresses {[::]:9300}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,081Z\", \"level\": \"INFO\", \"component\": \"o.e.c.c.Coordinator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,252Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.MasterService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,401Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.ClusterApplierService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.h.AbstractHttpServerTransport\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9200}, bound_addresses {[::]:9200}\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"started\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,162Z\", \"level\": \"INFO\", \"component\": \"o.e.l.LicenseService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,169Z\", \"level\": \"INFO\", \"component\": \"o.e.g.GatewayService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"recovered [15] indices into cluster_state\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:29,975Z\", \"level\": \"INFO\", \"component\": \"o.e.c.r.a.AllocationService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:02Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"auditTrail\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"visTypeXy\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"warning\",\"config\",\"deprecation\"],\"pid\":7,\"message\":\"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\\\"\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\ndocker_lamp_1 | + create_dot_env_local_file\ndocker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak\ndocker_lamp_1 | + create_dot_env\ndocker_lamp_1 | + [[ -f /home/jiminny/.env ]]\ndocker_lamp_1 | + return\ndocker_lamp_1 | + declare DB_ADMIN_PASSWORD\ndocker_lamp_1 | + declare DB_ADMIN_USERNAME\ndocker_lamp_1 | + declare DB_DEV_PASSWORD\ndocker_lamp_1 | + declare DB_DEV_USERNAME\ndocker_lamp_1 | + declare DB_ROOT_PASSWORD\ndocker_lamp_1 | + declare DB_ROOT_USERNAME\ndocker_lamp_1 | + declare DB_WEB_PASSWORD\ndocker_lamp_1 | + declare DB_WEB_USERNAME\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_PASSWORD='dgyt$rTe21-d'\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_USERNAME=root\ndocker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + readonly DB_ADMIN_PASSWORD\ndocker_lamp_1 | + readonly DB_ADMIN_USERNAME\ndocker_lamp_1 | + readonly DB_DEV_PASSWORD\ndocker_lamp_1 | + readonly DB_DEV_USERNAME\ndocker_lamp_1 | + readonly DB_ROOT_PASSWORD\ndocker_lamp_1 | + readonly DB_ROOT_USERNAME\ndocker_lamp_1 | + readonly DB_WEB_PASSWORD\ndocker_lamp_1 | + readonly DB_WEB_USERNAME\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=dgyt$rTe21-d~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.root\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml\"}\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Session cookies will be transmitted over insecure connections. This is not recommended.\"}\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local\ndocker_lamp_1 | + echo ''\ndocker_lamp_1 | + echo 'DB_ADMIN_PASSWORD=dgyt$rTe21-d'\ndocker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | + echo DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | + echo DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | + echo DB_ROOT_USERNAME=root\ndocker_lamp_1 | + echo DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + declare IP_ADDRESS\ndocker_lamp_1 | ++ curl ifconfig.co/\ndocker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current\ndocker_lamp_1 | Dload Upload Total Spent Left Speed\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"encryptedSavedObjects\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"ingestManager\"],\"pid\":7,\"message\":\"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\n100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224\ndocker_lamp_1 | + IP_ADDRESS=212.5.153.87\ndocker_lamp_1 | + readonly IP_ADDRESS\ndocker_lamp_1 | + sed -i -E 's~X-Forwarded-For \"\\$remote_addr\"~X-Forwarded-For \"IP_ADDRESS\"~g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + sed -i -E s~IP_ADDRESS~212.5.153.87~g /etc/nginx/nginx.conf\ndocker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \\{([^#]*)# QA Only End \\}//g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + storage_init\ndocker_lamp_1 | + false\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]\ndocker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1\ndocker_lamp_1 | ++ wc -l\ndocker_lamp_1 | + [[ 11 -lt 7 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]\ndocker_lamp_1 | + sleep 5\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Found 'server.host: \\\"0\\\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' is being automatically to the configuration. You can change the setting to 'server.host: 0.0.0.0' or add 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' in kibana.yml to prevent this message.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"actions\",\"actions\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"alerts\",\"plugins\",\"alerting\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\"],\"pid\":7,\"message\":\"config sourced from: production cluster\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Starting saved objects migrations\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\n\n\nv View in Docker Desktop o View Config w Enable Watch","is_focused":true},{"role":"AXButton","text":"Menu","depth":3,"bounds":{"left":0.50199467,"top":1.0,"width":0.004986702,"height":-0.06424582},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥1 DOCKER (docker-compose)","depth":3,"bounds":{"left":0.27925533,"top":1.0,"width":0.22107713,"height":-0.06464481},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","depth":5,"bounds":{"left":0.5093085,"top":0.11811652,"width":0.23969415,"height":0.8818835},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.74202126,"top":1.0,"width":0.004986702,"height":-0.06424582},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥2 PROD (ssh)","depth":4,"bounds":{"left":0.51828456,"top":1.0,"width":0.22207446,"height":-0.06464481},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","depth":5,"bounds":{"left":0.5093085,"top":0.21867518,"width":0.23969415,"height":0.7813248},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥3 EU (ssh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥4 STAGE (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥5 QA (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:48:07 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Thu Apr 16 15:48:07 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥6 FE (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:48:07 on ttys005\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Thu Apr 16 15:48:07 on ttys005\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥7 EXT (-zsh)","depth":4,"role_description":"text"},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.27027926,"top":1.0,"width":0.06732048,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.27227393,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.33759972,"top":1.0,"width":0.06732048,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.33959442,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.40492022,"top":1.0,"width":0.06732048,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.4069149,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.4722407,"top":1.0,"width":0.06732048,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.4742354,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.53956115,"top":1.0,"width":0.06715426,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.5415558,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"✳ Review screenpipe usage and Boosteroid integration (claude)","depth":2,"bounds":{"left":0.60671544,"top":1.0,"width":0.06715426,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.6087101,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.67386967,"top":1.0,"width":0.06715426,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.67586434,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.7273936,"top":1.0,"width":0.01861702,"height":-0.023144484},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"EU (ssh)","depth":1,"bounds":{"left":0.49966756,"top":1.0,"width":0.019614361,"height":-0.02394259},"role_description":"text"}]...
|
3578503106199058064
|
-8629985393044693234
|
app_switch
|
accessibility
|
NULL
|
asticsearch","data"],"pid":7,& asticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:41Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:42Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:49Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_ROOT_USERNAME=root
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local
docker_lamp_1 | + declare IP_ADDRESS
docker_lamp_1 | ++ curl ifconfig.co/
docker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current
docker_lamp_1 | Dload Upload Total Spent Left Speed
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","encryptedSavedObjects","config"],"pid":7,"message":"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","ingestManager"],"pid":7,"message":"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224
docker_lamp_1 | + IP_ADDRESS=[IP_ADDRESS]
docker_lamp_1 | + readonly IP_ADDRESS
docker_lamp_1 | + sed -i -E 's~X-Forwarded-For "\$remote_addr"~X-Forwarded-For "IP_ADDRESS"~g' /etc/nginx/nginx.conf
docker_lamp_1 | + sed -i -E s~IP_ADDRESS~[IP_ADDRESS]~g /etc/nginx/nginx.conf
docker_lamp_...
|
NULL
|
|
52156
|
1128
|
0
|
2026-04-20T06:42:06.354792+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667326354_m1.jpg...
|
iTerm2
|
EU (ssh)
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
ins","taskManager","taskManager ins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:42Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:49Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_ROOT_USERNAME=root
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local
docker_lamp_1 | + declare IP_ADDRESS
docker_lamp_1 | ++ curl ifconfig.co/
docker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current
docker_lamp_1 | Dload Upload Total Spent Left Speed
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","encryptedSavedObjects","config"],"pid":7,"message":"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","ingestManager"],"pid":7,"message":"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224
docker_lamp_1 | + IP_ADDRESS=[IP_ADDRESS]
docker_lamp_1 | + readonly IP_ADDRESS
docker_lamp_1 | + sed -i -E 's~X-Forwarded-For "\$remote_addr"~X-Forwarded-For "IP_ADDRESS"~g' /etc/nginx/nginx.conf
docker_lamp_1 | + sed -i -E s~IP_ADDRESS~[IP_ADDRESS]~g /etc/nginx/nginx.conf
docker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \{([^#]*)# QA Only End \}//g' /etc/nginx/nginx.conf
docker_lamp_1 | + storage_init
docker_lamp_1 | + false
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"ins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:42Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:45Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:48Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:49Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"reporting\",\"esqueue\",\"queue-worker\",\"error\"],\"pid\":7,\"message\":\"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\\n at process._tickCallback (internal/process/next_tick.js:61:11)\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:55Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:58Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:00Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:03Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:10Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work\nWARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion \n[+] Running 2/2\n ✔ Container redis Running 0.0s \n ✔ Container docker_lamp_1 Running 0.0s \nAttaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis\ndocker_lamp_1 | \ndocker_lamp_1 | Illuminate\\Database\\QueryException \ndocker_lamp_1 | \ndocker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)\ndocker_lamp_1 | \ndocker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838\ndocker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)\ndocker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class\ndocker_lamp_1 | 836▕ : QueryException::class;\ndocker_lamp_1 | 837▕ \ndocker_lamp_1 | ➜ 838▕ throw new $exceptionType(\ndocker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),\ndocker_lamp_1 | 840▕ $query,\ndocker_lamp_1 | 841▕ $this->prepareBindings($bindings),\ndocker_lamp_1 | 842▕ $e,\ndocker_lamp_1 | \ndocker_lamp_1 | +42 vendor frames \ndocker_lamp_1 | \ndocker_lamp_1 | 43 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \nmariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\nblackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.\nblackfire-1 | usage blackfire-agent [options]\nblackfire-1 | --collector=\"https://blackfire.io\": Sets the URL of Blackfire's data collector\nblackfire-1 | --config=\"/etc/blackfire/agent\": Sets the path to the configuration file\nblackfire-1 | -d: Prints the current configuration\nblackfire-1 | --http-proxy=\"\": Sets the HTTP proxy to use\nblackfire-1 | --https-proxy=\"\": Sets the HTTPS proxy to use\nblackfire-1 | --log-file=\"stderr\": Sets the path of the log file. Use stderr to log to stderr\nblackfire-1 | --log-level=\"1\": log verbosity level (4: debug, 3: info, 2: warning, 1: error)\nblackfire-1 | --register: Helps you with registering the agent\nblackfire-1 | --server-id=\"\": Sets the server id used to authenticate with Blackfire API\nblackfire-1 | --server-token=\"\": Sets the server token used to authenticate with Blackfire API. It is unsafe to set this from the command line\nblackfire-1 | --socket=\"unix:///var/run/blackfire/agent.sock\": Sets the socket the agent should read traces from. Possible value can be a unix socket or a TCP address. ie: unix:///var/run/blackfire/agent.sock or tcp://127.0.0.1:8307\nblackfire-1 | --test: Tests the configuration\nblackfire-1 | --timeout=\"15s\": Sets the Blackfire connection timeout\nblackfire-1 | -v: Prints the version number\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"no configuration paths supplied\"\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"using configuration at default config path\" path=/home/ngrok/.ngrok2/ngrok.yml\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"open config file\" path=/home/ngrok/.ngrok2/ngrok.yml err=nil\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"starting web service\" obj=web addr=0.0.0.0:4040\njiminny_ext-1 exited with code 0\nblackfire-1 exited with code 1\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"tunnel session started\" obj=tunnels.session\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"client session established\" obj=csess id=7d0b19e3215d\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=\"command_line (http)\" addr=http://lamp:3080 url=http://lukask.ngrok.io\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=command_line addr=http://lamp:3080 url=https://lukask.ngrok.io\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Warn] [Entrypoint]: /sys/fs/cgroup///memory.pressure not writable, functionality unavailable to MariaDB\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\ndatadog-1 | [s6-init] making user provided files available at /var/run/s6/etc...exited 0.\nngrok | t=2026-04-20T06:38:55+0000 lvl=info msg=\"update available\" obj=updater\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade information missing, assuming required\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade (mariadb-upgrade or creating healthcheck users) required, but skipped due to $MARIADB_AUTO_UPGRADE setting\nmariadb-1 | 2026-04-20 6:38:55 0 [Note] Starting MariaDB 11.4.5-MariaDB-ubu2404 source revision 0771110266ff5c04216af4bf1243c65f8c67ccf4 server_uid geAM8SSFn8P6Gz+0WMyyW27EUEM= as process 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Compressed tables use zlib 1.3\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Number of transaction pools: 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using ARMv8 crc32 + pmull instructions\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using liburing\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Completed initialization of buffer pool\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes)\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: End of log at LSN=7614257762\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Opened 3 undo tablespaces\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: 128 rollback segments in 3 undo tablespaces are active.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ...\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: log sequence number 7614257762; transaction id 11481082\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'FEEDBACK' is disabled.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'wsrep-provider' is disabled.\nmariadb-1 | 2026-04-20 6:38:58 0 [Note] InnoDB: Buffer pool(s) load completed at 260420 6:38:58\ndatadog-1 | [s6-init] ensuring user provided files have correct perms...exited 0.\ndatadog-1 | [fix-attrs.d] applying ownership & permissions fixes...\ndatadog-1 | [fix-attrs.d] done.\ndatadog-1 | [cont-init.d] executing container initialization scripts...\ndatadog-1 | [cont-init.d] 01-check-apikey.sh: executing... \ndatadog-1 | \ndatadog-1 | ==================================================================================\ndatadog-1 | You must set an DD_API_KEY environment variable to run the Datadog Agent container\ndatadog-1 | ==================================================================================\ndatadog-1 | \ndatadog-1 | [cont-init.d] 01-check-apikey.sh: exited 1.\ndatadog-1 exited with code 1\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '0.0.0.0'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '::'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: Event Scheduler: Loaded 0 events\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: ready for connections.\nmariadb-1 | Version: '11.4.5-MariaDB-ubu2404' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution\ndocker_lamp_1 | \ndocker_lamp_1 | INFO Configuration cached successfully. \ndocker_lamp_1 | \ndocker_lamp_1 | cp: cannot stat '/home/jiminny/bootstrap/cache/config-new.php': No such file or directory\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + [[ false == \\f\\a\\l\\s\\e ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + [[ -f /home/jiminny/storage/oauth-private.key ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + setup_local_environment\ndocker_lamp_1 | + storage_permissions_workaround\ndocker_lamp_1 | + [[ false != \\t\\r\\u\\e ]]\ndocker_lamp_1 | + ln -sf /home/jiminny/storage /home/jiminny/jiminny_storage\ndocker_lamp_1 | + return\ndocker_lamp_1 | + chmod 660 /home/jiminny/storage/oauth-private.key\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/views\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/cache\ndocker_lamp_1 | + chmod 775 /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/cache\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,413Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"version[7.10.2], pid[7], build[default/docker/747e1cc71def077253878a59143c1f785afa92b9/2021-01-13T04:42:47.157277Z], OS[Linux/6.12.54-linuxkit/aarch64], JVM[AdoptOpenJDK/OpenJDK 64-Bit Server VM/15.0.1/15.0.1+9]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM home [/usr/share/elasticsearch/jdk], using bundled JDK [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Djava.locale.providers=SPI,COMPAT, -Xms1g, -Xmx1g, -XX:+UseG1GC, -XX:G1ReservePercent=25, -XX:InitiatingHeapOccupancyPercent=30, -Djava.io.tmpdir=/tmp/elasticsearch-11712280666606670653, -XX:+HeapDumpOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -Des.cgroups.hierarchy.override=/, -Xms700m, -Xmx700m, -XX:MaxDirectMemorySize=367001600, -Des.path.home=/usr/share/elasticsearch, -Des.path.conf=/usr/share/elasticsearch/config, -Des.distribution.flavor=default, -Des.distribution.type=docker, -Des.bundled_jdk=true]\" }\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chmod g+w '{}' ';'\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,926Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [aggs-matrix-stats]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [analysis-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [constant-keyword]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [flattened]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [frozen-indices]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-geoip]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-user-agent]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [kibana]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-expression]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-mustache]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-painless]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-extras]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-version]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [parent-join]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [percolator]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [rank-eval]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [reindex]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repositories-metering-api]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repository-url]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [search-business-rules]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [searchable-snapshots]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [spatial]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transform]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transport-netty4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [unsigned-long]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [vectors]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [wildcard]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-analytics]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async-search]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-autoscaling]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ccr]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-core]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-data-streams]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-deprecation]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-enrich]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-eql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-graph]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-identity-provider]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,936Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ilm]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-logstash]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ml]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-monitoring]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-rollup]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-security]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-sql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-stack]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-voting-only-node]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,938Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-watcher]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,943Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"no plugins loaded\" }\nelasticsearch | {\"type\": \"deprecation\", \"timestamp\": \"2026-04-20T06:39:12,054Z\", \"level\": \"DEPRECATION\", \"component\": \"o.e.d.c.s.Settings\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[node.data] setting was deprecated in Elasticsearch and will be removed in a future release! See the breaking changes documentation for the next major version.\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using [1] data paths, mounts [[/usr/share/elasticsearch/data (/dev/vda1)]], net usable_space [14.5gb], net total_space [58.3gb], types [ext4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"heap size [700mb], compressed ordinary object pointers [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,356Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"node name [e802ad473a4f], node ID [e2ZKzgw4Q4aCf2w5ljWr1A], cluster name [docker-cluster], roles [transform, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:18,805Z\", \"level\": \"INFO\", \"component\": \"o.e.x.m.p.l.CppLogMessageHandler\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[controller/213] [Main.cc@114] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,521Z\", \"level\": \"INFO\", \"component\": \"o.e.t.NettyAllocator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,695Z\", \"level\": \"INFO\", \"component\": \"o.e.d.DiscoveryModule\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using discovery type [single-node] and seed hosts providers [settings]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:22,462Z\", \"level\": \"WARN\", \"component\": \"o.e.g.DanglingIndicesState\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,302Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"initialized\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,303Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"starting ...\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,565Z\", \"level\": \"INFO\", \"component\": \"o.e.t.TransportService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9300}, bound_addresses {[::]:9300}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,081Z\", \"level\": \"INFO\", \"component\": \"o.e.c.c.Coordinator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,252Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.MasterService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,401Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.ClusterApplierService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.h.AbstractHttpServerTransport\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9200}, bound_addresses {[::]:9200}\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"started\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,162Z\", \"level\": \"INFO\", \"component\": \"o.e.l.LicenseService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,169Z\", \"level\": \"INFO\", \"component\": \"o.e.g.GatewayService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"recovered [15] indices into cluster_state\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:29,975Z\", \"level\": \"INFO\", \"component\": \"o.e.c.r.a.AllocationService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:02Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"auditTrail\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"visTypeXy\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"warning\",\"config\",\"deprecation\"],\"pid\":7,\"message\":\"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\\\"\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\ndocker_lamp_1 | + create_dot_env_local_file\ndocker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak\ndocker_lamp_1 | + create_dot_env\ndocker_lamp_1 | + [[ -f /home/jiminny/.env ]]\ndocker_lamp_1 | + return\ndocker_lamp_1 | + declare DB_ADMIN_PASSWORD\ndocker_lamp_1 | + declare DB_ADMIN_USERNAME\ndocker_lamp_1 | + declare DB_DEV_PASSWORD\ndocker_lamp_1 | + declare DB_DEV_USERNAME\ndocker_lamp_1 | + declare DB_ROOT_PASSWORD\ndocker_lamp_1 | + declare DB_ROOT_USERNAME\ndocker_lamp_1 | + declare DB_WEB_PASSWORD\ndocker_lamp_1 | + declare DB_WEB_USERNAME\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_PASSWORD='dgyt$rTe21-d'\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_USERNAME=root\ndocker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + readonly DB_ADMIN_PASSWORD\ndocker_lamp_1 | + readonly DB_ADMIN_USERNAME\ndocker_lamp_1 | + readonly DB_DEV_PASSWORD\ndocker_lamp_1 | + readonly DB_DEV_USERNAME\ndocker_lamp_1 | + readonly DB_ROOT_PASSWORD\ndocker_lamp_1 | + readonly DB_ROOT_USERNAME\ndocker_lamp_1 | + readonly DB_WEB_PASSWORD\ndocker_lamp_1 | + readonly DB_WEB_USERNAME\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=dgyt$rTe21-d~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.root\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml\"}\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Session cookies will be transmitted over insecure connections. This is not recommended.\"}\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local\ndocker_lamp_1 | + echo ''\ndocker_lamp_1 | + echo 'DB_ADMIN_PASSWORD=dgyt$rTe21-d'\ndocker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | + echo DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | + echo DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | + echo DB_ROOT_USERNAME=root\ndocker_lamp_1 | + echo DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + declare IP_ADDRESS\ndocker_lamp_1 | ++ curl ifconfig.co/\ndocker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current\ndocker_lamp_1 | Dload Upload Total Spent Left Speed\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"encryptedSavedObjects\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"ingestManager\"],\"pid\":7,\"message\":\"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\n100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224\ndocker_lamp_1 | + IP_ADDRESS=212.5.153.87\ndocker_lamp_1 | + readonly IP_ADDRESS\ndocker_lamp_1 | + sed -i -E 's~X-Forwarded-For \"\\$remote_addr\"~X-Forwarded-For \"IP_ADDRESS\"~g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + sed -i -E s~IP_ADDRESS~212.5.153.87~g /etc/nginx/nginx.conf\ndocker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \\{([^#]*)# QA Only End \\}//g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + storage_init\ndocker_lamp_1 | + false\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]\ndocker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1\ndocker_lamp_1 | ++ wc -l\ndocker_lamp_1 | + [[ 11 -lt 7 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]\ndocker_lamp_1 | + sleep 5\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Found 'server.host: \\\"0\\\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' is being automatically to the configuration. You can change the setting to 'server.host: 0.0.0.0' or add 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' in kibana.yml to prevent this message.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"actions\",\"actions\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"alerts\",\"plugins\",\"alerting\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\"],\"pid\":7,\"message\":\"config sourced from: production cluster\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Starting saved objects migrations\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:42:04 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \n\n\nv View in Docker Desktop o View Config w Enable Watch","depth":4,"value":"ins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:42Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:44Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:45Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:47Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:48Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:49Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:50Z\",\"tags\":[\"error\",\"plugins\",\"reporting\",\"esqueue\",\"queue-worker\",\"error\"],\"pid\":7,\"message\":\"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\\n at process._tickCallback (internal/process/next_tick.js:61:11)\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"elasticsearch\",\"monitoring\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:51Z\",\"tags\":[\"warning\",\"plugins\",\"licensing\"],\"pid\":7,\"message\":\"License information could not be obtained from Elasticsearch due to Error: No Living connections error\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:55Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:58Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:00Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:03Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:10Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work\nWARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion \n[+] Running 2/2\n ✔ Container redis Running 0.0s \n ✔ Container docker_lamp_1 Running 0.0s \nAttaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis\ndocker_lamp_1 | \ndocker_lamp_1 | Illuminate\\Database\\QueryException \ndocker_lamp_1 | \ndocker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)\ndocker_lamp_1 | \ndocker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838\ndocker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)\ndocker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class\ndocker_lamp_1 | 836▕ : QueryException::class;\ndocker_lamp_1 | 837▕ \ndocker_lamp_1 | ➜ 838▕ throw new $exceptionType(\ndocker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),\ndocker_lamp_1 | 840▕ $query,\ndocker_lamp_1 | 841▕ $this->prepareBindings($bindings),\ndocker_lamp_1 | 842▕ $e,\ndocker_lamp_1 | \ndocker_lamp_1 | +42 vendor frames \ndocker_lamp_1 | \ndocker_lamp_1 | 43 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \nmariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\nblackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.\nblackfire-1 | usage blackfire-agent [options]\nblackfire-1 | --collector=\"https://blackfire.io\": Sets the URL of Blackfire's data collector\nblackfire-1 | --config=\"/etc/blackfire/agent\": Sets the path to the configuration file\nblackfire-1 | -d: Prints the current configuration\nblackfire-1 | --http-proxy=\"\": Sets the HTTP proxy to use\nblackfire-1 | --https-proxy=\"\": Sets the HTTPS proxy to use\nblackfire-1 | --log-file=\"stderr\": Sets the path of the log file. Use stderr to log to stderr\nblackfire-1 | --log-level=\"1\": log verbosity level (4: debug, 3: info, 2: warning, 1: error)\nblackfire-1 | --register: Helps you with registering the agent\nblackfire-1 | --server-id=\"\": Sets the server id used to authenticate with Blackfire API\nblackfire-1 | --server-token=\"\": Sets the server token used to authenticate with Blackfire API. It is unsafe to set this from the command line\nblackfire-1 | --socket=\"unix:///var/run/blackfire/agent.sock\": Sets the socket the agent should read traces from. Possible value can be a unix socket or a TCP address. ie: unix:///var/run/blackfire/agent.sock or tcp://127.0.0.1:8307\nblackfire-1 | --test: Tests the configuration\nblackfire-1 | --timeout=\"15s\": Sets the Blackfire connection timeout\nblackfire-1 | -v: Prints the version number\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"no configuration paths supplied\"\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"using configuration at default config path\" path=/home/ngrok/.ngrok2/ngrok.yml\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"open config file\" path=/home/ngrok/.ngrok2/ngrok.yml err=nil\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"starting web service\" obj=web addr=0.0.0.0:4040\njiminny_ext-1 exited with code 0\nblackfire-1 exited with code 1\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"tunnel session started\" obj=tunnels.session\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"client session established\" obj=csess id=7d0b19e3215d\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=\"command_line (http)\" addr=http://lamp:3080 url=http://lukask.ngrok.io\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=command_line addr=http://lamp:3080 url=https://lukask.ngrok.io\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Warn] [Entrypoint]: /sys/fs/cgroup///memory.pressure not writable, functionality unavailable to MariaDB\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\ndatadog-1 | [s6-init] making user provided files available at /var/run/s6/etc...exited 0.\nngrok | t=2026-04-20T06:38:55+0000 lvl=info msg=\"update available\" obj=updater\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade information missing, assuming required\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade (mariadb-upgrade or creating healthcheck users) required, but skipped due to $MARIADB_AUTO_UPGRADE setting\nmariadb-1 | 2026-04-20 6:38:55 0 [Note] Starting MariaDB 11.4.5-MariaDB-ubu2404 source revision 0771110266ff5c04216af4bf1243c65f8c67ccf4 server_uid geAM8SSFn8P6Gz+0WMyyW27EUEM= as process 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Compressed tables use zlib 1.3\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Number of transaction pools: 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using ARMv8 crc32 + pmull instructions\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using liburing\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Completed initialization of buffer pool\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes)\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: End of log at LSN=7614257762\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Opened 3 undo tablespaces\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: 128 rollback segments in 3 undo tablespaces are active.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ...\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: log sequence number 7614257762; transaction id 11481082\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'FEEDBACK' is disabled.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'wsrep-provider' is disabled.\nmariadb-1 | 2026-04-20 6:38:58 0 [Note] InnoDB: Buffer pool(s) load completed at 260420 6:38:58\ndatadog-1 | [s6-init] ensuring user provided files have correct perms...exited 0.\ndatadog-1 | [fix-attrs.d] applying ownership & permissions fixes...\ndatadog-1 | [fix-attrs.d] done.\ndatadog-1 | [cont-init.d] executing container initialization scripts...\ndatadog-1 | [cont-init.d] 01-check-apikey.sh: executing... \ndatadog-1 | \ndatadog-1 | ==================================================================================\ndatadog-1 | You must set an DD_API_KEY environment variable to run the Datadog Agent container\ndatadog-1 | ==================================================================================\ndatadog-1 | \ndatadog-1 | [cont-init.d] 01-check-apikey.sh: exited 1.\ndatadog-1 exited with code 1\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '0.0.0.0'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '::'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: Event Scheduler: Loaded 0 events\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: ready for connections.\nmariadb-1 | Version: '11.4.5-MariaDB-ubu2404' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution\ndocker_lamp_1 | \ndocker_lamp_1 | INFO Configuration cached successfully. \ndocker_lamp_1 | \ndocker_lamp_1 | cp: cannot stat '/home/jiminny/bootstrap/cache/config-new.php': No such file or directory\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + [[ false == \\f\\a\\l\\s\\e ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + [[ -f /home/jiminny/storage/oauth-private.key ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + setup_local_environment\ndocker_lamp_1 | + storage_permissions_workaround\ndocker_lamp_1 | + [[ false != \\t\\r\\u\\e ]]\ndocker_lamp_1 | + ln -sf /home/jiminny/storage /home/jiminny/jiminny_storage\ndocker_lamp_1 | + return\ndocker_lamp_1 | + chmod 660 /home/jiminny/storage/oauth-private.key\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/views\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/cache\ndocker_lamp_1 | + chmod 775 /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/cache\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,413Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"version[7.10.2], pid[7], build[default/docker/747e1cc71def077253878a59143c1f785afa92b9/2021-01-13T04:42:47.157277Z], OS[Linux/6.12.54-linuxkit/aarch64], JVM[AdoptOpenJDK/OpenJDK 64-Bit Server VM/15.0.1/15.0.1+9]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM home [/usr/share/elasticsearch/jdk], using bundled JDK [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Djava.locale.providers=SPI,COMPAT, -Xms1g, -Xmx1g, -XX:+UseG1GC, -XX:G1ReservePercent=25, -XX:InitiatingHeapOccupancyPercent=30, -Djava.io.tmpdir=/tmp/elasticsearch-11712280666606670653, -XX:+HeapDumpOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -Des.cgroups.hierarchy.override=/, -Xms700m, -Xmx700m, -XX:MaxDirectMemorySize=367001600, -Des.path.home=/usr/share/elasticsearch, -Des.path.conf=/usr/share/elasticsearch/config, -Des.distribution.flavor=default, -Des.distribution.type=docker, -Des.bundled_jdk=true]\" }\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chmod g+w '{}' ';'\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,926Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [aggs-matrix-stats]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [analysis-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [constant-keyword]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [flattened]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [frozen-indices]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-geoip]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-user-agent]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [kibana]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-expression]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-mustache]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-painless]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-extras]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-version]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [parent-join]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [percolator]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [rank-eval]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [reindex]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repositories-metering-api]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repository-url]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [search-business-rules]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [searchable-snapshots]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [spatial]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transform]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transport-netty4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [unsigned-long]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [vectors]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [wildcard]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-analytics]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async-search]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-autoscaling]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ccr]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-core]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-data-streams]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-deprecation]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-enrich]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-eql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-graph]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-identity-provider]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,936Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ilm]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-logstash]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ml]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-monitoring]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-rollup]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-security]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-sql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-stack]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-voting-only-node]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,938Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-watcher]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,943Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"no plugins loaded\" }\nelasticsearch | {\"type\": \"deprecation\", \"timestamp\": \"2026-04-20T06:39:12,054Z\", \"level\": \"DEPRECATION\", \"component\": \"o.e.d.c.s.Settings\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[node.data] setting was deprecated in Elasticsearch and will be removed in a future release! See the breaking changes documentation for the next major version.\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using [1] data paths, mounts [[/usr/share/elasticsearch/data (/dev/vda1)]], net usable_space [14.5gb], net total_space [58.3gb], types [ext4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"heap size [700mb], compressed ordinary object pointers [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,356Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"node name [e802ad473a4f], node ID [e2ZKzgw4Q4aCf2w5ljWr1A], cluster name [docker-cluster], roles [transform, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:18,805Z\", \"level\": \"INFO\", \"component\": \"o.e.x.m.p.l.CppLogMessageHandler\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[controller/213] [Main.cc@114] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,521Z\", \"level\": \"INFO\", \"component\": \"o.e.t.NettyAllocator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,695Z\", \"level\": \"INFO\", \"component\": \"o.e.d.DiscoveryModule\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using discovery type [single-node] and seed hosts providers [settings]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:22,462Z\", \"level\": \"WARN\", \"component\": \"o.e.g.DanglingIndicesState\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,302Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"initialized\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,303Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"starting ...\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,565Z\", \"level\": \"INFO\", \"component\": \"o.e.t.TransportService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9300}, bound_addresses {[::]:9300}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,081Z\", \"level\": \"INFO\", \"component\": \"o.e.c.c.Coordinator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,252Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.MasterService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,401Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.ClusterApplierService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.h.AbstractHttpServerTransport\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9200}, bound_addresses {[::]:9200}\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"started\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,162Z\", \"level\": \"INFO\", \"component\": \"o.e.l.LicenseService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,169Z\", \"level\": \"INFO\", \"component\": \"o.e.g.GatewayService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"recovered [15] indices into cluster_state\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:29,975Z\", \"level\": \"INFO\", \"component\": \"o.e.c.r.a.AllocationService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:02Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"auditTrail\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"visTypeXy\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"warning\",\"config\",\"deprecation\"],\"pid\":7,\"message\":\"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\\\"\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\ndocker_lamp_1 | + create_dot_env_local_file\ndocker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak\ndocker_lamp_1 | + create_dot_env\ndocker_lamp_1 | + [[ -f /home/jiminny/.env ]]\ndocker_lamp_1 | + return\ndocker_lamp_1 | + declare DB_ADMIN_PASSWORD\ndocker_lamp_1 | + declare DB_ADMIN_USERNAME\ndocker_lamp_1 | + declare DB_DEV_PASSWORD\ndocker_lamp_1 | + declare DB_DEV_USERNAME\ndocker_lamp_1 | + declare DB_ROOT_PASSWORD\ndocker_lamp_1 | + declare DB_ROOT_USERNAME\ndocker_lamp_1 | + declare DB_WEB_PASSWORD\ndocker_lamp_1 | + declare DB_WEB_USERNAME\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_PASSWORD='dgyt$rTe21-d'\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_USERNAME=root\ndocker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + readonly DB_ADMIN_PASSWORD\ndocker_lamp_1 | + readonly DB_ADMIN_USERNAME\ndocker_lamp_1 | + readonly DB_DEV_PASSWORD\ndocker_lamp_1 | + readonly DB_DEV_USERNAME\ndocker_lamp_1 | + readonly DB_ROOT_PASSWORD\ndocker_lamp_1 | + readonly DB_ROOT_USERNAME\ndocker_lamp_1 | + readonly DB_WEB_PASSWORD\ndocker_lamp_1 | + readonly DB_WEB_USERNAME\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=dgyt$rTe21-d~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.root\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml\"}\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Session cookies will be transmitted over insecure connections. This is not recommended.\"}\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local\ndocker_lamp_1 | + echo ''\ndocker_lamp_1 | + echo 'DB_ADMIN_PASSWORD=dgyt$rTe21-d'\ndocker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | + echo DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | + echo DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | + echo DB_ROOT_USERNAME=root\ndocker_lamp_1 | + echo DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + declare IP_ADDRESS\ndocker_lamp_1 | ++ curl ifconfig.co/\ndocker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current\ndocker_lamp_1 | Dload Upload Total Spent Left Speed\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"encryptedSavedObjects\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"ingestManager\"],\"pid\":7,\"message\":\"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\n100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224\ndocker_lamp_1 | + IP_ADDRESS=212.5.153.87\ndocker_lamp_1 | + readonly IP_ADDRESS\ndocker_lamp_1 | + sed -i -E 's~X-Forwarded-For \"\\$remote_addr\"~X-Forwarded-For \"IP_ADDRESS\"~g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + sed -i -E s~IP_ADDRESS~212.5.153.87~g /etc/nginx/nginx.conf\ndocker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \\{([^#]*)# QA Only End \\}//g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + storage_init\ndocker_lamp_1 | + false\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]\ndocker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1\ndocker_lamp_1 | ++ wc -l\ndocker_lamp_1 | + [[ 11 -lt 7 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]\ndocker_lamp_1 | + sleep 5\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Found 'server.host: \\\"0\\\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' is being automatically to the configuration. You can change the setting to 'server.host: 0.0.0.0' or add 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' in kibana.yml to prevent this message.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"actions\",\"actions\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"alerts\",\"plugins\",\"alerting\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\"],\"pid\":7,\"message\":\"config sourced from: production cluster\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Starting saved objects migrations\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:42:04 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \n\n\nv View in Docker Desktop o View Config w Enable Watch","is_focused":true},{"role":"AXButton","text":"Menu","depth":3,"bounds":{"left":0.48402777,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥1 DOCKER (docker-compose)","depth":3,"bounds":{"left":0.01875,"top":0.09,"width":0.46180555,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥2 PROD (ssh)","depth":4,"bounds":{"left":0.51805556,"top":0.09,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.24944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥3 EU (ssh)","depth":4,"bounds":{"left":0.51805556,"top":0.25,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.4422222,"width":0.50069445,"height":0.12222222},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.41944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥4 STAGE (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.42,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.5922222,"width":0.50069445,"height":0.12222222},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.5694444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥5 QA (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.57,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:48:07 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.74222225,"width":0.50069445,"height":0.12222222},"value":"Last login: Thu Apr 16 15:48:07 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.71944445,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥6 FE (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.72,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:48:07 on ttys005\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.87222224,"width":0.50069445,"height":0.12777776},"value":"Last login: Thu Apr 16 15:48:07 on ttys005\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.86944443,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥7 EXT (-zsh)","depth":4,"bounds":{"left":0.51805556,"top":0.87,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.140625,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.14479166,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.28125,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.28541666,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.421875,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.42604166,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.5625,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.56666666,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"✳ Review screenpipe usage and Boosteroid integration (claude)","depth":2,"bounds":{"left":0.7027778,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.70694447,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.84305555,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.8472222,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.9548611,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"EU (ssh)","depth":1,"bounds":{"left":0.47916666,"top":0.033333335,"width":0.04097222,"height":0.017777778},"role_description":"text"}]...
|
6298747085787097907
|
-8341755016892981490
|
app_switch
|
accessibility
|
NULL
|
ins","taskManager","taskManager ins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:42Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:49Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_ROOT_USERNAME=root
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local
docker_lamp_1 | + declare IP_ADDRESS
docker_lamp_1 | ++ curl ifconfig.co/
docker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current
docker_lamp_1 | Dload Upload Total Spent Left Speed
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","encryptedSavedObjects","config"],"pid":7,"message":"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","ingestManager"],"pid":7,"message":"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224
docker_lamp_1 | + IP_ADDRESS=[IP_ADDRESS]
docker_lamp_1 | + readonly IP_ADDRESS
docker_lamp_1 | + sed -i -E 's~X-Forwarded-For "\$remote_addr"~X-Forwarded-For "IP_ADDRESS"~g' /etc/nginx/nginx.conf
docker_lamp_1 | + sed -i -E s~IP_ADDRESS~[IP_ADDRESS]~g /etc/nginx/nginx.conf
docker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \{([^#]*)# QA Only End \}//g' /etc/nginx/nginx.conf
docker_lamp_1 | + storage_init
docker_lamp_1 | + false
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/...
|
NULL
|
|
52157
|
1129
|
0
|
2026-04-20T06:42:06.354728+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667326354_m2.jpg...
|
iTerm2
|
EU (ssh)
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
slackStoyan Tomov2 minutes ago - image.pngOrganiza slackStoyan Tomov2 minutes ago - image.pngOrganization SettingsGeneralUsersTeamsIntegrationsJob TitlesActivityRecordingAl ContextAl Automation BETASidekickDeal InsighisVocabularyTopicsKey Words ScoringPlaybooks & Coaching FrameworksNotificationsSettingsPlaybooks & Coaching Frameworks®SEARCHI> • AMTest »:• • Pulsar New Business 41> • Vuelio AM's - Agency »]> • Vuelio AM's - Brands 4]> • Vuelio AM's-NFP «:> • Vuelio AM's-PS»:> • Vuelio AMs Playbook »:> • Vuelio New Business »]• Vuelio Insights Team <A AUTODETECT DISABLEDA AUTODETECT DISABLED• AUTODETECT DISABLEDA, AUTODETECT DISABLED.A AUTONETECT DISARI ErA AUTODETECT DISABLEDA AUTODETECT DISABLEDA, AUTODETECT DISABLED.A, AUTODETECT DISABLEDPLAYBOOKVuelio Insights leamLog Activity to Salesforce asNefaultAutodetectTest activity type promptsToct if vour nromnts corroctlvDownload$0.ny@localhost]A HS_Jocal [jiminny@localhost]& console [PKol)A console [EU] XRetactor Connection (• Retactor UserPilot 1Plavaroundda jiminny~027 49 A23 X3 X 105 Abps<->1..n: ON bps.stage_id = s.idd = 6024tractsent'd IN (16352,20612, 18281, 7344,16378,16309,5036, 15223, 14535, 6293, 12098, 11607).me LIKE "%Pulsar Group%'; # 472, 380, 15138, [EMAIL] ceam 10 = 4/2 # event 220147ories WHERE playbook_id = 2290;KE 10 = 22014/es WHERE crm_field_id = 226147;Jons WHERE 10 = 3885id = t.owner id THEN ' (owner)' ELSE 'I END) AS user id!ccounts saciable id"= u.team id•provider = 'salesforce':MOn ZU AOr 9:42:00+0 ..OCI runtime exec failed: exec failed:noine axpcuatted: exec /ined: banbpe un start sunt tne prodese exee:1 file +61 -6>Ask anvthina (%4L)+ «>CodeS Adaptiveai_activity_type_detection_enabled T® Reject allAccept allO name Y89970 Trial Playbook89962 Trial Playbook II89962 Untitled Plavbook.89962 Tasks89970 Events Playbook89970 Sales Playbook89962 Vuelio New Business89962 Pulsar New Business39970 Vuelio AMs Playbook89962 Vuelio AMs Playbook39970 test89962 AM Test39962 Vuelio AM's - Brands39962 Vuelio AM's - PS89962 Vuelio AM's - Agency39962 Vuelio AM's - NFP89962 Vuelio AM's - PS÷ is_selectable YI created_at Y0 2025-04-29 09:24:050 2025-05-14 16:14:090 2025-05-29 10:11:090 2025-08-06 11:17:480 2025-08-06 11:17:580 2025-08-14 12:38:190 2025-08-15 15:17:160 2025-08-18 16:22:240 2025-09-04 19:31:250 2025-09-05 11:14:1462025-00-15 12.59.290 2025-12-08 16:24:580 2025-12-12 12:31:500 2025-12-12 13:17:320 2025-12-12 13:23:1702025-10-19 13.24-22A 2A0G 12.12 12.04,27updated_at T2025-08-15 15:18:262025-05-29 10:13:142025-05-29 10:11:232025-08-06 11:19:042025-08-06 11:18:582025-08-15 15:18:212025-12-12 16:45:222025-08-18 16:22:242025-09-04 19:32:402025-12-12 16:45:002025-00-15 12•50•312025-12-12 16:24:552025-12-12 12:31:502025-12-12 13:23:362025-12-12 13:23:172005-10-10 13.24•220005 101 17.01.99CS| 1 →| 0, 8,I• deleted_at P2025-08-15 15:18:262025-05-29 10:13:142025-05-29 10:11:232025-08-06 11:19:042025-08-06 11:18:582025-08-15 15:18:212025-09-04 19:32:40<null>2025-00-15 12.59•31<null><nul>2025-12-12 13:23:36<null)<null><null>1622:19 UTF-8 4 spaces...
|
NULL
|
1857871175362317742
|
NULL
|
app_switch
|
ocr
|
NULL
|
slackStoyan Tomov2 minutes ago - image.pngOrganiza slackStoyan Tomov2 minutes ago - image.pngOrganization SettingsGeneralUsersTeamsIntegrationsJob TitlesActivityRecordingAl ContextAl Automation BETASidekickDeal InsighisVocabularyTopicsKey Words ScoringPlaybooks & Coaching FrameworksNotificationsSettingsPlaybooks & Coaching Frameworks®SEARCHI> • AMTest »:• • Pulsar New Business 41> • Vuelio AM's - Agency »]> • Vuelio AM's - Brands 4]> • Vuelio AM's-NFP «:> • Vuelio AM's-PS»:> • Vuelio AMs Playbook »:> • Vuelio New Business »]• Vuelio Insights Team <A AUTODETECT DISABLEDA AUTODETECT DISABLED• AUTODETECT DISABLEDA, AUTODETECT DISABLED.A AUTONETECT DISARI ErA AUTODETECT DISABLEDA AUTODETECT DISABLEDA, AUTODETECT DISABLED.A, AUTODETECT DISABLEDPLAYBOOKVuelio Insights leamLog Activity to Salesforce asNefaultAutodetectTest activity type promptsToct if vour nromnts corroctlvDownload$0.ny@localhost]A HS_Jocal [jiminny@localhost]& console [PKol)A console [EU] XRetactor Connection (• Retactor UserPilot 1Plavaroundda jiminny~027 49 A23 X3 X 105 Abps<->1..n: ON bps.stage_id = s.idd = 6024tractsent'd IN (16352,20612, 18281, 7344,16378,16309,5036, 15223, 14535, 6293, 12098, 11607).me LIKE "%Pulsar Group%'; # 472, 380, 15138, [EMAIL] ceam 10 = 4/2 # event 220147ories WHERE playbook_id = 2290;KE 10 = 22014/es WHERE crm_field_id = 226147;Jons WHERE 10 = 3885id = t.owner id THEN ' (owner)' ELSE 'I END) AS user id!ccounts saciable id"= u.team id•provider = 'salesforce':MOn ZU AOr 9:42:00+0 ..OCI runtime exec failed: exec failed:noine axpcuatted: exec /ined: banbpe un start sunt tne prodese exee:1 file +61 -6>Ask anvthina (%4L)+ «>CodeS Adaptiveai_activity_type_detection_enabled T® Reject allAccept allO name Y89970 Trial Playbook89962 Trial Playbook II89962 Untitled Plavbook.89962 Tasks89970 Events Playbook89970 Sales Playbook89962 Vuelio New Business89962 Pulsar New Business39970 Vuelio AMs Playbook89962 Vuelio AMs Playbook39970 test89962 AM Test39962 Vuelio AM's - Brands39962 Vuelio AM's - PS89962 Vuelio AM's - Agency39962 Vuelio AM's - NFP89962 Vuelio AM's - PS÷ is_selectable YI created_at Y0 2025-04-29 09:24:050 2025-05-14 16:14:090 2025-05-29 10:11:090 2025-08-06 11:17:480 2025-08-06 11:17:580 2025-08-14 12:38:190 2025-08-15 15:17:160 2025-08-18 16:22:240 2025-09-04 19:31:250 2025-09-05 11:14:1462025-00-15 12.59.290 2025-12-08 16:24:580 2025-12-12 12:31:500 2025-12-12 13:17:320 2025-12-12 13:23:1702025-10-19 13.24-22A 2A0G 12.12 12.04,27updated_at T2025-08-15 15:18:262025-05-29 10:13:142025-05-29 10:11:232025-08-06 11:19:042025-08-06 11:18:582025-08-15 15:18:212025-12-12 16:45:222025-08-18 16:22:242025-09-04 19:32:402025-12-12 16:45:002025-00-15 12•50•312025-12-12 16:24:552025-12-12 12:31:502025-12-12 13:23:362025-12-12 13:23:172005-10-10 13.24•220005 101 17.01.99CS| 1 →| 0, 8,I• deleted_at P2025-08-15 15:18:262025-05-29 10:13:142025-05-29 10:11:232025-08-06 11:19:042025-08-06 11:18:582025-08-15 15:18:212025-09-04 19:32:40<null>2025-00-15 12.59•31<null><nul>2025-12-12 13:23:36<null)<null><null>1622:19 UTF-8 4 spaces...
|
NULL
|
|
52172
|
1128
|
7
|
2026-04-20T06:43:39.747692+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667419747_m1.jpg...
|
iTerm2
|
EU (ssh)
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
kibana | {"type":"log", kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_ROOT_USERNAME=root
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local
docker_lamp_1 | + declare IP_ADDRESS
docker_lamp_1 | ++ curl ifconfig.co/
docker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current
docker_lamp_1 | Dload Upload Total Spent Left Speed
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","encryptedSavedObjects","config"],"pid":7,"message":"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","ingestManager"],"pid":7,"message":"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224
docker_lamp_1 | + IP_ADDRESS=[IP_ADDRESS]
docker_lamp_1 | + readonly IP_ADDRESS
docker_lamp_1 | + sed -i -E 's~X-Forwarded-For "\$remote_addr"~X-Forwarded-For "IP_ADDRESS"~g' /etc/nginx/nginx.conf
docker_lamp_1 | + sed -i -E s~IP_ADDRESS~[IP_ADDRESS]~g /etc/nginx/nginx.conf
docker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \{([^#]*)# QA Only End \}//g' /etc/nginx/nginx.conf
docker_lamp_1 | + storage_init
docker_lamp_1 | + false
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]
docker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1
docker_lamp_1 | ++ wc -l
docker_lamp_1 | + [[ 11 -lt 7 ]]
docker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework
docker_lamp_1 | + [[ 775 -ne 775 ]]
docker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache
docker_lamp_1 | + [[ 775 -ne 775 ]]
docker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views
docker_lamp_1 | + [[ 775 -ne 775 ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]
docker_lamp_1 | + sleep 5
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","reporting","config"],"pid":7,"message":"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","reporting","config"],"pid":7,"message":"Found 'server.host: \"0\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: [IP_ADDRESS]' is being automatically to the configuration. You can change the setting to 'server.host: [IP_ADDRESS]' or add 'xpack.reporting.kibanaServer.hostname: [IP_ADDRESS]' in kibana.yml to prevent this message."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","reporting","config"],"pid":7,"message":"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","actions","actions"],"pid":7,"message":"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","alerts","plugins","alerting"],"pid":7,"message":"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:05Z","tags":["info","plugins","monitoring","monitoring"],"pid":7,"message":"config sourced from: production cluster"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:05Z","tags":["info","savedobjects-service"],"pid":7,"message":"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations..."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:05Z","tags":["info","savedobjects-service"],"pid":7,"message":"Starting saved objects migrations"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins-system"],"pid":7,"message":"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins","taskManager","taskManager"],"pid":7,"message":"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a"}
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:40:06,605Z", "level": "INFO", "component": "o.e.c.m.MetadataIndexTemplateService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "adding template [.management-beats] for index patterns [.management-beats]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
kibana | {"type":"log","@timestamp":"20...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"kibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:55Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:58Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:00Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:03Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:10Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work\nWARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion \n[+] Running 2/2\n ✔ Container redis Running 0.0s \n ✔ Container docker_lamp_1 Running 0.0s \nAttaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis\ndocker_lamp_1 | \ndocker_lamp_1 | Illuminate\\Database\\QueryException \ndocker_lamp_1 | \ndocker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)\ndocker_lamp_1 | \ndocker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838\ndocker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)\ndocker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class\ndocker_lamp_1 | 836▕ : QueryException::class;\ndocker_lamp_1 | 837▕ \ndocker_lamp_1 | ➜ 838▕ throw new $exceptionType(\ndocker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),\ndocker_lamp_1 | 840▕ $query,\ndocker_lamp_1 | 841▕ $this->prepareBindings($bindings),\ndocker_lamp_1 | 842▕ $e,\ndocker_lamp_1 | \ndocker_lamp_1 | +42 vendor frames \ndocker_lamp_1 | \ndocker_lamp_1 | 43 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \nmariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\nblackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.\nblackfire-1 | usage blackfire-agent [options]\nblackfire-1 | --collector=\"https://blackfire.io\": Sets the URL of Blackfire's data collector\nblackfire-1 | --config=\"/etc/blackfire/agent\": Sets the path to the configuration file\nblackfire-1 | -d: Prints the current configuration\nblackfire-1 | --http-proxy=\"\": Sets the HTTP proxy to use\nblackfire-1 | --https-proxy=\"\": Sets the HTTPS proxy to use\nblackfire-1 | --log-file=\"stderr\": Sets the path of the log file. Use stderr to log to stderr\nblackfire-1 | --log-level=\"1\": log verbosity level (4: debug, 3: info, 2: warning, 1: error)\nblackfire-1 | --register: Helps you with registering the agent\nblackfire-1 | --server-id=\"\": Sets the server id used to authenticate with Blackfire API\nblackfire-1 | --server-token=\"\": Sets the server token used to authenticate with Blackfire API. It is unsafe to set this from the command line\nblackfire-1 | --socket=\"unix:///var/run/blackfire/agent.sock\": Sets the socket the agent should read traces from. Possible value can be a unix socket or a TCP address. ie: unix:///var/run/blackfire/agent.sock or tcp://127.0.0.1:8307\nblackfire-1 | --test: Tests the configuration\nblackfire-1 | --timeout=\"15s\": Sets the Blackfire connection timeout\nblackfire-1 | -v: Prints the version number\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"no configuration paths supplied\"\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"using configuration at default config path\" path=/home/ngrok/.ngrok2/ngrok.yml\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"open config file\" path=/home/ngrok/.ngrok2/ngrok.yml err=nil\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"starting web service\" obj=web addr=0.0.0.0:4040\njiminny_ext-1 exited with code 0\nblackfire-1 exited with code 1\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"tunnel session started\" obj=tunnels.session\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"client session established\" obj=csess id=7d0b19e3215d\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=\"command_line (http)\" addr=http://lamp:3080 url=http://lukask.ngrok.io\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=command_line addr=http://lamp:3080 url=https://lukask.ngrok.io\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Warn] [Entrypoint]: /sys/fs/cgroup///memory.pressure not writable, functionality unavailable to MariaDB\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\ndatadog-1 | [s6-init] making user provided files available at /var/run/s6/etc...exited 0.\nngrok | t=2026-04-20T06:38:55+0000 lvl=info msg=\"update available\" obj=updater\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade information missing, assuming required\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade (mariadb-upgrade or creating healthcheck users) required, but skipped due to $MARIADB_AUTO_UPGRADE setting\nmariadb-1 | 2026-04-20 6:38:55 0 [Note] Starting MariaDB 11.4.5-MariaDB-ubu2404 source revision 0771110266ff5c04216af4bf1243c65f8c67ccf4 server_uid geAM8SSFn8P6Gz+0WMyyW27EUEM= as process 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Compressed tables use zlib 1.3\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Number of transaction pools: 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using ARMv8 crc32 + pmull instructions\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using liburing\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Completed initialization of buffer pool\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes)\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: End of log at LSN=7614257762\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Opened 3 undo tablespaces\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: 128 rollback segments in 3 undo tablespaces are active.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ...\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: log sequence number 7614257762; transaction id 11481082\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'FEEDBACK' is disabled.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'wsrep-provider' is disabled.\nmariadb-1 | 2026-04-20 6:38:58 0 [Note] InnoDB: Buffer pool(s) load completed at 260420 6:38:58\ndatadog-1 | [s6-init] ensuring user provided files have correct perms...exited 0.\ndatadog-1 | [fix-attrs.d] applying ownership & permissions fixes...\ndatadog-1 | [fix-attrs.d] done.\ndatadog-1 | [cont-init.d] executing container initialization scripts...\ndatadog-1 | [cont-init.d] 01-check-apikey.sh: executing... \ndatadog-1 | \ndatadog-1 | ==================================================================================\ndatadog-1 | You must set an DD_API_KEY environment variable to run the Datadog Agent container\ndatadog-1 | ==================================================================================\ndatadog-1 | \ndatadog-1 | [cont-init.d] 01-check-apikey.sh: exited 1.\ndatadog-1 exited with code 1\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '0.0.0.0'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '::'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: Event Scheduler: Loaded 0 events\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: ready for connections.\nmariadb-1 | Version: '11.4.5-MariaDB-ubu2404' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution\ndocker_lamp_1 | \ndocker_lamp_1 | INFO Configuration cached successfully. \ndocker_lamp_1 | \ndocker_lamp_1 | cp: cannot stat '/home/jiminny/bootstrap/cache/config-new.php': No such file or directory\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + [[ false == \\f\\a\\l\\s\\e ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + [[ -f /home/jiminny/storage/oauth-private.key ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + setup_local_environment\ndocker_lamp_1 | + storage_permissions_workaround\ndocker_lamp_1 | + [[ false != \\t\\r\\u\\e ]]\ndocker_lamp_1 | + ln -sf /home/jiminny/storage /home/jiminny/jiminny_storage\ndocker_lamp_1 | + return\ndocker_lamp_1 | + chmod 660 /home/jiminny/storage/oauth-private.key\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/views\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/cache\ndocker_lamp_1 | + chmod 775 /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/cache\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,413Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"version[7.10.2], pid[7], build[default/docker/747e1cc71def077253878a59143c1f785afa92b9/2021-01-13T04:42:47.157277Z], OS[Linux/6.12.54-linuxkit/aarch64], JVM[AdoptOpenJDK/OpenJDK 64-Bit Server VM/15.0.1/15.0.1+9]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM home [/usr/share/elasticsearch/jdk], using bundled JDK [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Djava.locale.providers=SPI,COMPAT, -Xms1g, -Xmx1g, -XX:+UseG1GC, -XX:G1ReservePercent=25, -XX:InitiatingHeapOccupancyPercent=30, -Djava.io.tmpdir=/tmp/elasticsearch-11712280666606670653, -XX:+HeapDumpOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -Des.cgroups.hierarchy.override=/, -Xms700m, -Xmx700m, -XX:MaxDirectMemorySize=367001600, -Des.path.home=/usr/share/elasticsearch, -Des.path.conf=/usr/share/elasticsearch/config, -Des.distribution.flavor=default, -Des.distribution.type=docker, -Des.bundled_jdk=true]\" }\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chmod g+w '{}' ';'\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,926Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [aggs-matrix-stats]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [analysis-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [constant-keyword]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [flattened]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [frozen-indices]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-geoip]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-user-agent]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [kibana]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-expression]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-mustache]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-painless]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-extras]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-version]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [parent-join]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [percolator]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [rank-eval]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [reindex]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repositories-metering-api]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repository-url]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [search-business-rules]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [searchable-snapshots]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [spatial]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transform]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transport-netty4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [unsigned-long]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [vectors]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [wildcard]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-analytics]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async-search]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-autoscaling]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ccr]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-core]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-data-streams]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-deprecation]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-enrich]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-eql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-graph]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-identity-provider]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,936Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ilm]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-logstash]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ml]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-monitoring]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-rollup]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-security]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-sql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-stack]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-voting-only-node]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,938Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-watcher]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,943Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"no plugins loaded\" }\nelasticsearch | {\"type\": \"deprecation\", \"timestamp\": \"2026-04-20T06:39:12,054Z\", \"level\": \"DEPRECATION\", \"component\": \"o.e.d.c.s.Settings\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[node.data] setting was deprecated in Elasticsearch and will be removed in a future release! See the breaking changes documentation for the next major version.\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using [1] data paths, mounts [[/usr/share/elasticsearch/data (/dev/vda1)]], net usable_space [14.5gb], net total_space [58.3gb], types [ext4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"heap size [700mb], compressed ordinary object pointers [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,356Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"node name [e802ad473a4f], node ID [e2ZKzgw4Q4aCf2w5ljWr1A], cluster name [docker-cluster], roles [transform, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:18,805Z\", \"level\": \"INFO\", \"component\": \"o.e.x.m.p.l.CppLogMessageHandler\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[controller/213] [Main.cc@114] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,521Z\", \"level\": \"INFO\", \"component\": \"o.e.t.NettyAllocator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,695Z\", \"level\": \"INFO\", \"component\": \"o.e.d.DiscoveryModule\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using discovery type [single-node] and seed hosts providers [settings]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:22,462Z\", \"level\": \"WARN\", \"component\": \"o.e.g.DanglingIndicesState\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,302Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"initialized\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,303Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"starting ...\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,565Z\", \"level\": \"INFO\", \"component\": \"o.e.t.TransportService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9300}, bound_addresses {[::]:9300}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,081Z\", \"level\": \"INFO\", \"component\": \"o.e.c.c.Coordinator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,252Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.MasterService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,401Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.ClusterApplierService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.h.AbstractHttpServerTransport\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9200}, bound_addresses {[::]:9200}\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"started\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,162Z\", \"level\": \"INFO\", \"component\": \"o.e.l.LicenseService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,169Z\", \"level\": \"INFO\", \"component\": \"o.e.g.GatewayService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"recovered [15] indices into cluster_state\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:29,975Z\", \"level\": \"INFO\", \"component\": \"o.e.c.r.a.AllocationService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:02Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"auditTrail\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"visTypeXy\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"warning\",\"config\",\"deprecation\"],\"pid\":7,\"message\":\"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\\\"\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\ndocker_lamp_1 | + create_dot_env_local_file\ndocker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak\ndocker_lamp_1 | + create_dot_env\ndocker_lamp_1 | + [[ -f /home/jiminny/.env ]]\ndocker_lamp_1 | + return\ndocker_lamp_1 | + declare DB_ADMIN_PASSWORD\ndocker_lamp_1 | + declare DB_ADMIN_USERNAME\ndocker_lamp_1 | + declare DB_DEV_PASSWORD\ndocker_lamp_1 | + declare DB_DEV_USERNAME\ndocker_lamp_1 | + declare DB_ROOT_PASSWORD\ndocker_lamp_1 | + declare DB_ROOT_USERNAME\ndocker_lamp_1 | + declare DB_WEB_PASSWORD\ndocker_lamp_1 | + declare DB_WEB_USERNAME\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_PASSWORD='dgyt$rTe21-d'\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_USERNAME=root\ndocker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + readonly DB_ADMIN_PASSWORD\ndocker_lamp_1 | + readonly DB_ADMIN_USERNAME\ndocker_lamp_1 | + readonly DB_DEV_PASSWORD\ndocker_lamp_1 | + readonly DB_DEV_USERNAME\ndocker_lamp_1 | + readonly DB_ROOT_PASSWORD\ndocker_lamp_1 | + readonly DB_ROOT_USERNAME\ndocker_lamp_1 | + readonly DB_WEB_PASSWORD\ndocker_lamp_1 | + readonly DB_WEB_USERNAME\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=dgyt$rTe21-d~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.root\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml\"}\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Session cookies will be transmitted over insecure connections. This is not recommended.\"}\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local\ndocker_lamp_1 | + echo ''\ndocker_lamp_1 | + echo 'DB_ADMIN_PASSWORD=dgyt$rTe21-d'\ndocker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | + echo DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | + echo DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | + echo DB_ROOT_USERNAME=root\ndocker_lamp_1 | + echo DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + declare IP_ADDRESS\ndocker_lamp_1 | ++ curl ifconfig.co/\ndocker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current\ndocker_lamp_1 | Dload Upload Total Spent Left Speed\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"encryptedSavedObjects\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"ingestManager\"],\"pid\":7,\"message\":\"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\n100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224\ndocker_lamp_1 | + IP_ADDRESS=212.5.153.87\ndocker_lamp_1 | + readonly IP_ADDRESS\ndocker_lamp_1 | + sed -i -E 's~X-Forwarded-For \"\\$remote_addr\"~X-Forwarded-For \"IP_ADDRESS\"~g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + sed -i -E s~IP_ADDRESS~212.5.153.87~g /etc/nginx/nginx.conf\ndocker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \\{([^#]*)# QA Only End \\}//g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + storage_init\ndocker_lamp_1 | + false\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]\ndocker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1\ndocker_lamp_1 | ++ wc -l\ndocker_lamp_1 | + [[ 11 -lt 7 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]\ndocker_lamp_1 | + sleep 5\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Found 'server.host: \\\"0\\\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' is being automatically to the configuration. You can change the setting to 'server.host: 0.0.0.0' or add 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' in kibana.yml to prevent this message.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"actions\",\"actions\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"alerts\",\"plugins\",\"alerting\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\"],\"pid\":7,\"message\":\"config sourced from: production cluster\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Starting saved objects migrations\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:42:04 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:06 Running ['artisan' dialers:monitor-activities] . 2s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:08 Running ['artisan' jiminny:monitor-social-accounts] 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:12 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:14 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:15 Running ['artisan' conference:monitor:count] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:16 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:42:20 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:42:20 Jiminny\\Jobs\\Mailbox\\CreateBatches ........ 44.79ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:43:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:06 Running ['artisan' jiminny:monitor-social-accounts] 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:11 Running ['artisan' mailbox:skip-lists:refresh] . 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:14 Running ['artisan' mailbox:batch:process --max-batches=15] 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 12.38ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' calendar:sync --dateMode=daily] 2026-04-20 06:43:31 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:32 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... 1s DONE\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\ndocker_lamp_1 | 14s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' calendar:sync --dateMode=daily > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents . 324.40ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\n\n\nv View in Docker Desktop o View Config w Enable Watch","depth":4,"value":"kibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:53Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:55Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:56Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:58Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:26:59Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:00Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:02Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:03Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:05Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:08Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\ndocker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:10Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200\"}\ndocker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"Unable to revive connection: http://elasticsearch:9200/\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"warning\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"No living connections\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-18T13:27:11Z\",\"tags\":[\"error\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"Failed to poll for work: Error: No Living connections\"}\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work\nWARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion \n[+] Running 2/2\n ✔ Container redis Running 0.0s \n ✔ Container docker_lamp_1 Running 0.0s \nAttaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis\ndocker_lamp_1 | \ndocker_lamp_1 | Illuminate\\Database\\QueryException \ndocker_lamp_1 | \ndocker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)\ndocker_lamp_1 | \ndocker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838\ndocker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)\ndocker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class\ndocker_lamp_1 | 836▕ : QueryException::class;\ndocker_lamp_1 | 837▕ \ndocker_lamp_1 | ➜ 838▕ throw new $exceptionType(\ndocker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),\ndocker_lamp_1 | 840▕ $query,\ndocker_lamp_1 | 841▕ $this->prepareBindings($bindings),\ndocker_lamp_1 | 842▕ $e,\ndocker_lamp_1 | \ndocker_lamp_1 | +42 vendor frames \ndocker_lamp_1 | \ndocker_lamp_1 | 43 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \nmariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\nblackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.\nblackfire-1 | usage blackfire-agent [options]\nblackfire-1 | --collector=\"https://blackfire.io\": Sets the URL of Blackfire's data collector\nblackfire-1 | --config=\"/etc/blackfire/agent\": Sets the path to the configuration file\nblackfire-1 | -d: Prints the current configuration\nblackfire-1 | --http-proxy=\"\": Sets the HTTP proxy to use\nblackfire-1 | --https-proxy=\"\": Sets the HTTPS proxy to use\nblackfire-1 | --log-file=\"stderr\": Sets the path of the log file. Use stderr to log to stderr\nblackfire-1 | --log-level=\"1\": log verbosity level (4: debug, 3: info, 2: warning, 1: error)\nblackfire-1 | --register: Helps you with registering the agent\nblackfire-1 | --server-id=\"\": Sets the server id used to authenticate with Blackfire API\nblackfire-1 | --server-token=\"\": Sets the server token used to authenticate with Blackfire API. It is unsafe to set this from the command line\nblackfire-1 | --socket=\"unix:///var/run/blackfire/agent.sock\": Sets the socket the agent should read traces from. Possible value can be a unix socket or a TCP address. ie: unix:///var/run/blackfire/agent.sock or tcp://127.0.0.1:8307\nblackfire-1 | --test: Tests the configuration\nblackfire-1 | --timeout=\"15s\": Sets the Blackfire connection timeout\nblackfire-1 | -v: Prints the version number\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"no configuration paths supplied\"\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"using configuration at default config path\" path=/home/ngrok/.ngrok2/ngrok.yml\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"open config file\" path=/home/ngrok/.ngrok2/ngrok.yml err=nil\nngrok | t=2026-04-20T06:38:52+0000 lvl=info msg=\"starting web service\" obj=web addr=0.0.0.0:4040\njiminny_ext-1 exited with code 0\nblackfire-1 exited with code 1\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"tunnel session started\" obj=tunnels.session\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"client session established\" obj=csess id=7d0b19e3215d\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=\"command_line (http)\" addr=http://lamp:3080 url=http://lukask.ngrok.io\nngrok | t=2026-04-20T06:38:54+0000 lvl=info msg=\"started tunnel\" obj=tunnels name=command_line addr=http://lamp:3080 url=https://lukask.ngrok.io\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Warn] [Entrypoint]: /sys/fs/cgroup///memory.pressure not writable, functionality unavailable to MariaDB\nmariadb-1 | 2026-04-20 06:38:54+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.\ndatadog-1 | [s6-init] making user provided files available at /var/run/s6/etc...exited 0.\nngrok | t=2026-04-20T06:38:55+0000 lvl=info msg=\"update available\" obj=updater\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade information missing, assuming required\nmariadb-1 | 2026-04-20 06:38:55+00:00 [Note] [Entrypoint]: MariaDB upgrade (mariadb-upgrade or creating healthcheck users) required, but skipped due to $MARIADB_AUTO_UPGRADE setting\nmariadb-1 | 2026-04-20 6:38:55 0 [Note] Starting MariaDB 11.4.5-MariaDB-ubu2404 source revision 0771110266ff5c04216af4bf1243c65f8c67ccf4 server_uid geAM8SSFn8P6Gz+0WMyyW27EUEM= as process 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Compressed tables use zlib 1.3\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Number of transaction pools: 1\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using ARMv8 crc32 + pmull instructions\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Using liburing\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: Completed initialization of buffer pool\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes)\nmariadb-1 | 2026-04-20 6:38:56 0 [Note] InnoDB: End of log at LSN=7614257762\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Opened 3 undo tablespaces\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: 128 rollback segments in 3 undo tablespaces are active.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ...\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: log sequence number 7614257762; transaction id 11481082\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'FEEDBACK' is disabled.\nmariadb-1 | 2026-04-20 6:38:57 0 [Note] Plugin 'wsrep-provider' is disabled.\nmariadb-1 | 2026-04-20 6:38:58 0 [Note] InnoDB: Buffer pool(s) load completed at 260420 6:38:58\ndatadog-1 | [s6-init] ensuring user provided files have correct perms...exited 0.\ndatadog-1 | [fix-attrs.d] applying ownership & permissions fixes...\ndatadog-1 | [fix-attrs.d] done.\ndatadog-1 | [cont-init.d] executing container initialization scripts...\ndatadog-1 | [cont-init.d] 01-check-apikey.sh: executing... \ndatadog-1 | \ndatadog-1 | ==================================================================================\ndatadog-1 | You must set an DD_API_KEY environment variable to run the Datadog Agent container\ndatadog-1 | ==================================================================================\ndatadog-1 | \ndatadog-1 | [cont-init.d] 01-check-apikey.sh: exited 1.\ndatadog-1 exited with code 1\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '0.0.0.0'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] Server socket created on IP: '::'.\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: Event Scheduler: Loaded 0 events\nmariadb-1 | 2026-04-20 6:39:01 0 [Note] mariadbd: ready for connections.\nmariadb-1 | Version: '11.4.5-MariaDB-ubu2404' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution\ndocker_lamp_1 | \ndocker_lamp_1 | INFO Configuration cached successfully. \ndocker_lamp_1 | \ndocker_lamp_1 | cp: cannot stat '/home/jiminny/bootstrap/cache/config-new.php': No such file or directory\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + [[ false == \\f\\a\\l\\s\\e ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + [[ -f /home/jiminny/storage/oauth-private.key ]]\ndocker_lamp_1 | + [[ 0 -eq 1 ]]\ndocker_lamp_1 | + setup_local_environment\ndocker_lamp_1 | + storage_permissions_workaround\ndocker_lamp_1 | + [[ false != \\t\\r\\u\\e ]]\ndocker_lamp_1 | + ln -sf /home/jiminny/storage /home/jiminny/jiminny_storage\ndocker_lamp_1 | + return\ndocker_lamp_1 | + chmod 660 /home/jiminny/storage/oauth-private.key\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/views\ndocker_lamp_1 | + chmod 775 /home/jiminny/storage/framework/cache\ndocker_lamp_1 | + chmod 775 /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/cache\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,413Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"version[7.10.2], pid[7], build[default/docker/747e1cc71def077253878a59143c1f785afa92b9/2021-01-13T04:42:47.157277Z], OS[Linux/6.12.54-linuxkit/aarch64], JVM[AdoptOpenJDK/OpenJDK 64-Bit Server VM/15.0.1/15.0.1+9]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM home [/usr/share/elasticsearch/jdk], using bundled JDK [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:06,417Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Djava.locale.providers=SPI,COMPAT, -Xms1g, -Xmx1g, -XX:+UseG1GC, -XX:G1ReservePercent=25, -XX:InitiatingHeapOccupancyPercent=30, -Djava.io.tmpdir=/tmp/elasticsearch-11712280666606670653, -XX:+HeapDumpOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -Des.cgroups.hierarchy.override=/, -Xms700m, -Xmx700m, -XX:MaxDirectMemorySize=367001600, -Des.path.home=/usr/share/elasticsearch, -Des.path.conf=/usr/share/elasticsearch/config, -Des.distribution.flavor=default, -Des.distribution.type=docker, -Des.bundled_jdk=true]\" }\ndocker_lamp_1 | + chmod -R 775 /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chmod g+w '{}' ';'\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,926Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [aggs-matrix-stats]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [analysis-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [constant-keyword]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [flattened]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,927Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [frozen-indices]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-common]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-geoip]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [ingest-user-agent]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,928Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [kibana]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-expression]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-mustache]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,929Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [lang-painless]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-extras]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [mapper-version]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,930Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [parent-join]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [percolator]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [rank-eval]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,931Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [reindex]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repositories-metering-api]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [repository-url]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,932Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [search-business-rules]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [searchable-snapshots]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,933Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [spatial]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transform]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [transport-netty4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [unsigned-long]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [vectors]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [wildcard]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-analytics]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-async-search]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,934Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-autoscaling]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ccr]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-core]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-data-streams]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-deprecation]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-enrich]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-eql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-graph]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,935Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-identity-provider]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,936Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ilm]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-logstash]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ml]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-monitoring]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-ql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-rollup]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-security]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-sql]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-stack]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,937Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-voting-only-node]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,938Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"loaded module [x-pack-watcher]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:11,943Z\", \"level\": \"INFO\", \"component\": \"o.e.p.PluginsService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"no plugins loaded\" }\nelasticsearch | {\"type\": \"deprecation\", \"timestamp\": \"2026-04-20T06:39:12,054Z\", \"level\": \"DEPRECATION\", \"component\": \"o.e.d.c.s.Settings\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[node.data] setting was deprecated in Elasticsearch and will be removed in a future release! See the breaking changes documentation for the next major version.\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using [1] data paths, mounts [[/usr/share/elasticsearch/data (/dev/vda1)]], net usable_space [14.5gb], net total_space [58.3gb], types [ext4]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,102Z\", \"level\": \"INFO\", \"component\": \"o.e.e.NodeEnvironment\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"heap size [700mb], compressed ordinary object pointers [true]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:12,356Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"node name [e802ad473a4f], node ID [e2ZKzgw4Q4aCf2w5ljWr1A], cluster name [docker-cluster], roles [transform, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:18,805Z\", \"level\": \"INFO\", \"component\": \"o.e.x.m.p.l.CppLogMessageHandler\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[controller/213] [Main.cc@114] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,521Z\", \"level\": \"INFO\", \"component\": \"o.e.t.NettyAllocator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:21,695Z\", \"level\": \"INFO\", \"component\": \"o.e.d.DiscoveryModule\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"using discovery type [single-node] and seed hosts providers [settings]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:22,462Z\", \"level\": \"WARN\", \"component\": \"o.e.g.DanglingIndicesState\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,302Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"initialized\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,303Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"starting ...\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:23,565Z\", \"level\": \"INFO\", \"component\": \"o.e.t.TransportService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9300}, bound_addresses {[::]:9300}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,081Z\", \"level\": \"INFO\", \"component\": \"o.e.c.c.Coordinator\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,252Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.MasterService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,401Z\", \"level\": \"INFO\", \"component\": \"o.e.c.s.ClusterApplierService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{172.18.0.7}{172.18.0.7:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.h.AbstractHttpServerTransport\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"publish_address {172.18.0.7:9200}, bound_addresses {[::]:9200}\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:24,521Z\", \"level\": \"INFO\", \"component\": \"o.e.n.Node\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"started\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,162Z\", \"level\": \"INFO\", \"component\": \"o.e.l.LicenseService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:25,169Z\", \"level\": \"INFO\", \"component\": \"o.e.g.GatewayService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"recovered [15] indices into cluster_state\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:39:29,975Z\", \"level\": \"INFO\", \"component\": \"o.e.c.r.a.AllocationService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:02Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"auditTrail\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-service\"],\"pid\":7,\"message\":\"Plugin \\\"visTypeXy\\\" is disabled.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"warning\",\"config\",\"deprecation\"],\"pid\":7,\"message\":\"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\\\"\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:03Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\ndocker_lamp_1 | + create_dot_env_local_file\ndocker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak\ndocker_lamp_1 | + create_dot_env\ndocker_lamp_1 | + [[ -f /home/jiminny/.env ]]\ndocker_lamp_1 | + return\ndocker_lamp_1 | + declare DB_ADMIN_PASSWORD\ndocker_lamp_1 | + declare DB_ADMIN_USERNAME\ndocker_lamp_1 | + declare DB_DEV_PASSWORD\ndocker_lamp_1 | + declare DB_DEV_USERNAME\ndocker_lamp_1 | + declare DB_ROOT_PASSWORD\ndocker_lamp_1 | + declare DB_ROOT_USERNAME\ndocker_lamp_1 | + declare DB_WEB_PASSWORD\ndocker_lamp_1 | + declare DB_WEB_USERNAME\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_PASSWORD='dgyt$rTe21-d'\ndocker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_ROOT_USERNAME=root\ndocker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json\ndocker_lamp_1 | + DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + readonly DB_ADMIN_PASSWORD\ndocker_lamp_1 | + readonly DB_ADMIN_USERNAME\ndocker_lamp_1 | + readonly DB_DEV_PASSWORD\ndocker_lamp_1 | + readonly DB_DEV_USERNAME\ndocker_lamp_1 | + readonly DB_ROOT_PASSWORD\ndocker_lamp_1 | + readonly DB_ROOT_USERNAME\ndocker_lamp_1 | + readonly DB_WEB_PASSWORD\ndocker_lamp_1 | + readonly DB_WEB_USERNAME\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=dgyt$rTe21-d~g' /home/jiminny/.env\ndocker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.migrate\ndocker_lamp_1 | + sed -i -E 's~DB_PASSWORD=.*$~DB_PASSWORD=b7h5-1fH3e54J~g' /home/jiminny/.env.root\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml\"}\ndocker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"security\",\"config\"],\"pid\":7,\"message\":\"Session cookies will be transmitted over insecure connections. This is not recommended.\"}\ndocker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local\ndocker_lamp_1 | + echo ''\ndocker_lamp_1 | + echo 'DB_ADMIN_PASSWORD=dgyt$rTe21-d'\ndocker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin\ndocker_lamp_1 | + echo DB_DEV_PASSWORD=rTr4sdQA65-Ad\ndocker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev\ndocker_lamp_1 | + echo DB_ROOT_PASSWORD=b7h5-1fH3e54J\ndocker_lamp_1 | + echo DB_ROOT_USERNAME=root\ndocker_lamp_1 | + echo DB_WEB_PASSWORD=aR5-EWf23b8da\ndocker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb\ndocker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local\ndocker_lamp_1 | + declare IP_ADDRESS\ndocker_lamp_1 | ++ curl ifconfig.co/\ndocker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current\ndocker_lamp_1 | Dload Upload Total Spent Left Speed\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"encryptedSavedObjects\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"ingestManager\"],\"pid\":7,\"message\":\"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\n100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224\ndocker_lamp_1 | + IP_ADDRESS=212.5.153.87\ndocker_lamp_1 | + readonly IP_ADDRESS\ndocker_lamp_1 | + sed -i -E 's~X-Forwarded-For \"\\$remote_addr\"~X-Forwarded-For \"IP_ADDRESS\"~g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + sed -i -E s~IP_ADDRESS~212.5.153.87~g /etc/nginx/nginx.conf\ndocker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \\{([^#]*)# QA Only End \\}//g' /etc/nginx/nginx.conf\ndocker_lamp_1 | + storage_init\ndocker_lamp_1 | + false\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]\ndocker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1\ndocker_lamp_1 | ++ wc -l\ndocker_lamp_1 | + [[ 11 -lt 7 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views\ndocker_lamp_1 | + [[ 775 -ne 775 ]]\ndocker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]\ndocker_lamp_1 | + sleep 5\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Found 'server.host: \\\"0\\\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' is being automatically to the configuration. You can change the setting to 'server.host: 0.0.0.0' or add 'xpack.reporting.kibanaServer.hostname: 0.0.0.0' in kibana.yml to prevent this message.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\",\"config\"],\"pid\":7,\"message\":\"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"actions\",\"actions\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:04Z\",\"tags\":[\"warning\",\"plugins\",\"alerts\",\"plugins\",\"alerting\"],\"pid\":7,\"message\":\"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\"],\"pid\":7,\"message\":\"config sourced from: production cluster\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:05Z\",\"tags\":[\"info\",\"savedobjects-service\"],\"pid\":7,\"message\":\"Starting saved objects migrations\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins-system\"],\"pid\":7,\"message\":\"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:42:04 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:06 Running ['artisan' dialers:monitor-activities] . 2s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:08 Running ['artisan' jiminny:monitor-social-accounts] 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:12 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:14 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:15 Running ['artisan' conference:monitor:count] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:16 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:42:20 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:42:20 Jiminny\\Jobs\\Mailbox\\CreateBatches ........ 44.79ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:43:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:06 Running ['artisan' jiminny:monitor-social-accounts] 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:11 Running ['artisan' mailbox:skip-lists:refresh] . 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:14 Running ['artisan' mailbox:batch:process --max-batches=15] 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 12.38ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' calendar:sync --dateMode=daily] 2026-04-20 06:43:31 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:32 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... 1s DONE\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\ndocker_lamp_1 | 14s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' calendar:sync --dateMode=daily > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents . 324.40ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\n\n\nv View in Docker Desktop o View Config w Enable Watch","is_focused":true},{"role":"AXButton","text":"Menu","depth":3,"bounds":{"left":0.48402777,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥1 DOCKER (docker-compose)","depth":3,"bounds":{"left":0.01875,"top":0.09,"width":0.46180555,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥2 PROD (ssh)","depth":4,"bounds":{"left":0.51805556,"top":0.09,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98541665,"top":0.24944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥3 EU (ssh)","depth":4,"bounds":{"left":0.51805556,"top":0.25,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.49930555,"top":0.4422222,"width":0.50069445,"height":0.12222222},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true}]...
|
5696478135866045708
|
-8341755017428807922
|
app_switch
|
accessibility
|
NULL
|
kibana | {"type":"log", kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_ROOT_USERNAME=root
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local
docker_lamp_1 | + declare IP_ADDRESS
docker_lamp_1 | ++ curl ifconfig.co/
docker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current
docker_lamp_1 | Dload Upload Total Spent Left Speed
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","encryptedSavedObjects","config"],"pid":7,"message":"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","ingestManager"],"pid":7,"message":"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224
docker_lamp_1 | + IP_ADDRESS=[IP_ADDRESS]
docker_lamp_1 | + readonly IP_ADDRESS
docker_lamp_1 | + sed -i -E 's~X-Forwarded-For "\$remote_addr"~X-Forwarded-For "IP_ADDRESS"~g' /etc/nginx/nginx.conf
docker_lamp_1 | + sed -i -E s~IP_ADDRESS~[IP_ADDRESS]~g /etc/nginx/nginx.conf
docker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \{([^#]*)# QA Only End \}//g' /etc/nginx/nginx.conf
docker_lamp_1 | + storage_init
docker_lamp_1 | + false
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/services.php ]]
docker_lamp_1 | ++ find /home/jiminny/jiminny_storage/ -maxdepth 1
docker_lamp_1 | ++ wc -l
docker_lamp_1 | + [[ 11 -lt 7 ]]
docker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework
docker_lamp_1 | + [[ 775 -ne 775 ]]
docker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/cache
docker_lamp_1 | + [[ 775 -ne 775 ]]
docker_lamp_1 | ++ stat -c %a /home/jiminny/jiminny_storage/framework/views
docker_lamp_1 | + [[ 775 -ne 775 ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/jiminny_storage/oauth-private.key ]]
docker_lamp_1 | + sleep 5
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","reporting","config"],"pid":7,"message":"Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","reporting","config"],"pid":7,"message":"Found 'server.host: \"0\"' in Kibana configuration. This is incompatible with Reporting. To enable Reporting to work, 'xpack.reporting.kibanaServer.hostname: [IP_ADDRESS]' is being automatically to the configuration. You can change the setting to 'server.host: [IP_ADDRESS]' or add 'xpack.reporting.kibanaServer.hostname: [IP_ADDRESS]' in kibana.yml to prevent this message."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","reporting","config"],"pid":7,"message":"Chromium sandbox provides an additional layer of protection, but is not supported for Linux CentOS 8.3.2011\n OS. Automatically setting 'xpack.reporting.capture.browser.chromium.disableSandbox: true'."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","actions","actions"],"pid":7,"message":"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","alerts","plugins","alerting"],"pid":7,"message":"APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:05Z","tags":["info","plugins","monitoring","monitoring"],"pid":7,"message":"config sourced from: production cluster"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:05Z","tags":["info","savedobjects-service"],"pid":7,"message":"Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations..."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:05Z","tags":["info","savedobjects-service"],"pid":7,"message":"Starting saved objects migrations"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins-system"],"pid":7,"message":"Starting [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins","taskManager","taskManager"],"pid":7,"message":"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a"}
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:40:06,605Z", "level": "INFO", "component": "o.e.c.m.MetadataIndexTemplateService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "adding template [.management-beats] for index patterns [.management-beats]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
kibana | {"type":"log","@timestamp":"20...
|
NULL
|
|
52173
|
1129
|
9
|
2026-04-20T06:43:39.759445+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667419759_m2.jpg...
|
iTerm2
|
EU (ssh)
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
FV faVsco.js vcode#11894 on JY-18909-automated-rep FV faVsco.js vcode#11894 on JY-18909-automated-reports-ask-jProjectyE ids.txtEinfection.json.distM+INSlALL.moM-INTckNAL webnook sclur.moiiminny storageM+ licenses.mdM MakerileO package-lock.json= phostan.neon.distE phpstan-baseline.neon< ohounit xmllTo raw_sqL_query.sqlM4 README mdsonar-project.properties= test.py<> Untitled Diagram.xmlus vetur.config.jsM+ WEBHOOK_FILTERING_IMPLEMENTATK›db External Librariesv =°Scratches and Consolesv M Datahase ConsolesV AEU& console cullA DEAL RISKS [EU]A DI (EU]Msuisin© Service.phpo riela.pnpyrelakepository.ohp© Playbook.php© ReportController.php= custom.logA console (STAGING]E laravel.logA SF [jiminny@localhost]A HS_Jocal [jiminny@localhost]A console (PROD]A console [EU] X16031© CreatereldActuivitvevent.ohoc TrackProviderinstalledzvent.ono© ActivityLoqqed.php(C) CreateActivityLoagedEvent.phpC) UserPilotActivityListener.phoC) AutomatedReportsCallbackService.php© RequestGenerateAskJiminnyReportJob.php(C) RequestGenerateReportJob.php1605160616071608Tx: AutoPlavaroundJOIN users u on u.id = sa.sociable_idJOIN teams t1..N<->1: on t.1d = U.team 1dWHERE u. team_id = 459 and sa.provider ='hubspot';dojiminny v027 49 A23 X3 X 105 AC) AutomatedReportResult.ohdC) AutomatedReport.ohoclass TrackAutomatedReportGeneratedEventTest extends TestCasetestHandLeTracksCreatorForAskJiminnyReport(): voids: Screator,aucomared-report-generaced1610SELECT os.stage_id, s.crm_provider_id, s.name, COUNT(*) as cntFROM opportunity_stages osJOIN stages s 1.n<->1: ON s.id = os.stage_idWHERE Os.opportunity_id = 7594349GROUP BY os.stage_id, s.crm_provider_id, s.nameORDER BY cnt DESC;Reject16121f26131614'ask-jiminny-report-generated",1615['report_type' => AutomatedReportsService::TYPE_ASK_JIMINNY, 'frequency' => 'weekly']161611617$listener = $this->makeListener();ustener.?nand lelsuuus-lakecvent sredonuee1618161911620116211622 VSELECT s.id, s.crm_provider_id, s.name, s.team_id, s.crm_configuration_idrkon scages sJOIN business_process_stages bps 1<-›1..n: ON bps.stage_id = s.idwHEkE bos.bus1ness_process 1d = 6024AND s.crm_provider_id = 'contractsent';select * from stages where id IN (16352,20612,18281, 7344, 16378, 16309, 5036, 15223, 14535, 6293, 12098, 11607).public function testHandleSkipsTrackingWhenAskJiminnyCreatorIsNulL(): voidSELECT * FROM teams WHERE name LIKE '%Pulsar Group%': # 472. 380. 15138. [EMAIL] * from plavbooks where team 1d = 412: # event 226147SELECT * FROM playbook_categories WHERE playbook_id = 2288;SELECT * FROMIcrm Fields WHERE10 = 226147:Ask anvthina (*4L)AdaptiveServices+,o,c|v Database4 console 1 s 509 ms#liminnv@localhostA SF#HS locallV A PROD#concaleV A STAGINGA consoleg DockerOutputtil# 472,380, 15138, [EMAIL] xTx22 rowsv50,0Did Ywuid CUUID with time-low and time-..2072 dechfc0k-6412-40a3-055c-85407057d12h12169 d6f9a38f-936e-4be8-86b1-5e0be1ed283b2176 24fe7887-47a0-4504-9a1f-8b2d9d15be022177 b4996f98-33a0-4dcf-85a8-73f58aa12c2d2178 8defle23-a5ad-458b-857b-d892afdcdadfi2179 02e1hf6d-a8c6-4972-ad68-c007F0f<1ffd2180 33/05073-4392-4711-a268-6097442759f72287 631c1554-ae45-4bcf-9685-0ad4b12e56702288 5618c7a6-c98f-4298-8511-97c435a88ab72289 340ff88а-7963-46bf-a9ce-1a34d7632chcl1200 001c/640-6038-44df-01h5-d52c007400062291 526bec4f-777c-4bff-ad76-2ee44ab3bb23# Q G 6 4ĐT : team_id T÷ Dactivity_type Y472 event472 task472 task472 task472 task472 task472 task472 task472 task472 task472 event472 task÷ @activity_field_id Y+name 7189970 test189962 AM Test189962 Vuelio AM's - Brands189962 Vuelio AM's - PS189962 VueLio AM's- Agency189962 Vuelio AM's -• NE189962 Vuelio AM'S - PS189962 Vuelio Insights Team189962 Vuelio Insights Team189962 Test Task Playbook180070 Test Event Pavhook189962 Vuelio Insights Team÷ is_selectable YWai_activity_type_detection_enabled YDally - Platrorm • In 2m100% S2MOn ZU AOr 9.43.39eporceneratedeventlestvCascadeRetactor Connection• Retactor UserPilot 1coverage 2>&1 | head -100vee runcale exec lanteu. exec Tatteu.vendor/e inphuaited: exec /aeled: binbphpunistant such rate orodiset erec:Checked command statusinougntsCommand docker. headbash: line 1: •/vendor/bin/phpunit: No such file or directoryChecked command status+0 ..I created_at Y0 2025-09-15 12:50:250 2025-12-08 16:24:58© 2025-12-12 12:31:500 2025-12-12 13:17:320 2025-12-12 13:23:1702025-12-12 13-24-320 2025-12-12 13:24:570 2026-04-16 10:29:520 2026-04-16 10:36:550 2026-04-17 13:13:430 2026-04-17 13:14:0102026-94-20 06-37-59I updated_at Y2025-09-15 12-59•342025-12-12 16:24:552025-12-12 12:31:502025-12-12 13:23:362025-12-12 13:23:172025-12-12 13:24-322025-12-12 13:24:572026-04-16 10:36:422026-04-16 10:37:032026-04-17 13:14:242026-04-17 13:14:182026-04-20 06-40•38®Reject allAccept all® :-CSVvI deleted_at Y2025-00-15 12•50•34<null>2025-12-12 13:23:36<null><null>2026-04-16 10:36:422026-04-16 10:37:032026-04-17 13:14:242026-04-17 13:14:182026-04-20 06-40•3822 rows retrieved starting from 1 in 845 ms (execution: 135 ms, fetching: 710 ms)W Windsurf Teams1622:20 UTF-84 spaces...
|
NULL
|
5690775399036335859
|
NULL
|
app_switch
|
ocr
|
NULL
|
FV faVsco.js vcode#11894 on JY-18909-automated-rep FV faVsco.js vcode#11894 on JY-18909-automated-reports-ask-jProjectyE ids.txtEinfection.json.distM+INSlALL.moM-INTckNAL webnook sclur.moiiminny storageM+ licenses.mdM MakerileO package-lock.json= phostan.neon.distE phpstan-baseline.neon< ohounit xmllTo raw_sqL_query.sqlM4 README mdsonar-project.properties= test.py<> Untitled Diagram.xmlus vetur.config.jsM+ WEBHOOK_FILTERING_IMPLEMENTATK›db External Librariesv =°Scratches and Consolesv M Datahase ConsolesV AEU& console cullA DEAL RISKS [EU]A DI (EU]Msuisin© Service.phpo riela.pnpyrelakepository.ohp© Playbook.php© ReportController.php= custom.logA console (STAGING]E laravel.logA SF [jiminny@localhost]A HS_Jocal [jiminny@localhost]A console (PROD]A console [EU] X16031© CreatereldActuivitvevent.ohoc TrackProviderinstalledzvent.ono© ActivityLoqqed.php(C) CreateActivityLoagedEvent.phpC) UserPilotActivityListener.phoC) AutomatedReportsCallbackService.php© RequestGenerateAskJiminnyReportJob.php(C) RequestGenerateReportJob.php1605160616071608Tx: AutoPlavaroundJOIN users u on u.id = sa.sociable_idJOIN teams t1..N<->1: on t.1d = U.team 1dWHERE u. team_id = 459 and sa.provider ='hubspot';dojiminny v027 49 A23 X3 X 105 AC) AutomatedReportResult.ohdC) AutomatedReport.ohoclass TrackAutomatedReportGeneratedEventTest extends TestCasetestHandLeTracksCreatorForAskJiminnyReport(): voids: Screator,aucomared-report-generaced1610SELECT os.stage_id, s.crm_provider_id, s.name, COUNT(*) as cntFROM opportunity_stages osJOIN stages s 1.n<->1: ON s.id = os.stage_idWHERE Os.opportunity_id = 7594349GROUP BY os.stage_id, s.crm_provider_id, s.nameORDER BY cnt DESC;Reject16121f26131614'ask-jiminny-report-generated",1615['report_type' => AutomatedReportsService::TYPE_ASK_JIMINNY, 'frequency' => 'weekly']161611617$listener = $this->makeListener();ustener.?nand lelsuuus-lakecvent sredonuee1618161911620116211622 VSELECT s.id, s.crm_provider_id, s.name, s.team_id, s.crm_configuration_idrkon scages sJOIN business_process_stages bps 1<-›1..n: ON bps.stage_id = s.idwHEkE bos.bus1ness_process 1d = 6024AND s.crm_provider_id = 'contractsent';select * from stages where id IN (16352,20612,18281, 7344, 16378, 16309, 5036, 15223, 14535, 6293, 12098, 11607).public function testHandleSkipsTrackingWhenAskJiminnyCreatorIsNulL(): voidSELECT * FROM teams WHERE name LIKE '%Pulsar Group%': # 472. 380. 15138. [EMAIL] * from plavbooks where team 1d = 412: # event 226147SELECT * FROM playbook_categories WHERE playbook_id = 2288;SELECT * FROMIcrm Fields WHERE10 = 226147:Ask anvthina (*4L)AdaptiveServices+,o,c|v Database4 console 1 s 509 ms#liminnv@localhostA SF#HS locallV A PROD#concaleV A STAGINGA consoleg DockerOutputtil# 472,380, 15138, [EMAIL] xTx22 rowsv50,0Did Ywuid CUUID with time-low and time-..2072 dechfc0k-6412-40a3-055c-85407057d12h12169 d6f9a38f-936e-4be8-86b1-5e0be1ed283b2176 24fe7887-47a0-4504-9a1f-8b2d9d15be022177 b4996f98-33a0-4dcf-85a8-73f58aa12c2d2178 8defle23-a5ad-458b-857b-d892afdcdadfi2179 02e1hf6d-a8c6-4972-ad68-c007F0f<1ffd2180 33/05073-4392-4711-a268-6097442759f72287 631c1554-ae45-4bcf-9685-0ad4b12e56702288 5618c7a6-c98f-4298-8511-97c435a88ab72289 340ff88а-7963-46bf-a9ce-1a34d7632chcl1200 001c/640-6038-44df-01h5-d52c007400062291 526bec4f-777c-4bff-ad76-2ee44ab3bb23# Q G 6 4ĐT : team_id T÷ Dactivity_type Y472 event472 task472 task472 task472 task472 task472 task472 task472 task472 task472 event472 task÷ @activity_field_id Y+name 7189970 test189962 AM Test189962 Vuelio AM's - Brands189962 Vuelio AM's - PS189962 VueLio AM's- Agency189962 Vuelio AM's -• NE189962 Vuelio AM'S - PS189962 Vuelio Insights Team189962 Vuelio Insights Team189962 Test Task Playbook180070 Test Event Pavhook189962 Vuelio Insights Team÷ is_selectable YWai_activity_type_detection_enabled YDally - Platrorm • In 2m100% S2MOn ZU AOr 9.43.39eporceneratedeventlestvCascadeRetactor Connection• Retactor UserPilot 1coverage 2>&1 | head -100vee runcale exec lanteu. exec Tatteu.vendor/e inphuaited: exec /aeled: binbphpunistant such rate orodiset erec:Checked command statusinougntsCommand docker. headbash: line 1: •/vendor/bin/phpunit: No such file or directoryChecked command status+0 ..I created_at Y0 2025-09-15 12:50:250 2025-12-08 16:24:58© 2025-12-12 12:31:500 2025-12-12 13:17:320 2025-12-12 13:23:1702025-12-12 13-24-320 2025-12-12 13:24:570 2026-04-16 10:29:520 2026-04-16 10:36:550 2026-04-17 13:13:430 2026-04-17 13:14:0102026-94-20 06-37-59I updated_at Y2025-09-15 12-59•342025-12-12 16:24:552025-12-12 12:31:502025-12-12 13:23:362025-12-12 13:23:172025-12-12 13:24-322025-12-12 13:24:572026-04-16 10:36:422026-04-16 10:37:032026-04-17 13:14:242026-04-17 13:14:182026-04-20 06-40•38®Reject allAccept all® :-CSVvI deleted_at Y2025-00-15 12•50•34<null>2025-12-12 13:23:36<null><null>2026-04-16 10:36:422026-04-16 10:37:032026-04-17 13:14:242026-04-17 13:14:182026-04-20 06-40•3822 rows retrieved starting from 1 in 845 ms (execution: 135 ms, fetching: 710 ms)W Windsurf Teams1622:20 UTF-84 spaces...
|
NULL
|
|
52196
|
1129
|
16
|
2026-04-20T06:45:04.669623+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667504669_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
rirerox••0!IY [SRD-6793] Les Mills activity t XJY- rirerox••0!IY [SRD-6793] Les Mills activity t XJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tab1 Useroilot I EventsPronllesBookmarksQ Search bookmarksv a bookmarks loolba• Sprint Board# SRD QueueGithub8 Jiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....v # Bookmarks Menu> → Mozilla Firefox#Other BookmarksI JIMINNYg For you• Recent# Starred0+ AppsQ Spaces+ ...Jiminny (New)ID Platform Teamii SE KanbanŒI Capture TeamI Enterprise Stability I...Processing Team( Service-Desk•, QueuesE Service requestsA Incidentslul Reports© OperationsE Knowledge Base& Customers• Channels• Email logs|%› Developer escalations: Slack integration& Reporting Center[ Add shortcut# Archived work items= More spaces= FiltersI DashboardsOperations&& Confiuence:: Teams'= Customise sidebar•y0 liil oj Daily - Platform • nowQ Search+ Create100% L2• Mon ZU AOr 9.40:04Ask RovoSpaces / F7 Service-Desk / #F SRD-6793Steps to reproduceCreate a new playbook in Les Mills's instanceCustomer typeEnterpriseActual outcomeNo activity types are being imported upon new Playbook creationExpected outcomeThe activity types to get importedSeverity level(S2ImpactNoneRoot causeVia have no access to FieldDefinition (needed to sync activity types).Linked work itemsrelates to** JY-20698 Les Mills activity types not pulling inActivityAll Comments History Work log Approvals@ Add internal note Reply to customerNormal text v B I Mv m z z ₴ z₴ +v 0WiteType /ai for Atlassian Intelligence or @ to mention and notify someone.0. Passina to riaht team|| 0. Pronose solution|| A. Reauest more info|| 0. Suaaest replv!CancellStoyan Tomov 4 days ago & Internal noteFYI, Pulsar Group (EU) is facing the same issue.€ • Edit • Delete1CODE REVIEWo Summarise comments[ Canned responsesIn ProgressDetailsAssignee@ Lukas Kovalikkevorier@ Stoyan TomovRequest Type# Report a bugKnowledge base View relate..Priority levelP2 MediumDev TeamPlatform teamOrganization1oc MillcCanny LinksOpen Canny Links> More fields Labels. Time tracking, Tvpe of I...> Automation 4 Rule executions> featureOS Open featureOSIntercomShowing 1 out of 1 linked conversationsQ Christopher Wilton ? =..>> Sentry ell! Linked Issuessôr Confiaure...
|
NULL
|
-3779784927519369284
|
NULL
|
click
|
ocr
|
NULL
|
rirerox••0!IY [SRD-6793] Les Mills activity t XJY- rirerox••0!IY [SRD-6793] Les Mills activity t XJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tab1 Useroilot I EventsPronllesBookmarksQ Search bookmarksv a bookmarks loolba• Sprint Board# SRD QueueGithub8 Jiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....v # Bookmarks Menu> → Mozilla Firefox#Other BookmarksI JIMINNYg For you• Recent# Starred0+ AppsQ Spaces+ ...Jiminny (New)ID Platform Teamii SE KanbanŒI Capture TeamI Enterprise Stability I...Processing Team( Service-Desk•, QueuesE Service requestsA Incidentslul Reports© OperationsE Knowledge Base& Customers• Channels• Email logs|%› Developer escalations: Slack integration& Reporting Center[ Add shortcut# Archived work items= More spaces= FiltersI DashboardsOperations&& Confiuence:: Teams'= Customise sidebar•y0 liil oj Daily - Platform • nowQ Search+ Create100% L2• Mon ZU AOr 9.40:04Ask RovoSpaces / F7 Service-Desk / #F SRD-6793Steps to reproduceCreate a new playbook in Les Mills's instanceCustomer typeEnterpriseActual outcomeNo activity types are being imported upon new Playbook creationExpected outcomeThe activity types to get importedSeverity level(S2ImpactNoneRoot causeVia have no access to FieldDefinition (needed to sync activity types).Linked work itemsrelates to** JY-20698 Les Mills activity types not pulling inActivityAll Comments History Work log Approvals@ Add internal note Reply to customerNormal text v B I Mv m z z ₴ z₴ +v 0WiteType /ai for Atlassian Intelligence or @ to mention and notify someone.0. Passina to riaht team|| 0. Pronose solution|| A. Reauest more info|| 0. Suaaest replv!CancellStoyan Tomov 4 days ago & Internal noteFYI, Pulsar Group (EU) is facing the same issue.€ • Edit • Delete1CODE REVIEWo Summarise comments[ Canned responsesIn ProgressDetailsAssignee@ Lukas Kovalikkevorier@ Stoyan TomovRequest Type# Report a bugKnowledge base View relate..Priority levelP2 MediumDev TeamPlatform teamOrganization1oc MillcCanny LinksOpen Canny Links> More fields Labels. Time tracking, Tvpe of I...> Automation 4 Rule executions> featureOS Open featureOSIntercomShowing 1 out of 1 linked conversationsQ Christopher Wilton ? =..>> Sentry ell! Linked Issuessôr Confiaure...
|
52195
|
|
52197
|
1128
|
24
|
2026-04-20T06:45:04.769247+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667504769_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
FirefoxFileEditView→HistoryBookmarksProfilesToolsW FirefoxFileEditView→HistoryBookmarksProfilesToolsWindowHelpalol§ Daily - Platform • nowmeet.google.com/agt-teir-cwt?authuser=lukas.kovalik%40jiminny.com100% 0 8•Mon 20 Apr 9:45:05C Meet - Daily - Platform+ New TabBookmarksQ Search bookmarksXv [* Bookmarks ToolbarSprint Board• SRD QueueGithub8 Jiminny DEVAsk Jiminny Reports by nikolay-yankov - Pull ...v E Bookmarks Menu> [ Mozilla FirefoxOther [EMAIL] accountLukas Kovalikf s...AS... •I F...-[ B... -Daily - PlatformSteliyan Georgiev is in this call* Use Gemini to take notesShare notes and transcriptStartJoin nowOther ways to join~cce nodado naten on peroea in uanta are e n are coenano i mроnку...
|
NULL
|
-4234584306121702301
|
NULL
|
click
|
ocr
|
NULL
|
FirefoxFileEditView→HistoryBookmarksProfilesToolsW FirefoxFileEditView→HistoryBookmarksProfilesToolsWindowHelpalol§ Daily - Platform • nowmeet.google.com/agt-teir-cwt?authuser=lukas.kovalik%40jiminny.com100% 0 8•Mon 20 Apr 9:45:05C Meet - Daily - Platform+ New TabBookmarksQ Search bookmarksXv [* Bookmarks ToolbarSprint Board• SRD QueueGithub8 Jiminny DEVAsk Jiminny Reports by nikolay-yankov - Pull ...v E Bookmarks Menu> [ Mozilla FirefoxOther [EMAIL] accountLukas Kovalikf s...AS... •I F...-[ B... -Daily - PlatformSteliyan Georgiev is in this call* Use Gemini to take notesShare notes and transcriptStartJoin nowOther ways to join~cce nodado naten on peroea in uanta are e n are coenano i mроnку...
|
52194
|
|
52200
|
1128
|
26
|
2026-04-20T06:45:08.452207+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667508452_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Firefox•FileEditViewHistoryBookmarks→ CProfilesToo Firefox•FileEditViewHistoryBookmarks→ CProfilesToolsWindowHelp>0 klmeet.google.com/agt-teir-cwt?authuser=lukas.kovalik%40jiminny.comDaily - Platform • nowБГ100% & 8• Mon 20 Apr 9:45:081 [EMAIL] account+Lukas KovalikDaily - PlatformNikolay Nikolov, Nikolay Yankov and Steliyan Georgiev are in thiscall. Use Gemini to take notesShare notes and transcriptStartJoin nowOther ways to join{ soundcore A...4 System Defa...• FaceTime HD...g Backgrounds...Gemini is available in Meet as your personal in-meeting assistant. It can analyze conversation via temporaryaccess to meeting captions. Using Ask Gemini won't create a recording or store meeting data. The meetinghost can turn it off. Learn moreLộ3...
|
NULL
|
-1578154115932039129
|
NULL
|
click
|
ocr
|
NULL
|
Firefox•FileEditViewHistoryBookmarks→ CProfilesToo Firefox•FileEditViewHistoryBookmarks→ CProfilesToolsWindowHelp>0 klmeet.google.com/agt-teir-cwt?authuser=lukas.kovalik%40jiminny.comDaily - Platform • nowБГ100% & 8• Mon 20 Apr 9:45:081 [EMAIL] account+Lukas KovalikDaily - PlatformNikolay Nikolov, Nikolay Yankov and Steliyan Georgiev are in thiscall. Use Gemini to take notesShare notes and transcriptStartJoin nowOther ways to join{ soundcore A...4 System Defa...• FaceTime HD...g Backgrounds...Gemini is available in Meet as your personal in-meeting assistant. It can analyze conversation via temporaryaccess to meeting captions. Using Ask Gemini won't create a recording or store meeting data. The meetinghost can turn it off. Learn moreLộ3...
|
52198
|
|
52201
|
1129
|
18
|
2026-04-20T06:45:08.550827+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667508550_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
rirerox••0!IY [SRD-6793] Les Mills activity t XJY- rirerox••0!IY [SRD-6793] Les Mills activity t XJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tab1 Useroilot I EventsPronllesBookmarksQ Search bookmarksv a bookmarks loolba• Sprint Board# SRD QueueGithub8 Jiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....v # Bookmarks Menu> → Mozilla Firefox#Other BookmarksI JIMINNYg For you• Recent# Starred0+ AppsQ Spaces+ ...Jiminny (New)ID Platform Teamii SE KanbanŒI Capture TeamI Enterprise Stability I...Processing Team( Service-Desk•, QueuesE Service requestsA Incidentslul Reports© OperationsE Knowledge Base& Customers• Channels• Email logs|%› Developer escalations: Slack integration& Reporting Center[ Add shortcut# Archived work items= More spaces= FiltersI DashboardsOperations&& Confiuence:: Teams'= Customise sidebar•y0 liil oj Daily - Platform • nowQ Search+ Create100% L2• Mon ZU AOr 9.40:00Ask RovoSpaces / F7 Service-Desk / #F SRD-6793Steps to reproduceCreate a new playbook in Les Mills's instanceCustomer typeEnterpriseActual outcomeNo activity types are being imported upon new Playbook creationExpected outcomeThe activity types to get importedSeverity level(S2ImpactNoneRoot causeVia have no access to FieldDefinition (needed to sync activity types).Linked work itemsrelates to** JY-20698 Les Mills activity types not pulling inActivityAll Comments History Work log Approvals@ Add internal note Reply to customerNormal text v B I Mv m z z ₴ z₴ +v 0WiteType /ai for Atlassian Intelligence or @ to mention and notify someone.0. Passina to riaht team|| 0. Pronose solution|| A. Reauest more info|| 0. Suaaest replv!CancellStoyan Tomov 4 days ago & Internal noteFYI, Pulsar Group (EU) is facing the same issue.€ • Edit • Delete1CODE REVIEWo Summarise comments[ Canned responsesIn ProgressDetailsAssignee@ Lukas Kovalikkevorier@ Stoyan TomovRequest Type# Report a bugKnowledge base View relate..Priority levelP2 MediumDev TeamPlatform teamOrganization1oc MillcCanny LinksOpen Canny Links> More fields Labels. Time tracking, Tvpe of I...> Automation 4 Rule executions> featureOS Open featureOSIntercomShowing 1 out of 1 linked conversationsQ Christopher Wilton ? =..>> Sentry ell! Linked Issuessôr Confiaure...
|
NULL
|
1270557903591530571
|
NULL
|
click
|
ocr
|
NULL
|
rirerox••0!IY [SRD-6793] Les Mills activity t XJY- rirerox••0!IY [SRD-6793] Les Mills activity t XJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tab1 Useroilot I EventsPronllesBookmarksQ Search bookmarksv a bookmarks loolba• Sprint Board# SRD QueueGithub8 Jiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....v # Bookmarks Menu> → Mozilla Firefox#Other BookmarksI JIMINNYg For you• Recent# Starred0+ AppsQ Spaces+ ...Jiminny (New)ID Platform Teamii SE KanbanŒI Capture TeamI Enterprise Stability I...Processing Team( Service-Desk•, QueuesE Service requestsA Incidentslul Reports© OperationsE Knowledge Base& Customers• Channels• Email logs|%› Developer escalations: Slack integration& Reporting Center[ Add shortcut# Archived work items= More spaces= FiltersI DashboardsOperations&& Confiuence:: Teams'= Customise sidebar•y0 liil oj Daily - Platform • nowQ Search+ Create100% L2• Mon ZU AOr 9.40:00Ask RovoSpaces / F7 Service-Desk / #F SRD-6793Steps to reproduceCreate a new playbook in Les Mills's instanceCustomer typeEnterpriseActual outcomeNo activity types are being imported upon new Playbook creationExpected outcomeThe activity types to get importedSeverity level(S2ImpactNoneRoot causeVia have no access to FieldDefinition (needed to sync activity types).Linked work itemsrelates to** JY-20698 Les Mills activity types not pulling inActivityAll Comments History Work log Approvals@ Add internal note Reply to customerNormal text v B I Mv m z z ₴ z₴ +v 0WiteType /ai for Atlassian Intelligence or @ to mention and notify someone.0. Passina to riaht team|| 0. Pronose solution|| A. Reauest more info|| 0. Suaaest replv!CancellStoyan Tomov 4 days ago & Internal noteFYI, Pulsar Group (EU) is facing the same issue.€ • Edit • Delete1CODE REVIEWo Summarise comments[ Canned responsesIn ProgressDetailsAssignee@ Lukas Kovalikkevorier@ Stoyan TomovRequest Type# Report a bugKnowledge base View relate..Priority levelP2 MediumDev TeamPlatform teamOrganization1oc MillcCanny LinksOpen Canny Links> More fields Labels. Time tracking, Tvpe of I...> Automation 4 Rule executions> featureOS Open featureOSIntercomShowing 1 out of 1 linked conversationsQ Christopher Wilton ? =..>> Sentry ell! Linked Issuessôr Confiaure...
|
52199
|
|
52202
|
1128
|
27
|
2026-04-20T06:45:10.242035+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776667510242_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
FirefoxFileEditViewHistoryBookmarksProfilesToolsWi FirefoxFileEditViewHistoryBookmarksProfilesToolsWindow→ CI MeetHelpalolmeet.google.com/agt-teir-cwt?authuser=lukas.kovalik%40jiminny.c+Lukas Kovalik:Daily - Platform • nowAll-In-OneCapture AreaCapture Previous Area• Capture FullscreenCapture Windowj Scrolling CaptureSelf-Timer AAa Capture Text (OCR)• Record ScreenHide Desktop IconsOpen...Pin to the Screen...Capture History...About CleanShot...Check for Updates...Settings...Quit100% <7*8 • Mon 20 Apr 9:45:100&5+*4+₴[EMAIL] accountPlatformand Steliyan Georgiev are in thisallcake notesranscriptStart|28,i nowOther ways to join{ soundcore A...4 System Defa...• FaceTime HD...g Backgrounds...Gemini is available in Meet as your personal in-meeting assistant. It can analyze conversation via temporaryaccess to meeting captions. Using Ask Gemini won't create a recording or store meeting data. The meetinghost can turn it off. Learn moreLộ3...
|
NULL
|
-4501979073449577389
|
NULL
|
visual_change
|
ocr
|
NULL
|
FirefoxFileEditViewHistoryBookmarksProfilesToolsWi FirefoxFileEditViewHistoryBookmarksProfilesToolsWindow→ CI MeetHelpalolmeet.google.com/agt-teir-cwt?authuser=lukas.kovalik%40jiminny.c+Lukas Kovalik:Daily - Platform • nowAll-In-OneCapture AreaCapture Previous Area• Capture FullscreenCapture Windowj Scrolling CaptureSelf-Timer AAa Capture Text (OCR)• Record ScreenHide Desktop IconsOpen...Pin to the Screen...Capture History...About CleanShot...Check for Updates...Settings...Quit100% <7*8 • Mon 20 Apr 9:45:100&5+*4+₴[EMAIL] accountPlatformand Steliyan Georgiev are in thisallcake notesranscriptStart|28,i nowOther ways to join{ soundcore A...4 System Defa...• FaceTime HD...g Backgrounds...Gemini is available in Meet as your personal in-meeting assistant. It can analyze conversation via temporaryaccess to meeting captions. Using Ask Gemini won't create a recording or store meeting data. The meetinghost can turn it off. Learn moreLộ3...
|
NULL
|
|
52460
|
1132
|
82
|
2026-04-20T07:03:59.465642+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776668639465_m1.jpg...
|
iTerm2
|
EU (ssh)
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
FirefoxFileEditViewHistoryBookmarksProfilesToolsWi FirefoxFileEditViewHistoryBookmarksProfilesToolsWindowHelp(ahol§ Daily - Platform • 2 m left→meet.google.com/agt-teir-cwt?authuser=lukas.kovalik%40jiminny.comNikolay Yankov (Presenting)Edit® For you© Recent|# Starred8? Apps(ID Processing TeamIID SE Kanban= More spaces= FitersCB Dashboards€ Operations|20& Confluence38 Team.5 Customise sidebarC ASkSFа жC fot xProjects2) SSHDatadosQ Search |Spaces / Jiminny Ovea) / 0 Jy-15971 / A Jy-20500oluonPS$IPhpStormLinked work itemsclonerJY-20200 Batch update deals, leads, accounts and contacts for Salestorceis cloned byA JY-20501 Batch Initial sync for Hubspot |ActivityWork logAdd a comment.….Who is working on this.?Pro tip: press M to corl10:03 AM | Daily - PlatformPro X2 ( x0 GF Pro, X8-Mon 20 Apr 10:0312 (YC Tasks |0 Dev+ CreateAsk RovoBacklog~# Improve Story|DetailsDEPLOYED@ Open with VS CodeGreate comeComponentsTieounSub-ProductAdd optionsLabelsNone100% <78• Mon 20 Apr 10:03:592a 7Nikolay NikolovGalya DimitrovaNikolay YankovLukas Kovalik18:40...
|
NULL
|
6354512290029311017
|
NULL
|
visual_change
|
ocr
|
NULL
|
FirefoxFileEditViewHistoryBookmarksProfilesToolsWi FirefoxFileEditViewHistoryBookmarksProfilesToolsWindowHelp(ahol§ Daily - Platform • 2 m left→meet.google.com/agt-teir-cwt?authuser=lukas.kovalik%40jiminny.comNikolay Yankov (Presenting)Edit® For you© Recent|# Starred8? Apps(ID Processing TeamIID SE Kanban= More spaces= FitersCB Dashboards€ Operations|20& Confluence38 Team.5 Customise sidebarC ASkSFа жC fot xProjects2) SSHDatadosQ Search |Spaces / Jiminny Ovea) / 0 Jy-15971 / A Jy-20500oluonPS$IPhpStormLinked work itemsclonerJY-20200 Batch update deals, leads, accounts and contacts for Salestorceis cloned byA JY-20501 Batch Initial sync for Hubspot |ActivityWork logAdd a comment.….Who is working on this.?Pro tip: press M to corl10:03 AM | Daily - PlatformPro X2 ( x0 GF Pro, X8-Mon 20 Apr 10:0312 (YC Tasks |0 Dev+ CreateAsk RovoBacklog~# Improve Story|DetailsDEPLOYED@ Open with VS CodeGreate comeComponentsTieounSub-ProductAdd optionsLabelsNone100% <78• Mon 20 Apr 10:03:592a 7Nikolay NikolovGalya DimitrovaNikolay YankovLukas Kovalik18:40...
|
52459
|
|
52461
|
1133
|
40
|
2026-04-20T07:04:00.484783+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776668640484_m2.jpg...
|
iTerm2
|
EU (ssh)
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
,apm,uptime]"}
kibana | {"type" ,apm,uptime]"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins","taskManager","taskManager"],"pid":7,"message":"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a"}
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:40:06,605Z", "level": "INFO", "component": "o.e.c.m.MetadataIndexTemplateService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "adding template [.management-beats] for index patterns [.management-beats]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins","crossClusterReplication"],"pid":7,"message":"Your basic license does not support crossClusterReplication. Please upgrade your license."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins","watcher"],"pid":7,"message":"Your basic license does not support watcher. Please upgrade your license."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins","monitoring","monitoring","kibana-monitoring"],"pid":7,"message":"Starting monitoring stats collection"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["listening","info"],"pid":7,"message":"Server running at [URL_WITH_CREDENTIALS] server running at [URL_WITH_CREDENTIALS] the Chromium sandbox provides an additional layer of protection."}
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:40:14,068Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][50] overhead, spent [362ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + trap _term SIGTERM
docker_lamp_1 | + super_pid=13640
docker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf
docker_lamp_1 | + nginx
docker_lamp_1 | + php-fpm
docker_lamp_1 | + wait 13640
docker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642
docker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections
docker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/artisan-schedule.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-1.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-2.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-3.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-4.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-5.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-analytics.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-audio.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-calendar.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-conferences.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-crm-sync.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-crm-update.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-download.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-emails.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-es-update.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-nudges.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker.conf" during parsing
docker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock
docker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized
docker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking
docker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640
docker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653
docker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654
docker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655
docker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656
docker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657
docker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658
docker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660
docker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664
docker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668
docker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669
docker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670
docker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672
docker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673
docker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674
docker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675
docker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676
docker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677
docker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679
docker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:39 Jiminny\Jobs\Mailbox\CreateBatches [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:40:40 Jiminny\Jobs\Mailbox\CreateBatches ....... 505.47ms DONE
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:42:04 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:06 Running ['artisan' dialers:monitor-activities] . 2s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:08 Running ['artisan' jiminny:monitor-social-accounts] 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:12 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:14 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:15 Running ['artisan' conference:monitor:count] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:16 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 | 2026-04-20 06:42:20 Jiminny\Jobs\Mailbox\CreateBatches [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:42:20 Jiminny\Jobs\Mailbox\CreateBatches [PASSWORD_DOTS] 44.79ms DONE
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:43:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:43:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:43:06 Running ['artisan' jiminny:monitor-social-accounts] 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:43:11 Running ['artisan' mailbox:skip-lists:refresh] . 3s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:43:14 Running ['artisan' mailbox:batch:process --max-batches=15] 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 12.38ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' calendar:sync --dateMode=daily] 2026-04-20 06:43:31 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... RUNNING
docker_lamp_1 | 2026-04-20 06:43:32 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... 1s DONE
docker_lamp_1 | 2026-04-20 06:43:34 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... RUNNING
docker_lamp_1 | 14s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' calendar:sync --dateMode=daily > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 | 2026-04-20 06:43:34 Jiminny\Jobs\Calendar\SyncCalendarEvents . 324.40ms DONE
docker_lamp_1 | 2026-04-20 06:43:34 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... RUNNING
docker_lamp_1 | 2026-04-20 06:43:43 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... 8s DONE
docker_lamp_1 | 2026-04-20 06:43:43 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:44 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 887.95ms DONE
docker_lamp_1 | 2026-04-20 06:43:44 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:45 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 902.35ms DONE
docker_lamp_1 | 2026-04-20 06:43:45 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:46 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 959.92ms DONE
docker_lamp_1 | 2026-04-20 06:43:46 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:47 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 855.24ms DONE
docker_lamp_1 | 2026-04-20 06:43:47 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:48 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 933.68ms DONE
docker_lamp_1 | 2026-04-20 06:43:48 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:48 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 952.67ms DONE
docker_lamp_1 | 2026-04-20 06:43:49 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:49 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 933.63ms DONE
docker_lamp_1 | 2026-04-20 06:43:49 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:51 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 1s DONE
docker_lamp_1 | 2026-04-20 06:43:51 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:52 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 1s DONE
docker_lamp_1 | 2026-04-20 06:43:52 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:53 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 1s DONE
docker_lamp_1 | 2026-04-20 06:43:53 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:54 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 986.58ms DONE
docker_lamp_1 | 2026-04-20 06:43:54 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:56 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 2s DONE
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:44:03 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:06 Running ['artisan' jiminny:monitor-social-accounts] 2s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:09 Running ['artisan' mailbox:skip-lists:refresh] . 2s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:11 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:13 Running ['artisan' conference:monitor:count] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:14 Running ['artisan' crm:sync-objects] Syncing objects for Salesforce Team since 2026-03-19 20:14:10
docker_lamp_1 | Syncing objects for Hubspot since 2026-04-18 13:14:12
docker_lamp_1 | Syncing objects for Edge Communications since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Burlington Textiles Corp of America since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Uber since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Vonage since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Nexmo since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for SF since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for NewAccount since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for TestV since 2026-02-17 15:05:47
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for KioskAccount since 2026-02-17 15:06:27
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Pipedrive, Inc. since 2026-04-08 18:14:13
docker_lamp_1 | Syncing objects for Copper since 2026-04-18 13:14:12
docker_lamp_1 | Syncing objects for Pipedrive External Test since 2026-04-08 18:14:16
docker_lamp_1 | Syncing objects for InsightSquared since 2026-02-17 15:09:59
docker_lamp_1 | Syncing objects for Close since 2026-04-18 13:14:13
docker_lamp_1 | Syncing objects for DeewanyAccount since 2026-02-19 13:27:51
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Aircall Demo since 2026-02-17 15:06:51
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05
docker_lamp_1 | Syncing objects for GoStudent UAT since 2026-02-17 15:09:59
docker_lamp_1 | Syncing objects for Horen test since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed Hamadi since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed Testing since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed jiminny since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for JustCall since 2026-02-17 15:07:41
docker_lamp_1 | Syncing objects for Testers Inc since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Copper Team since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for TheBestPlace Ever since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Horen's apartments since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for New Org Test since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Pipedrive test org since 2026-02-19 13:27:42
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for BigChairs Inc. since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for GL 500 since 2026-02-19 13:26:04
docker_lamp_1 | Syncing objects for Loren X since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Laravel Company 2024 since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Barbara Fuchs since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Twilio Video since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for My Test Account 3000 since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for test since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Test since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Test salesforce auto sync since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for test ogg auto sync since 2026-02-17 15:07:16
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Globo Tech since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Laravel 11 Company 2024 since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 |
docker_lamp_1 | Error
docker_lamp_1 |
docker_lamp_1 | Call to a member function getAttribute() on null
docker_lamp_1 |
docker_lamp_1 | at app/Console/Commands/Crm/SyncObjects.php:42
docker_lamp_1 | 38▕
docker_lamp_1 | 39▕ foreach ($teams as $team) {
docker_lamp_1 | 40▕ $config = $team->getCrmConfiguration();
docker_lamp_1 | 41▕
docker_lamp_1 | ➜ 42▕ if ($config->getAttribute('sync_objects') === false) {
docker_lamp_1 | 43▕ continue;
docker_lamp_1 | 44▕ }
docker_lamp_1 | 45▕
docker_lamp_1 | 46▕ $this->info(
docker_lamp_1 |
docker_lamp_1 | 1 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:36
docker_lamp_1 | Jiminny\Console\Commands\Crm\SyncObjects::handle()
docker_lamp_1 |
docker_lamp_1 | 2 vendor/laravel/framework/src/Illuminate/Container/Util.php:43
docker_lamp_1 | Illuminate\Container\BoundMethod::Illuminate\Container\{closure}()
docker_lamp_1 |
docker_lamp_1 | 3 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:96
docker_lamp_1 | Illuminate\Container\Util::unwrapIfClosure(Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 4 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:35
docker_lamp_1 | Illuminate\Container\BoundMethod::callBoundMethod(Object(Illuminate\Foundation\Application), Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 5 vendor/laravel/framework/src/Illuminate/Container/Container.php:799
docker_lamp_1 | Illuminate\Container\BoundMethod::call(Object(Illuminate\Foundation\Application), [])
docker_lamp_1 |
docker_lamp_1 | 6 vendor/laravel/framework/src/Illuminate/Console/Command.php:211
docker_lamp_1 | Illuminate\Container\Container::call()
docker_lamp_1 |
docker_lamp_1 | 7 vendor/symfony/console/Command/Command.php:341
docker_lamp_1 | Illuminate\Console\Command::execute(Object(Symfony\Component\Console\Input\ArgvInput), Object(Illuminate\Console\OutputStyle))
docker_lamp_1 |
docker_lamp_1 | 8 vendor/laravel/framework/src/Illuminate/Console/Command.php:180
docker_lamp_1 | Symfony\Component\Console\Command\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Illuminate\Console\OutputStyle))
docker_lamp_1 |
docker_lamp_1 | 9 app/Console/Commands/Command.php:41
docker_lamp_1 | Illuminate\Console\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 10 vendor/symfony/console/Application.php:1117
docker_lamp_1 | Jiminny\Console\Commands\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 11 vendor/symfony/console/Application.php:356
docker_lamp_1 | Symfony\Component\Console\Application::doRunCommand(Object(Jiminny\Console\Commands\Crm\SyncObjects), Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 12 vendor/symfony/console/Application.php:195
docker_lamp_1 | Symfony\Component\Console\Application::doRun(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 13 vendor/laravel/framework/src/Illuminate/Foundation/Console/Kernel.php:198
docker_lamp_1 | Symfony\Component\Console\Application::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 14 vendor/laravel/framework/src/Illuminate/Foundation/Application.php:1235
docker_lamp_1 | Illuminate\Foundation\Console\Kernel::handle(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 15 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
docker_lamp_1 | [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:sync-objects > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 | 2026-04-20 06:44:15 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:15 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 98.77ms DONE
docker_lamp_1 | 2026-04-20 06:44:15 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:17 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-20 06:44:17 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:17 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 14.35ms DONE
docker_lamp_1 | 2026-04-20 06:44:17 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 14.13ms DONE
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 10.93ms DONE
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 13.44ms DONE
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 11.33ms DONE
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 10.38ms DONE
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 10.50ms DONE
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob RUNNING
docker_lamp_1 | 2026-04-20 06:44:20 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob 821.63ms DONE
docker_lamp_1 | 2026-04-20 06:44:20 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob RUNNING
docker_lamp_1 | 2026-04-20 06:44:20 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob 8.20ms DONE
docker_lamp_1 | 2026-04-20 06:44:25 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob RUNNING
docker_lamp_1 | 2026-04-20 06:44:25 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob 26.88ms DONE
docker_lamp_1 | 2026-04-20 06:44:30 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob RUNNING
docker_lamp_1 | 2026-04-20 06:44:30 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob 10.59ms DONE
docker_lamp_1 | 2026-04-20 06:44:35 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob RUNNING
docker_lamp_1 | 2026-04-20 06:44:35 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob 52.49ms DONE
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:45:06 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:45:13 Running ['artisan' dialers:monitor-activities] . 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:45:17 Running ['artisan' jiminny:monitor-social-accounts] 13s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:45:30 Running ['artisan' mailbox:skip-lists:refresh] 14s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:45:45 Running ['artisan' mailbox:batch:process --max-batches=15] 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:45:54 Running ['artisan' activity:purge-stale] ...... 10s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:purge-stale > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:05 Running ['artisan' mailbox:text-relay:sync] {
docker_lamp_1 | "error": "invalid_request",
docker_lamp_1 | "error_description": "Invalid impersonation \u0026quot;sub\u0026quot; field: @"
docker_lamp_1 | }
docker_lamp_1 | .... 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:text-relay:sync > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:14 Running ['artisan' conference:pre-meeting-notification] 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:pre-meeting-notification > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:18 Running ['artisan' conference:monitor:start] ... 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:start > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:27 Running ['artisan' conference:monitor:end] ..... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:end > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:29 Running ['artisan' jiminny:fix-hubspot-tokens] . 3s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:fix-hubspot-tokens > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' conference:pre-meeting-reminder] in background 1.51ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' conference:pre-meeting-reminder > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-805efb160ee8d9da02e60364ace7970eb2b35f31" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' hubspot:journal-poll --start] in background 3.11ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' hubspot:journal-poll --start > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-e26d77f915d2c55fe91ca4148a230e32eaa1865e" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' datadog:report:processing-sla-activities] 9s DONE
docker_lamp_1 | 🚀 Starting HubSpot journal polling service...
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' datadog:report:processing-sla-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:42 Running ['artisan' activity:sync --from='2026-04-20 06:29:00' --to='2026-04-20 06:45:00' --skipProviders='ringcentral' --skipProviders='avaya' --skipProviders='telus' --skipProviders='talkdesk'] 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:sync --from='2026-04-20 06:29:00' --to='2026-04-20 06:45:00' --skipProviders='ringcentral' --skipProviders='avaya' --skipProviders='telus' --skipProviders='talkdesk' > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:51 Running ['artisan' crm:bullhorn:ping --heartbeat] 2026-04-20 06:46:53 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:46:54 Jiminny\Jobs\Activity\SyncActivity ....... 959.91ms DONE
docker_lamp_1 | 2026-04-20 06:46:55 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:46:58 Jiminny\Notifications\Crm\UserAccountDisconnected RUNNING
docker_lamp_1 | 2026-04-20 06:46:59 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] 4s DONE
docker_lamp_1 | 2026-04-20 06:46:59 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:47:01 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] 2s DONE
docker_lamp_1 | 2026-04-20 06:47:01 Jiminny\Notifications\Crm\UserAccountDisconnected 2026-04-20 06:47:01 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 3s DONE
docker_lamp_1 | 2026-04-20 06:47:03 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-20 06:47:03 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:47:04 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-20 06:47:04 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:47:08 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] 4s DONE
docker_lamp_1 | 0 social account(s) to be processed ...
docker_lamp_1 |
docker_lamp_1 | Done!
docker_lamp_1 | 20s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:bullhorn:ping --heartbeat > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:47:11 Running ['artisan' nudges:send --silent] ....... 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' nudges:send --silent > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:47:17 Running ['artisan' jiminny:playlists:normalize-sort] 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:playlists:normalize-sort > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:48:06 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:48:11 Running ['artisan' dialers:monitor-activities] . 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:48:18 Running ['artisan' jiminny:monitor-social-accounts] 11s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:48:30 Running ['artisan' mailbox:skip-lists:refresh] 14s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:48:47,842Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][559] overhead, spent [265ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | 2026-04-20 06:48:45 Running ['artisan' mailbox:batch:process --max-batches=15] 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:48:50 Running ['artisan' conference:monitor:count] ... 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:48:58 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 14.85ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:48:58 Running ['artisan' crm:autolog-delayed] Dispatched autolog delayed jobs for all applicable teams:
docker_lamp_1 | ....... 10s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:autolog-delayed > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:50:07 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:50:13 Running ['artisan' dialers:monitor-activities] . 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:50:18 Running ['artisan' jiminny:monitor-social-accounts] 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:50:25 Running ['artisan' mailbox:skip-lists:refresh] 17s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:50:43 Running ['artisan' mailbox:batch:process --max-batches=15] 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:50:51 Running ['artisan' conference:monitor:count] .. 11s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:03 Running ['artisan' activity:purge-stale] ...... 12s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:purge-stale > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:15 Running ['artisan' mailbox:text-relay:sync] {
docker_lamp_1 | "error": "invalid_request",
docker_lamp_1 | "error_description": "Invalid impersonation \u0026quot;sub\u0026quot; field: @"
docker_lamp_1 | }
docker_lamp_1 | .... 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:text-relay:sync > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:23 Running ['artisan' conference:pre-meeting-notification] 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:pre-meeting-notification > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:30 Running ['artisan' conference:monitor:start] ... 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:start > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:39 Running ['artisan' conference:monitor:end] ..... 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:end > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:47 Running ['artisan' jiminny:fix-hubspot-tokens] . 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:fix-hubspot-tokens > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' conference:pre-meeting-reminder] in background 3.72ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' conference:pre-meeting-reminder > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-805efb160ee8d9da02e60364ace7970eb2b35f31" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' hubspot:journal-poll --start] in background 2.73ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' hubspot:journal-poll --start > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-e26d77f915d2c55fe91ca4148a230e32eaa1865e" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' jiminny:transcription:retry-failed] No failed transcriptions found.
docker_lamp_1 | 🚀 Starting HubSpot journal polling service...
docker_lamp_1 | 12s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:transcription:retry-failed > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:52:08 Running ['artisan' crm:reset-governor] [PASSWORD_DOTS] 17s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:reset-governor > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:52:26 Running ['artisan' crm:bullhorn:ping --heartbeat] 0 social account(s) to be processed ...
docker_lamp_1 |
docker_lamp_1 | Done!
docker_lamp_1 | 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:bullhorn:ping --heartbeat > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:53:07 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:53:15 Running ['artisan' dialers:monitor-activities] 11s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:53:26 Running ['artisan' jiminny:monitor-social-accounts] 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:53:35 Running ['artisan' mailbox:skip-lists:refresh] . 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:53:40 Running ['artisan' mailbox:batch:process --max-batches=15] 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:53:50 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 16.19ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:54:09 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:15 Running ['artisan' dialers:monitor-activities] . 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:19 Running ['artisan' jiminny:monitor-social-accounts] 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:24 Running ['artisan' mailbox:skip-lists:refresh] . 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:28 Running ['artisan' mailbox:batch:process --max-batches=15] 12s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:40 Running ['artisan' conference:monitor:count] ... 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:46 Running ['artisan' activity:aircall:check-and-renew] 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:aircall:check-and-renew > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:53 Running ['artisan' track:retry-failed-downloads] 11s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' track:retry-failed-downloads > '/proc/1/fd/...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":",apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:42:04 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:06 Running ['artisan' dialers:monitor-activities] . 2s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:08 Running ['artisan' jiminny:monitor-social-accounts] 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:12 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:14 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:15 Running ['artisan' conference:monitor:count] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:16 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:42:20 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:42:20 Jiminny\\Jobs\\Mailbox\\CreateBatches ........ 44.79ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:43:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:06 Running ['artisan' jiminny:monitor-social-accounts] 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:11 Running ['artisan' mailbox:skip-lists:refresh] . 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:14 Running ['artisan' mailbox:batch:process --max-batches=15] 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 12.38ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' calendar:sync --dateMode=daily] 2026-04-20 06:43:31 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:32 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... 1s DONE\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\ndocker_lamp_1 | 14s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' calendar:sync --dateMode=daily > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents . 324.40ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:43 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... 8s DONE\ndocker_lamp_1 | 2026-04-20 06:43:43 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:44 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 887.95ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:44 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:45 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 902.35ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:45 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:46 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 959.92ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:46 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:47 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 855.24ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:47 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:48 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 933.68ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:48 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:48 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 952.67ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:49 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:49 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 933.63ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:49 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:51 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 1s DONE\ndocker_lamp_1 | 2026-04-20 06:43:51 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:52 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 1s DONE\ndocker_lamp_1 | 2026-04-20 06:43:52 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:53 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 1s DONE\ndocker_lamp_1 | 2026-04-20 06:43:53 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:54 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 986.58ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:54 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:56 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 2s DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:44:03 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:06 Running ['artisan' jiminny:monitor-social-accounts] 2s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:09 Running ['artisan' mailbox:skip-lists:refresh] . 2s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:11 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:13 Running ['artisan' conference:monitor:count] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:14 Running ['artisan' crm:sync-objects] Syncing objects for Salesforce Team since 2026-03-19 20:14:10\ndocker_lamp_1 | Syncing objects for Hubspot since 2026-04-18 13:14:12\ndocker_lamp_1 | Syncing objects for Edge Communications since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Burlington Textiles Corp of America since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Uber since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Vonage since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Nexmo since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for SF since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for NewAccount since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for TestV since 2026-02-17 15:05:47\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for KioskAccount since 2026-02-17 15:06:27\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Pipedrive, Inc. since 2026-04-08 18:14:13\ndocker_lamp_1 | Syncing objects for Copper since 2026-04-18 13:14:12\ndocker_lamp_1 | Syncing objects for Pipedrive External Test since 2026-04-08 18:14:16\ndocker_lamp_1 | Syncing objects for InsightSquared since 2026-02-17 15:09:59\ndocker_lamp_1 | Syncing objects for Close since 2026-04-18 13:14:13\ndocker_lamp_1 | Syncing objects for DeewanyAccount since 2026-02-19 13:27:51\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Aircall Demo since 2026-02-17 15:06:51\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05\ndocker_lamp_1 | Syncing objects for GoStudent UAT since 2026-02-17 15:09:59\ndocker_lamp_1 | Syncing objects for Horen test since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Ahmed Hamadi since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Ahmed Testing since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Ahmed jiminny since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for JustCall since 2026-02-17 15:07:41\ndocker_lamp_1 | Syncing objects for Testers Inc since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Copper Team since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for TheBestPlace Ever since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Horen's apartments since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for New Org Test since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Pipedrive test org since 2026-02-19 13:27:42\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for BigChairs Inc. since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for GL 500 since 2026-02-19 13:26:04\ndocker_lamp_1 | Syncing objects for Loren X since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Laravel Company 2024 since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Barbara Fuchs since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Twilio Video since 2026-02-17 15:09:59\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for My Test Account 3000 since 2026-02-17 15:09:59\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for test since 2026-02-17 15:09:59\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Test since 2026-02-17 15:09:59\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Test salesforce auto sync since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for test ogg auto sync since 2026-02-17 15:07:16\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Globo Tech since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Laravel 11 Company 2024 since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | \ndocker_lamp_1 | Error \ndocker_lamp_1 | \ndocker_lamp_1 | Call to a member function getAttribute() on null\ndocker_lamp_1 | \ndocker_lamp_1 | at app/Console/Commands/Crm/SyncObjects.php:42\ndocker_lamp_1 | 38▕ \ndocker_lamp_1 | 39▕ foreach ($teams as $team) {\ndocker_lamp_1 | 40▕ $config = $team->getCrmConfiguration();\ndocker_lamp_1 | 41▕ \ndocker_lamp_1 | ➜ 42▕ if ($config->getAttribute('sync_objects') === false) {\ndocker_lamp_1 | 43▕ continue;\ndocker_lamp_1 | 44▕ }\ndocker_lamp_1 | 45▕ \ndocker_lamp_1 | 46▕ $this->info(\ndocker_lamp_1 | \ndocker_lamp_1 | 1 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:36\ndocker_lamp_1 | Jiminny\\Console\\Commands\\Crm\\SyncObjects::handle()\ndocker_lamp_1 | \ndocker_lamp_1 | 2 vendor/laravel/framework/src/Illuminate/Container/Util.php:43\ndocker_lamp_1 | Illuminate\\Container\\BoundMethod::Illuminate\\Container\\{closure}()\ndocker_lamp_1 | \ndocker_lamp_1 | 3 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:96\ndocker_lamp_1 | Illuminate\\Container\\Util::unwrapIfClosure(Object(Closure))\ndocker_lamp_1 | \ndocker_lamp_1 | 4 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:35\ndocker_lamp_1 | Illuminate\\Container\\BoundMethod::callBoundMethod(Object(Illuminate\\Foundation\\Application), Object(Closure))\ndocker_lamp_1 | \ndocker_lamp_1 | 5 vendor/laravel/framework/src/Illuminate/Container/Container.php:799\ndocker_lamp_1 | Illuminate\\Container\\BoundMethod::call(Object(Illuminate\\Foundation\\Application), [])\ndocker_lamp_1 | \ndocker_lamp_1 | 6 vendor/laravel/framework/src/Illuminate/Console/Command.php:211\ndocker_lamp_1 | Illuminate\\Container\\Container::call()\ndocker_lamp_1 | \ndocker_lamp_1 | 7 vendor/symfony/console/Command/Command.php:341\ndocker_lamp_1 | Illuminate\\Console\\Command::execute(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Illuminate\\Console\\OutputStyle))\ndocker_lamp_1 | \ndocker_lamp_1 | 8 vendor/laravel/framework/src/Illuminate/Console/Command.php:180\ndocker_lamp_1 | Symfony\\Component\\Console\\Command\\Command::run(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Illuminate\\Console\\OutputStyle))\ndocker_lamp_1 | \ndocker_lamp_1 | 9 app/Console/Commands/Command.php:41\ndocker_lamp_1 | Illuminate\\Console\\Command::run(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 10 vendor/symfony/console/Application.php:1117\ndocker_lamp_1 | Jiminny\\Console\\Commands\\Command::run(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 11 vendor/symfony/console/Application.php:356\ndocker_lamp_1 | Symfony\\Component\\Console\\Application::doRunCommand(Object(Jiminny\\Console\\Commands\\Crm\\SyncObjects), Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 12 vendor/symfony/console/Application.php:195\ndocker_lamp_1 | Symfony\\Component\\Console\\Application::doRun(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 13 vendor/laravel/framework/src/Illuminate/Foundation/Console/Kernel.php:198\ndocker_lamp_1 | Symfony\\Component\\Console\\Application::run(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 14 vendor/laravel/framework/src/Illuminate/Foundation/Application.php:1235\ndocker_lamp_1 | Illuminate\\Foundation\\Console\\Kernel::handle(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 15 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \ndocker_lamp_1 | ........... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:sync-objects > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:44:15 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:15 Jiminny\\Jobs\\Crm\\SyncObjects .............. 98.77ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:15 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:17 Jiminny\\Jobs\\Crm\\SyncObjects ................... 1s DONE\ndocker_lamp_1 | 2026-04-20 06:44:17 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:17 Jiminny\\Jobs\\Crm\\SyncObjects .............. 14.35ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:17 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects ................... 1s DONE\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects .............. 14.13ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects .............. 10.93ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects ................... 1s DONE\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects .............. 13.44ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects .............. 11.33ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects .............. 10.38ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects .............. 10.50ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:20 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob 821.63ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:20 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:20 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob 8.20ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:25 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:25 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob 26.88ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:30 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:30 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob 10.59ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:35 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:35 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob 52.49ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:45:06 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:45:13 Running ['artisan' dialers:monitor-activities] . 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:45:17 Running ['artisan' jiminny:monitor-social-accounts] 13s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:45:30 Running ['artisan' mailbox:skip-lists:refresh] 14s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:45:45 Running ['artisan' mailbox:batch:process --max-batches=15] 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:45:54 Running ['artisan' activity:purge-stale] ...... 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:purge-stale > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:05 Running ['artisan' mailbox:text-relay:sync] {\ndocker_lamp_1 | \"error\": \"invalid_request\",\ndocker_lamp_1 | \"error_description\": \"Invalid impersonation \\u0026quot;sub\\u0026quot; field: @\"\ndocker_lamp_1 | }\ndocker_lamp_1 | .... 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:text-relay:sync > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:14 Running ['artisan' conference:pre-meeting-notification] 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:pre-meeting-notification > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:18 Running ['artisan' conference:monitor:start] ... 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:start > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:27 Running ['artisan' conference:monitor:end] ..... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:end > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:29 Running ['artisan' jiminny:fix-hubspot-tokens] . 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:fix-hubspot-tokens > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' conference:pre-meeting-reminder] in background 1.51ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' conference:pre-meeting-reminder > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-805efb160ee8d9da02e60364ace7970eb2b35f31\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' hubspot:journal-poll --start] in background 3.11ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' hubspot:journal-poll --start > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-e26d77f915d2c55fe91ca4148a230e32eaa1865e\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' datadog:report:processing-sla-activities] 9s DONE\ndocker_lamp_1 | 🚀\u0000 Starting HubSpot journal polling service...\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' datadog:report:processing-sla-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:42 Running ['artisan' activity:sync --from='2026-04-20 06:29:00' --to='2026-04-20 06:45:00' --skipProviders='ringcentral' --skipProviders='avaya' --skipProviders='telus' --skipProviders='talkdesk'] 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:sync --from='2026-04-20 06:29:00' --to='2026-04-20 06:45:00' --skipProviders='ringcentral' --skipProviders='avaya' --skipProviders='telus' --skipProviders='talkdesk' > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:51 Running ['artisan' crm:bullhorn:ping --heartbeat] 2026-04-20 06:46:53 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:46:54 Jiminny\\Jobs\\Activity\\SyncActivity ....... 959.91ms DONE\ndocker_lamp_1 | 2026-04-20 06:46:55 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:46:58 Jiminny\\Notifications\\Crm\\UserAccountDisconnected RUNNING\ndocker_lamp_1 | 2026-04-20 06:46:59 Jiminny\\Jobs\\Activity\\SyncActivity ............. 4s DONE\ndocker_lamp_1 | 2026-04-20 06:46:59 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:47:01 Jiminny\\Jobs\\Activity\\SyncActivity ............. 2s DONE\ndocker_lamp_1 | 2026-04-20 06:47:01 Jiminny\\Notifications\\Crm\\UserAccountDisconnected 2026-04-20 06:47:01 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 3s DONE\ndocker_lamp_1 | 2026-04-20 06:47:03 Jiminny\\Jobs\\Activity\\SyncActivity ............. 1s DONE\ndocker_lamp_1 | 2026-04-20 06:47:03 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:47:04 Jiminny\\Jobs\\Activity\\SyncActivity ............. 1s DONE\ndocker_lamp_1 | 2026-04-20 06:47:04 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:47:08 Jiminny\\Jobs\\Activity\\SyncActivity ............. 4s DONE\ndocker_lamp_1 | 0 social account(s) to be processed ...\ndocker_lamp_1 | \ndocker_lamp_1 | Done!\ndocker_lamp_1 | 20s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:bullhorn:ping --heartbeat > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:47:11 Running ['artisan' nudges:send --silent] ....... 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' nudges:send --silent > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:47:17 Running ['artisan' jiminny:playlists:normalize-sort] 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:playlists:normalize-sort > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:48:06 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:48:11 Running ['artisan' dialers:monitor-activities] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:48:18 Running ['artisan' jiminny:monitor-social-accounts] 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:48:30 Running ['artisan' mailbox:skip-lists:refresh] 14s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:48:47,842Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][559] overhead, spent [265ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | 2026-04-20 06:48:45 Running ['artisan' mailbox:batch:process --max-batches=15] 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:48:50 Running ['artisan' conference:monitor:count] ... 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:48:58 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 14.85ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:48:58 Running ['artisan' crm:autolog-delayed] Dispatched autolog delayed jobs for all applicable teams: \ndocker_lamp_1 | ....... 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:autolog-delayed > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:50:07 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:50:13 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:50:18 Running ['artisan' jiminny:monitor-social-accounts] 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:50:25 Running ['artisan' mailbox:skip-lists:refresh] 17s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:50:43 Running ['artisan' mailbox:batch:process --max-batches=15] 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:50:51 Running ['artisan' conference:monitor:count] .. 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:03 Running ['artisan' activity:purge-stale] ...... 12s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:purge-stale > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:15 Running ['artisan' mailbox:text-relay:sync] {\ndocker_lamp_1 | \"error\": \"invalid_request\",\ndocker_lamp_1 | \"error_description\": \"Invalid impersonation \\u0026quot;sub\\u0026quot; field: @\"\ndocker_lamp_1 | }\ndocker_lamp_1 | .... 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:text-relay:sync > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:23 Running ['artisan' conference:pre-meeting-notification] 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:pre-meeting-notification > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:30 Running ['artisan' conference:monitor:start] ... 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:start > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:39 Running ['artisan' conference:monitor:end] ..... 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:end > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:47 Running ['artisan' jiminny:fix-hubspot-tokens] . 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:fix-hubspot-tokens > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' conference:pre-meeting-reminder] in background 3.72ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' conference:pre-meeting-reminder > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-805efb160ee8d9da02e60364ace7970eb2b35f31\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' hubspot:journal-poll --start] in background 2.73ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' hubspot:journal-poll --start > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-e26d77f915d2c55fe91ca4148a230e32eaa1865e\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' jiminny:transcription:retry-failed] No failed transcriptions found.\ndocker_lamp_1 | 🚀\u0000 Starting HubSpot journal polling service...\ndocker_lamp_1 | 12s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:transcription:retry-failed > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:52:08 Running ['artisan' crm:reset-governor] ........ 17s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:reset-governor > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:52:26 Running ['artisan' crm:bullhorn:ping --heartbeat] 0 social account(s) to be processed ...\ndocker_lamp_1 | \ndocker_lamp_1 | Done!\ndocker_lamp_1 | 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:bullhorn:ping --heartbeat > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:53:07 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:53:15 Running ['artisan' dialers:monitor-activities] 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:53:26 Running ['artisan' jiminny:monitor-social-accounts] 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:53:35 Running ['artisan' mailbox:skip-lists:refresh] . 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:53:40 Running ['artisan' mailbox:batch:process --max-batches=15] 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:53:50 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 16.19ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:54:09 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:15 Running ['artisan' dialers:monitor-activities] . 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:19 Running ['artisan' jiminny:monitor-social-accounts] 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:24 Running ['artisan' mailbox:skip-lists:refresh] . 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:28 Running ['artisan' mailbox:batch:process --max-batches=15] 12s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:40 Running ['artisan' conference:monitor:count] ... 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:46 Running ['artisan' activity:aircall:check-and-renew] 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:aircall:check-and-renew > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:53 Running ['artisan' track:retry-failed-downloads] 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' track:retry-failed-downloads > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:56:08 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:56:13 Running ['artisan' dialers:monitor-activities] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:56:24 Running ['artisan' jiminny:monitor-social-accounts] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:56:35 Running ['artisan' mailbox:skip-lists:refresh] 24s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:57:11,425Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][1055] overhead, spent [616ms] collecting in the last [1.6s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | 2026-04-20 06:57:00 Running ['artisan' mailbox:batch:process --max-batches=15] 18s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:57:18 Running ['artisan' conference:monitor:count] .. 13s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:57:31 Running ['artisan' activity:notify-not-logged] 15s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:notify-not-logged > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:57:46 Running ['artisan' activity:status-count] []..... 17s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:status-count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:58:04 Running ['artisan' mailbox:sync] Queueing 2 inbox(es) for sync.\ndocker_lamp_1 | ............... 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:sync > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:58:13 Jiminny\\Jobs\\Mailbox\\SyncInbox ................. RUNNING\ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:58:14 Jiminny\\Jobs\\Mailbox\\SyncInbox ........... 829.46ms DONE\ndocker_lamp_1 | 2026-04-20 06:58:14 Jiminny\\Jobs\\Mailbox\\SyncInbox ................. RUNNING\ndocker_lamp_1 | 2026-04-20 06:58:15 Jiminny\\Jobs\\Mailbox\\SyncInbox ................. 1s DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:59:10 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:59:17 Running ['artisan' dialers:monitor-activities] 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:59:28 Running ['artisan' jiminny:monitor-social-accounts] 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:59:36 Running ['artisan' mailbox:skip-lists:refresh] 20s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:59:56 Running ['artisan' mailbox:batch:process --max-batches=15] 17s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 07:01:09 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:01:15 Running ['artisan' dialers:monitor-activities] . 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:01:23 Running ['artisan' jiminny:monitor-social-accounts] 13s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:01:36 Running ['artisan' mailbox:skip-lists:refresh] 13s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:01:49 Running ['artisan' mailbox:batch:process --max-batches=15] 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:01:59 Running ['artisan' activity:sync 'ringcentral' --from='2026-04-20 01:01:00' --to='2026-04-20 07:01:00'] 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:sync 'ringcentral' --from='2026-04-20 01:01:00' --to='2026-04-20 07:01:00' > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 07:03:08 Running ['artisan' meeting-bot:schedule-bot] ... 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:03:18 Running ['artisan' dialers:monitor-activities] 22s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:03:40 Running ['artisan' jiminny:monitor-social-accounts] 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \n\n\nv View in Docker Desktop o View Config w Enable Watch","depth":4,"bounds":{"left":0.27027926,"top":1.0,"width":0.23869681,"height":-0.08060658},"value":",apm,uptime]\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"taskManager\",\"taskManager\"],\"pid\":7,\"message\":\"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:06,605Z\", \"level\": \"INFO\", \"component\": \"o.e.c.m.MetadataIndexTemplateService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"adding template [.management-beats] for index patterns [.management-beats]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"crossClusterReplication\"],\"pid\":7,\"message\":\"Your basic license does not support crossClusterReplication. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"watcher\"],\"pid\":7,\"message\":\"Your basic license does not support watcher. Please upgrade your license.\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:06Z\",\"tags\":[\"info\",\"plugins\",\"monitoring\",\"monitoring\",\"kibana-monitoring\"],\"pid\":7,\"message\":\"Starting monitoring stats collection\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"error\",\"elasticsearch\",\"data\"],\"pid\":7,\"message\":\"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:07Z\",\"tags\":[\"listening\",\"info\"],\"pid\":7,\"message\":\"Server running at http://0:5601\"}\ndocker_lamp_1 | + mkdir -p /home/jiminny_storage\ndocker_lamp_1 | + cp -prf /home/jiminny/jiminny_storage/app /home/jiminny/jiminny_storage/debugbar /home/jiminny/jiminny_storage/framework /home/jiminny/jiminny_storage/logs /home/jiminny/jiminny_storage/oauth-private.key /home/jiminny/jiminny_storage/oauth-public.key /home/jiminny/jiminny_storage/storage /home/jiminny/jiminny_storage/supervisord.pid /home/jiminny/jiminny_storage/text-relay.json /home/jiminny_storage/\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:10Z\",\"tags\":[\"info\",\"http\",\"server\",\"Kibana\"],\"pid\":7,\"message\":\"http server running at http://0:5601\"}\nkibana | {\"type\":\"log\",\"@timestamp\":\"2026-04-20T06:40:11Z\",\"tags\":[\"warning\",\"plugins\",\"reporting\"],\"pid\":7,\"message\":\"Enabling the Chromium sandbox provides an additional layer of protection.\"}\nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:40:14,068Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][50] overhead, spent [362ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | + trap _term SIGTERM\ndocker_lamp_1 | + super_pid=13640\ndocker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf\ndocker_lamp_1 | + nginx\ndocker_lamp_1 | + php-fpm\ndocker_lamp_1 | + wait 13640\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642\ndocker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections\ndocker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/artisan-schedule.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-1.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-2.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-3.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-4.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-5.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-analytics.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-audio.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-calendar.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-conferences.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-sync.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-crm-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-download.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-emails.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-es-update.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker-nudges.conf\" during parsing\ndocker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file \"/etc/supervisor/conf.d/worker.conf\" during parsing\ndocker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized\ndocker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking\ndocker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640\ndocker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653\ndocker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654\ndocker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655\ndocker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656\ndocker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657\ndocker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658\ndocker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660\ndocker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664\ndocker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668\ndocker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669\ndocker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670\ndocker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672\ndocker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673\ndocker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674\ndocker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675\ndocker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676\ndocker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677\ndocker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679\ndocker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)\ndocker_lamp_1 | 2026-04-20 06:40:39 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:40:40 Jiminny\\Jobs\\Mailbox\\CreateBatches ....... 505.47ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:42:04 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:06 Running ['artisan' dialers:monitor-activities] . 2s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:08 Running ['artisan' jiminny:monitor-social-accounts] 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:12 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:14 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:15 Running ['artisan' conference:monitor:count] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:42:16 Running ['artisan' mailbox:batch:create] ....... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:42:20 Jiminny\\Jobs\\Mailbox\\CreateBatches ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:42:20 Jiminny\\Jobs\\Mailbox\\CreateBatches ........ 44.79ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:43:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:06 Running ['artisan' jiminny:monitor-social-accounts] 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:11 Running ['artisan' mailbox:skip-lists:refresh] . 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:14 Running ['artisan' mailbox:batch:process --max-batches=15] 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 12.38ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' calendar:sync --dateMode=daily] 2026-04-20 06:43:31 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:32 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... 1s DONE\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\ndocker_lamp_1 | 14s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' calendar:sync --dateMode=daily > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents . 324.40ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:34 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:43 Jiminny\\Jobs\\Calendar\\SyncCalendarEvents ....... 8s DONE\ndocker_lamp_1 | 2026-04-20 06:43:43 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:44 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 887.95ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:44 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:45 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 902.35ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:45 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:46 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 959.92ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:46 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:47 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 855.24ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:47 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:48 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 933.68ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:48 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:48 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 952.67ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:49 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:49 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 933.63ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:49 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:51 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 1s DONE\ndocker_lamp_1 | 2026-04-20 06:43:51 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:52 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 1s DONE\ndocker_lamp_1 | 2026-04-20 06:43:52 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:53 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 1s DONE\ndocker_lamp_1 | 2026-04-20 06:43:53 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:54 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 986.58ms DONE\ndocker_lamp_1 | 2026-04-20 06:43:54 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents RUNNING\ndocker_lamp_1 | 2026-04-20 06:43:56 Jiminny\\Jobs\\Calendar\\SyncOfficeRecurringEvents 2s DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:44:03 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:04 Running ['artisan' dialers:monitor-activities] . 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:06 Running ['artisan' jiminny:monitor-social-accounts] 2s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:09 Running ['artisan' mailbox:skip-lists:refresh] . 2s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:11 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:13 Running ['artisan' conference:monitor:count] ... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:44:14 Running ['artisan' crm:sync-objects] Syncing objects for Salesforce Team since 2026-03-19 20:14:10\ndocker_lamp_1 | Syncing objects for Hubspot since 2026-04-18 13:14:12\ndocker_lamp_1 | Syncing objects for Edge Communications since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Burlington Textiles Corp of America since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Uber since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Vonage since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Nexmo since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for SF since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for NewAccount since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for TestV since 2026-02-17 15:05:47\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for KioskAccount since 2026-02-17 15:06:27\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Pipedrive, Inc. since 2026-04-08 18:14:13\ndocker_lamp_1 | Syncing objects for Copper since 2026-04-18 13:14:12\ndocker_lamp_1 | Syncing objects for Pipedrive External Test since 2026-04-08 18:14:16\ndocker_lamp_1 | Syncing objects for InsightSquared since 2026-02-17 15:09:59\ndocker_lamp_1 | Syncing objects for Close since 2026-04-18 13:14:13\ndocker_lamp_1 | Syncing objects for DeewanyAccount since 2026-02-19 13:27:51\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Aircall Demo since 2026-02-17 15:06:51\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05\ndocker_lamp_1 | Syncing objects for GoStudent UAT since 2026-02-17 15:09:59\ndocker_lamp_1 | Syncing objects for Horen test since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Ahmed Hamadi since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Ahmed Testing since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Ahmed jiminny since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for JustCall since 2026-02-17 15:07:41\ndocker_lamp_1 | Syncing objects for Testers Inc since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Copper Team since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for TheBestPlace Ever since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Horen's apartments since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for New Org Test since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Pipedrive test org since 2026-02-19 13:27:42\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for BigChairs Inc. since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for GL 500 since 2026-02-19 13:26:04\ndocker_lamp_1 | Syncing objects for Loren X since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Laravel Company 2024 since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Barbara Fuchs since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Twilio Video since 2026-02-17 15:09:59\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for My Test Account 3000 since 2026-02-17 15:09:59\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for test since 2026-02-17 15:09:59\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Test since 2026-02-17 15:09:59\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Test salesforce auto sync since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for test ogg auto sync since 2026-02-17 15:07:16\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Globo Tech since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | Syncing objects for Laravel 11 Company 2024 since 2026-02-19 13:26:04\ndocker_lamp_1 | Team is not yet assigned an owner. skipping...\ndocker_lamp_1 | \ndocker_lamp_1 | Error \ndocker_lamp_1 | \ndocker_lamp_1 | Call to a member function getAttribute() on null\ndocker_lamp_1 | \ndocker_lamp_1 | at app/Console/Commands/Crm/SyncObjects.php:42\ndocker_lamp_1 | 38▕ \ndocker_lamp_1 | 39▕ foreach ($teams as $team) {\ndocker_lamp_1 | 40▕ $config = $team->getCrmConfiguration();\ndocker_lamp_1 | 41▕ \ndocker_lamp_1 | ➜ 42▕ if ($config->getAttribute('sync_objects') === false) {\ndocker_lamp_1 | 43▕ continue;\ndocker_lamp_1 | 44▕ }\ndocker_lamp_1 | 45▕ \ndocker_lamp_1 | 46▕ $this->info(\ndocker_lamp_1 | \ndocker_lamp_1 | 1 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:36\ndocker_lamp_1 | Jiminny\\Console\\Commands\\Crm\\SyncObjects::handle()\ndocker_lamp_1 | \ndocker_lamp_1 | 2 vendor/laravel/framework/src/Illuminate/Container/Util.php:43\ndocker_lamp_1 | Illuminate\\Container\\BoundMethod::Illuminate\\Container\\{closure}()\ndocker_lamp_1 | \ndocker_lamp_1 | 3 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:96\ndocker_lamp_1 | Illuminate\\Container\\Util::unwrapIfClosure(Object(Closure))\ndocker_lamp_1 | \ndocker_lamp_1 | 4 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:35\ndocker_lamp_1 | Illuminate\\Container\\BoundMethod::callBoundMethod(Object(Illuminate\\Foundation\\Application), Object(Closure))\ndocker_lamp_1 | \ndocker_lamp_1 | 5 vendor/laravel/framework/src/Illuminate/Container/Container.php:799\ndocker_lamp_1 | Illuminate\\Container\\BoundMethod::call(Object(Illuminate\\Foundation\\Application), [])\ndocker_lamp_1 | \ndocker_lamp_1 | 6 vendor/laravel/framework/src/Illuminate/Console/Command.php:211\ndocker_lamp_1 | Illuminate\\Container\\Container::call()\ndocker_lamp_1 | \ndocker_lamp_1 | 7 vendor/symfony/console/Command/Command.php:341\ndocker_lamp_1 | Illuminate\\Console\\Command::execute(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Illuminate\\Console\\OutputStyle))\ndocker_lamp_1 | \ndocker_lamp_1 | 8 vendor/laravel/framework/src/Illuminate/Console/Command.php:180\ndocker_lamp_1 | Symfony\\Component\\Console\\Command\\Command::run(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Illuminate\\Console\\OutputStyle))\ndocker_lamp_1 | \ndocker_lamp_1 | 9 app/Console/Commands/Command.php:41\ndocker_lamp_1 | Illuminate\\Console\\Command::run(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 10 vendor/symfony/console/Application.php:1117\ndocker_lamp_1 | Jiminny\\Console\\Commands\\Command::run(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 11 vendor/symfony/console/Application.php:356\ndocker_lamp_1 | Symfony\\Component\\Console\\Application::doRunCommand(Object(Jiminny\\Console\\Commands\\Crm\\SyncObjects), Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 12 vendor/symfony/console/Application.php:195\ndocker_lamp_1 | Symfony\\Component\\Console\\Application::doRun(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 13 vendor/laravel/framework/src/Illuminate/Foundation/Console/Kernel.php:198\ndocker_lamp_1 | Symfony\\Component\\Console\\Application::run(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 14 vendor/laravel/framework/src/Illuminate/Foundation/Application.php:1235\ndocker_lamp_1 | Illuminate\\Foundation\\Console\\Kernel::handle(Object(Symfony\\Component\\Console\\Input\\ArgvInput), Object(Symfony\\Component\\Console\\Output\\ConsoleOutput))\ndocker_lamp_1 | \ndocker_lamp_1 | 15 artisan:13\ndocker_lamp_1 | Illuminate\\Foundation\\Application::handleCommand(Object(Symfony\\Component\\Console\\Input\\ArgvInput))\ndocker_lamp_1 | \ndocker_lamp_1 | ........... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:sync-objects > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:44:15 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:15 Jiminny\\Jobs\\Crm\\SyncObjects .............. 98.77ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:15 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:17 Jiminny\\Jobs\\Crm\\SyncObjects ................... 1s DONE\ndocker_lamp_1 | 2026-04-20 06:44:17 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:17 Jiminny\\Jobs\\Crm\\SyncObjects .............. 14.35ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:17 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects ................... 1s DONE\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects .............. 14.13ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects .............. 10.93ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:18 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects ................... 1s DONE\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects .............. 13.44ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects .............. 11.33ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects .............. 10.38ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects ................... RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\SyncObjects .............. 10.50ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:19 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:20 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob 821.63ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:20 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:20 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob 8.20ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:25 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:25 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob 26.88ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:30 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:30 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob 10.59ms DONE\ndocker_lamp_1 | 2026-04-20 06:44:35 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob RUNNING\ndocker_lamp_1 | 2026-04-20 06:44:35 Jiminny\\Jobs\\Crm\\Salesforce\\FetchSalesforceEntitiesJob 52.49ms DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:45:06 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:45:13 Running ['artisan' dialers:monitor-activities] . 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:45:17 Running ['artisan' jiminny:monitor-social-accounts] 13s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:45:30 Running ['artisan' mailbox:skip-lists:refresh] 14s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:45:45 Running ['artisan' mailbox:batch:process --max-batches=15] 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:45:54 Running ['artisan' activity:purge-stale] ...... 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:purge-stale > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:05 Running ['artisan' mailbox:text-relay:sync] {\ndocker_lamp_1 | \"error\": \"invalid_request\",\ndocker_lamp_1 | \"error_description\": \"Invalid impersonation \\u0026quot;sub\\u0026quot; field: @\"\ndocker_lamp_1 | }\ndocker_lamp_1 | .... 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:text-relay:sync > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:14 Running ['artisan' conference:pre-meeting-notification] 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:pre-meeting-notification > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:18 Running ['artisan' conference:monitor:start] ... 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:start > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:27 Running ['artisan' conference:monitor:end] ..... 1s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:end > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:29 Running ['artisan' jiminny:fix-hubspot-tokens] . 3s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:fix-hubspot-tokens > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' conference:pre-meeting-reminder] in background 1.51ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' conference:pre-meeting-reminder > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-805efb160ee8d9da02e60364ace7970eb2b35f31\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' hubspot:journal-poll --start] in background 3.11ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' hubspot:journal-poll --start > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-e26d77f915d2c55fe91ca4148a230e32eaa1865e\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' datadog:report:processing-sla-activities] 9s DONE\ndocker_lamp_1 | 🚀\u0000 Starting HubSpot journal polling service...\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' datadog:report:processing-sla-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:42 Running ['artisan' activity:sync --from='2026-04-20 06:29:00' --to='2026-04-20 06:45:00' --skipProviders='ringcentral' --skipProviders='avaya' --skipProviders='telus' --skipProviders='talkdesk'] 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:sync --from='2026-04-20 06:29:00' --to='2026-04-20 06:45:00' --skipProviders='ringcentral' --skipProviders='avaya' --skipProviders='telus' --skipProviders='talkdesk' > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:46:51 Running ['artisan' crm:bullhorn:ping --heartbeat] 2026-04-20 06:46:53 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:46:54 Jiminny\\Jobs\\Activity\\SyncActivity ....... 959.91ms DONE\ndocker_lamp_1 | 2026-04-20 06:46:55 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:46:58 Jiminny\\Notifications\\Crm\\UserAccountDisconnected RUNNING\ndocker_lamp_1 | 2026-04-20 06:46:59 Jiminny\\Jobs\\Activity\\SyncActivity ............. 4s DONE\ndocker_lamp_1 | 2026-04-20 06:46:59 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:47:01 Jiminny\\Jobs\\Activity\\SyncActivity ............. 2s DONE\ndocker_lamp_1 | 2026-04-20 06:47:01 Jiminny\\Notifications\\Crm\\UserAccountDisconnected 2026-04-20 06:47:01 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 3s DONE\ndocker_lamp_1 | 2026-04-20 06:47:03 Jiminny\\Jobs\\Activity\\SyncActivity ............. 1s DONE\ndocker_lamp_1 | 2026-04-20 06:47:03 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:47:04 Jiminny\\Jobs\\Activity\\SyncActivity ............. 1s DONE\ndocker_lamp_1 | 2026-04-20 06:47:04 Jiminny\\Jobs\\Activity\\SyncActivity ............. RUNNING\ndocker_lamp_1 | 2026-04-20 06:47:08 Jiminny\\Jobs\\Activity\\SyncActivity ............. 4s DONE\ndocker_lamp_1 | 0 social account(s) to be processed ...\ndocker_lamp_1 | \ndocker_lamp_1 | Done!\ndocker_lamp_1 | 20s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:bullhorn:ping --heartbeat > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:47:11 Running ['artisan' nudges:send --silent] ....... 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' nudges:send --silent > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:47:17 Running ['artisan' jiminny:playlists:normalize-sort] 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:playlists:normalize-sort > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:48:06 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:48:11 Running ['artisan' dialers:monitor-activities] . 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:48:18 Running ['artisan' jiminny:monitor-social-accounts] 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:48:30 Running ['artisan' mailbox:skip-lists:refresh] 14s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:48:47,842Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][559] overhead, spent [265ms] collecting in the last [1s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | 2026-04-20 06:48:45 Running ['artisan' mailbox:batch:process --max-batches=15] 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:48:50 Running ['artisan' conference:monitor:count] ... 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:48:58 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 14.85ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:48:58 Running ['artisan' crm:autolog-delayed] Dispatched autolog delayed jobs for all applicable teams: \ndocker_lamp_1 | ....... 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:autolog-delayed > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:50:07 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:50:13 Running ['artisan' dialers:monitor-activities] . 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:50:18 Running ['artisan' jiminny:monitor-social-accounts] 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:50:25 Running ['artisan' mailbox:skip-lists:refresh] 17s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:50:43 Running ['artisan' mailbox:batch:process --max-batches=15] 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:50:51 Running ['artisan' conference:monitor:count] .. 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:03 Running ['artisan' activity:purge-stale] ...... 12s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:purge-stale > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:15 Running ['artisan' mailbox:text-relay:sync] {\ndocker_lamp_1 | \"error\": \"invalid_request\",\ndocker_lamp_1 | \"error_description\": \"Invalid impersonation \\u0026quot;sub\\u0026quot; field: @\"\ndocker_lamp_1 | }\ndocker_lamp_1 | .... 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:text-relay:sync > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:23 Running ['artisan' conference:pre-meeting-notification] 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:pre-meeting-notification > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:30 Running ['artisan' conference:monitor:start] ... 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:start > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:39 Running ['artisan' conference:monitor:end] ..... 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:end > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:47 Running ['artisan' jiminny:fix-hubspot-tokens] . 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:fix-hubspot-tokens > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' conference:pre-meeting-reminder] in background 3.72ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' conference:pre-meeting-reminder > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-805efb160ee8d9da02e60364ace7970eb2b35f31\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' hubspot:journal-poll --start] in background 2.73ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' hubspot:journal-poll --start > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-e26d77f915d2c55fe91ca4148a230e32eaa1865e\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' jiminny:transcription:retry-failed] No failed transcriptions found.\ndocker_lamp_1 | 🚀\u0000 Starting HubSpot journal polling service...\ndocker_lamp_1 | 12s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:transcription:retry-failed > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:52:08 Running ['artisan' crm:reset-governor] ........ 17s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:reset-governor > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:52:26 Running ['artisan' crm:bullhorn:ping --heartbeat] 0 social account(s) to be processed ...\ndocker_lamp_1 | \ndocker_lamp_1 | Done!\ndocker_lamp_1 | 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:bullhorn:ping --heartbeat > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:53:07 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:53:15 Running ['artisan' dialers:monitor-activities] 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:53:26 Running ['artisan' jiminny:monitor-social-accounts] 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:53:35 Running ['artisan' mailbox:skip-lists:refresh] . 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:53:40 Running ['artisan' mailbox:batch:process --max-batches=15] 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:53:50 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 16.19ms DONE\ndocker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish \"framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11\" \"$?\") > '/dev/null' 2>&1 & \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:54:09 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:15 Running ['artisan' dialers:monitor-activities] . 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:19 Running ['artisan' jiminny:monitor-social-accounts] 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:24 Running ['artisan' mailbox:skip-lists:refresh] . 4s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:28 Running ['artisan' mailbox:batch:process --max-batches=15] 12s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:40 Running ['artisan' conference:monitor:count] ... 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:46 Running ['artisan' activity:aircall:check-and-renew] 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:aircall:check-and-renew > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:54:53 Running ['artisan' track:retry-failed-downloads] 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' track:retry-failed-downloads > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:56:08 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:56:13 Running ['artisan' dialers:monitor-activities] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:56:24 Running ['artisan' jiminny:monitor-social-accounts] 10s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:56:35 Running ['artisan' mailbox:skip-lists:refresh] 24s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \nelasticsearch | {\"type\": \"server\", \"timestamp\": \"2026-04-20T06:57:11,425Z\", \"level\": \"INFO\", \"component\": \"o.e.m.j.JvmGcMonitorService\", \"cluster.name\": \"docker-cluster\", \"node.name\": \"e802ad473a4f\", \"message\": \"[gc][1055] overhead, spent [616ms] collecting in the last [1.6s]\", \"cluster.uuid\": \"8uh2w1CUSGyWYR_OvaKx6g\", \"node.id\": \"e2ZKzgw4Q4aCf2w5ljWr1A\" }\ndocker_lamp_1 | 2026-04-20 06:57:00 Running ['artisan' mailbox:batch:process --max-batches=15] 18s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:57:18 Running ['artisan' conference:monitor:count] .. 13s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:57:31 Running ['artisan' activity:notify-not-logged] 15s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:notify-not-logged > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:57:46 Running ['artisan' activity:status-count] []..... 17s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:status-count > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:58:04 Running ['artisan' mailbox:sync] Queueing 2 inbox(es) for sync.\ndocker_lamp_1 | ............... 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:sync > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:58:13 Jiminny\\Jobs\\Mailbox\\SyncInbox ................. RUNNING\ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | 2026-04-20 06:58:14 Jiminny\\Jobs\\Mailbox\\SyncInbox ........... 829.46ms DONE\ndocker_lamp_1 | 2026-04-20 06:58:14 Jiminny\\Jobs\\Mailbox\\SyncInbox ................. RUNNING\ndocker_lamp_1 | 2026-04-20 06:58:15 Jiminny\\Jobs\\Mailbox\\SyncInbox ................. 1s DONE\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 06:59:10 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:59:17 Running ['artisan' dialers:monitor-activities] 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:59:28 Running ['artisan' jiminny:monitor-social-accounts] 7s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:59:36 Running ['artisan' mailbox:skip-lists:refresh] 20s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 06:59:56 Running ['artisan' mailbox:batch:process --max-batches=15] 17s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 07:01:09 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:01:15 Running ['artisan' dialers:monitor-activities] . 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:01:23 Running ['artisan' jiminny:monitor-social-accounts] 13s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:01:36 Running ['artisan' mailbox:skip-lists:refresh] 13s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:01:49 Running ['artisan' mailbox:batch:process --max-batches=15] 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:01:59 Running ['artisan' activity:sync 'ringcentral' --from='2026-04-20 01:01:00' --to='2026-04-20 07:01:00'] 8s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:sync 'ringcentral' --from='2026-04-20 01:01:00' --to='2026-04-20 07:01:00' > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | \ndocker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run\ndocker_lamp_1 | \ndocker_lamp_1 | 2026-04-20 07:03:08 Running ['artisan' meeting-bot:schedule-bot] ... 9s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:03:18 Running ['artisan' dialers:monitor-activities] 22s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1 \ndocker_lamp_1 | 2026-04-20 07:03:40 Running ['artisan' jiminny:monitor-social-accounts] 11s DONE\ndocker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1 \n\n\nv View in Docker Desktop o View Config w Enable Watch","is_focused":true},{"role":"AXButton","text":"Menu","depth":3,"bounds":{"left":0.50199467,"top":1.0,"width":0.004986702,"height":-0.06424582},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥1 DOCKER (docker-compose)","depth":3,"bounds":{"left":0.27925533,"top":1.0,"width":0.22107713,"height":-0.06464481},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","depth":5,"bounds":{"left":0.5093085,"top":0.11811652,"width":0.23969415,"height":0.8818835},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:21 UTC 2026\n\n System load: 0.0 Processes: 123\n Usage of /: 59.3% of 7.57GB Users logged in: 1\n Memory usage: 34% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n42 updates can be applied immediately.\n10 of these updates are standard security updates.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:09 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ prod\n(lukas@jiminny-prod-bastion) Verification code: \nWelcome to Ubuntu 22.04.5 LTS (GNU/Linux 6.8.0-1041-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:39 UTC 2026\n\n System load: 0.0 Processes: 139\n Usage of /: 58.6% of 7.57GB Users logged in: 3\n Memory usage: 37% IPv4 address for eth0: 10.30.45.167\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n37 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:22 2026 from 212.39.71.189\nlukas@jiminny-prod-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.74202126,"top":1.0,"width":0.004986702,"height":-0.06424582},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥2 PROD (ssh)","depth":4,"bounds":{"left":0.51828456,"top":1.0,"width":0.22207446,"height":-0.06464481},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","depth":5,"bounds":{"left":0.5093085,"top":0.21867518,"width":0.23969415,"height":0.7813248},"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Fri Apr 17 05:56:31 UTC 2026\n\n System load: 0.0 Processes: 118\n Usage of /: 58.0% of 7.57GB Users logged in: 1\n Memory usage: 18% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n90 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Thu Apr 16 06:55:03 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$ client_loop: send disconnect: Broken pipe\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ eu\n(lukas@jiminny-eu-bastion) Verification code: \nWelcome to Ubuntu 22.04.4 LTS (GNU/Linux 6.8.0-1047-aws x86_64)\n\n * Documentation: https://help.ubuntu.com\n * Management: https://landscape.canonical.com\n * Support: https://ubuntu.com/pro\n\n System information as of Mon Apr 20 06:40:48 UTC 2026\n\n System load: 0.0 Processes: 134\n Usage of /: 58.0% of 7.57GB Users logged in: 2\n Memory usage: 22% IPv4 address for eth0: 10.20.163.228\n Swap usage: 0%\n\n * Ubuntu Pro delivers the most comprehensive open source security and\n compliance features.\n\n https://ubuntu.com/aws/pro\n\nExpanded Security Maintenance for Applications is not enabled.\n\n85 updates can be applied immediately.\nTo see these additional updates run: apt list --upgradable\n\nEnable ESM Apps to receive additional future security updates.\nSee https://ubuntu.com/esm or run: sudo pro status\n\nNew release '24.04.4 LTS' available.\nRun 'do-release-upgrade' to upgrade to it.\n\n\n*** System restart required ***\nLast login: Fri Apr 17 05:56:31 2026 from 212.39.71.189\nlukas@jiminny-eu-bastion:~$","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥3 EU (ssh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Thu Apr 16 15:43:43 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true}]...
|
-7668142984030789381
|
-4834148495309075580
|
app_switch
|
accessibility
|
NULL
|
,apm,uptime]"}
kibana | {"type" ,apm,uptime]"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins","taskManager","taskManager"],"pid":7,"message":"TaskManager is identified by the Kibana UUID: bf01f365-e094-4cde-940d-3e0db65fa22a"}
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:40:06,605Z", "level": "INFO", "component": "o.e.c.m.MetadataIndexTemplateService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "adding template [.management-beats] for index patterns [.management-beats]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins","crossClusterReplication"],"pid":7,"message":"Your basic license does not support crossClusterReplication. Please upgrade your license."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins","watcher"],"pid":7,"message":"Your basic license does not support watcher. Please upgrade your license."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:06Z","tags":["info","plugins","monitoring","monitoring","kibana-monitoring"],"pid":7,"message":"Starting monitoring stats collection"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[version_conflict_engine_exception]: [task:Lens-lens_telemetry]: version conflict, document already exists (current version [736])"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[version_conflict_engine_exception]: [task:endpoint:user-artifact-packager:1.0.0]: version conflict, document already exists (current version [308723])"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[version_conflict_engine_exception]: [task:apm-telemetry-task]: version conflict, document already exists (current version [1213])"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[version_conflict_engine_exception]: [task:Alerting-alerting_telemetry]: version conflict, document already exists (current version [736])"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[version_conflict_engine_exception]: [task:Actions-actions_telemetry]: version conflict, document already exists (current version [736])"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:07Z","tags":["listening","info"],"pid":7,"message":"Server running at [URL_WITH_CREDENTIALS] server running at [URL_WITH_CREDENTIALS] the Chromium sandbox provides an additional layer of protection."}
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:40:14,068Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][50] overhead, spent [362ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + trap _term SIGTERM
docker_lamp_1 | + super_pid=13640
docker_lamp_1 | + supervisord --configuration /etc/supervisor/supervisord.conf
docker_lamp_1 | + nginx
docker_lamp_1 | + php-fpm
docker_lamp_1 | + wait 13640
docker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: fpm is running, pid 13642
docker_lamp_1 | [20-Apr-2026 06:40:30] NOTICE: ready to handle connections
docker_lamp_1 | 2026-04-20 06:40:30,437 CRIT Supervisor is running as root. Privileges were not dropped because no user is specified in the config file. If you intend to run as root, you can set user=root in the config file to avoid this message.
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/artisan-schedule.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-1.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-2.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-3.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-4.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-5.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/jiminny-worker-processing-delayed.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-analytics.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-audio.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-calendar.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-conferences.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-crm-sync.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-crm-update.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-download.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-emails.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-es-update.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker-nudges.conf" during parsing
docker_lamp_1 | 2026-04-20 06:40:30,437 INFO Included extra file "/etc/supervisor/conf.d/worker.conf" during parsing
docker_lamp_1 | Unlinking stale socket /var/run/supervisor.sock
docker_lamp_1 | 2026-04-20 06:40:30,751 INFO RPC interface 'supervisor' initialized
docker_lamp_1 | 2026-04-20 06:40:30,751 CRIT Server 'unix_http_server' running without any HTTP authentication checking
docker_lamp_1 | 2026-04-20 06:40:30,751 INFO supervisord started with pid 13640
docker_lamp_1 | 2026-04-20 06:40:31,759 INFO spawned: 'artisan-schedule_00' with pid 13653
docker_lamp_1 | 2026-04-20 06:40:31,763 INFO spawned: 'jiminny-worker-processing-1_00' with pid 13654
docker_lamp_1 | 2026-04-20 06:40:31,764 INFO spawned: 'jiminny-worker-processing-2_00' with pid 13655
docker_lamp_1 | 2026-04-20 06:40:31,766 INFO spawned: 'jiminny-worker-processing-3_00' with pid 13656
docker_lamp_1 | 2026-04-20 06:40:31,769 INFO spawned: 'jiminny-worker-processing-4_00' with pid 13657
docker_lamp_1 | 2026-04-20 06:40:31,772 INFO spawned: 'jiminny-worker-processing-5_00' with pid 13658
docker_lamp_1 | 2026-04-20 06:40:31,773 INFO spawned: 'jiminny-worker-processing-delayed_00' with pid 13660
docker_lamp_1 | 2026-04-20 06:40:31,776 INFO spawned: 'worker_00' with pid 13664
docker_lamp_1 | 2026-04-20 06:40:31,782 INFO spawned: 'worker-analytics_00' with pid 13668
docker_lamp_1 | 2026-04-20 06:40:31,790 INFO spawned: 'worker-audio_00' with pid 13669
docker_lamp_1 | 2026-04-20 06:40:31,793 INFO spawned: 'worker-calendar_00' with pid 13670
docker_lamp_1 | 2026-04-20 06:40:31,796 INFO spawned: 'worker-conferences_00' with pid 13672
docker_lamp_1 | 2026-04-20 06:40:31,799 INFO spawned: 'worker-crm-sync_00' with pid 13673
docker_lamp_1 | 2026-04-20 06:40:31,801 INFO spawned: 'worker-crm-update_00' with pid 13674
docker_lamp_1 | 2026-04-20 06:40:31,805 INFO spawned: 'worker-download_00' with pid 13675
docker_lamp_1 | 2026-04-20 06:40:31,808 INFO spawned: 'worker-emails_00' with pid 13676
docker_lamp_1 | 2026-04-20 06:40:31,812 INFO spawned: 'worker-es-update_00' with pid 13677
docker_lamp_1 | 2026-04-20 06:40:31,814 INFO spawned: 'worker-nudges_00' with pid 13679
docker_lamp_1 | 2026-04-20 06:40:32,821 INFO success: artisan-schedule_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-1_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-2_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-3_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-4_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-5_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: jiminny-worker-processing-delayed_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-analytics_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-audio_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-calendar_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-conferences_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-sync_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-crm-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-download_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-emails_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-es-update_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:32,824 INFO success: worker-nudges_00 entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
docker_lamp_1 | 2026-04-20 06:40:39 Jiminny\Jobs\Mailbox\CreateBatches [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:40:40 Jiminny\Jobs\Mailbox\CreateBatches ....... 505.47ms DONE
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:41:03 Running ['artisan' meeting-bot:schedule-bot] ... 3s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:41:06 Running ['artisan' dialers:monitor-activities] . 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:41:12 Running ['artisan' jiminny:monitor-social-accounts] 3s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:41:15 Running ['artisan' mailbox:skip-lists:refresh] . 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:41:23 Running ['artisan' mailbox:batch:process --max-batches=15] 10s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:42:04 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:06 Running ['artisan' dialers:monitor-activities] . 2s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:08 Running ['artisan' jiminny:monitor-social-accounts] 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:12 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:14 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:15 Running ['artisan' conference:monitor:count] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:42:16 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 | 2026-04-20 06:42:20 Jiminny\Jobs\Mailbox\CreateBatches [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:42:20 Jiminny\Jobs\Mailbox\CreateBatches [PASSWORD_DOTS] 44.79ms DONE
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:43:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:43:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:43:06 Running ['artisan' jiminny:monitor-social-accounts] 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:43:11 Running ['artisan' mailbox:skip-lists:refresh] . 3s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:43:14 Running ['artisan' mailbox:batch:process --max-batches=15] 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 12.38ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:43:20 Running ['artisan' calendar:sync --dateMode=daily] 2026-04-20 06:43:31 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... RUNNING
docker_lamp_1 | 2026-04-20 06:43:32 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... 1s DONE
docker_lamp_1 | 2026-04-20 06:43:34 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... RUNNING
docker_lamp_1 | 14s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' calendar:sync --dateMode=daily > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 | 2026-04-20 06:43:34 Jiminny\Jobs\Calendar\SyncCalendarEvents . 324.40ms DONE
docker_lamp_1 | 2026-04-20 06:43:34 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... RUNNING
docker_lamp_1 | 2026-04-20 06:43:43 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... 8s DONE
docker_lamp_1 | 2026-04-20 06:43:43 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:44 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 887.95ms DONE
docker_lamp_1 | 2026-04-20 06:43:44 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:45 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 902.35ms DONE
docker_lamp_1 | 2026-04-20 06:43:45 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:46 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 959.92ms DONE
docker_lamp_1 | 2026-04-20 06:43:46 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:47 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 855.24ms DONE
docker_lamp_1 | 2026-04-20 06:43:47 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:48 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 933.68ms DONE
docker_lamp_1 | 2026-04-20 06:43:48 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:48 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 952.67ms DONE
docker_lamp_1 | 2026-04-20 06:43:49 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:49 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 933.63ms DONE
docker_lamp_1 | 2026-04-20 06:43:49 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:51 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 1s DONE
docker_lamp_1 | 2026-04-20 06:43:51 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:52 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 1s DONE
docker_lamp_1 | 2026-04-20 06:43:52 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:53 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 1s DONE
docker_lamp_1 | 2026-04-20 06:43:53 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:54 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 986.58ms DONE
docker_lamp_1 | 2026-04-20 06:43:54 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents RUNNING
docker_lamp_1 | 2026-04-20 06:43:56 Jiminny\Jobs\Calendar\SyncOfficeRecurringEvents 2s DONE
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:44:03 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:06 Running ['artisan' jiminny:monitor-social-accounts] 2s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:09 Running ['artisan' mailbox:skip-lists:refresh] . 2s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:11 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:13 Running ['artisan' conference:monitor:count] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:44:14 Running ['artisan' crm:sync-objects] Syncing objects for Salesforce Team since 2026-03-19 20:14:10
docker_lamp_1 | Syncing objects for Hubspot since 2026-04-18 13:14:12
docker_lamp_1 | Syncing objects for Edge Communications since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Burlington Textiles Corp of America since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Uber since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Vonage since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Nexmo since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for SF since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for NewAccount since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for TestV since 2026-02-17 15:05:47
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for KioskAccount since 2026-02-17 15:06:27
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Pipedrive, Inc. since 2026-04-08 18:14:13
docker_lamp_1 | Syncing objects for Copper since 2026-04-18 13:14:12
docker_lamp_1 | Syncing objects for Pipedrive External Test since 2026-04-08 18:14:16
docker_lamp_1 | Syncing objects for InsightSquared since 2026-02-17 15:09:59
docker_lamp_1 | Syncing objects for Close since 2026-04-18 13:14:13
docker_lamp_1 | Syncing objects for DeewanyAccount since 2026-02-19 13:27:51
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Aircall Demo since 2026-02-17 15:06:51
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05
docker_lamp_1 | Syncing objects for GoStudent UAT since 2026-02-17 15:09:59
docker_lamp_1 | Syncing objects for Horen test since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed Hamadi since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed Testing since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed jiminny since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for JustCall since 2026-02-17 15:07:41
docker_lamp_1 | Syncing objects for Testers Inc since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Copper Team since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for TheBestPlace Ever since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Horen's apartments since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for New Org Test since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Pipedrive test org since 2026-02-19 13:27:42
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for BigChairs Inc. since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for GL 500 since 2026-02-19 13:26:04
docker_lamp_1 | Syncing objects for Loren X since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Laravel Company 2024 since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Barbara Fuchs since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Twilio Video since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for My Test Account 3000 since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for test since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Test since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Test salesforce auto sync since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for test ogg auto sync since 2026-02-17 15:07:16
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Globo Tech since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Laravel 11 Company 2024 since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 |
docker_lamp_1 | Error
docker_lamp_1 |
docker_lamp_1 | Call to a member function getAttribute() on null
docker_lamp_1 |
docker_lamp_1 | at app/Console/Commands/Crm/SyncObjects.php:42
docker_lamp_1 | 38▕
docker_lamp_1 | 39▕ foreach ($teams as $team) {
docker_lamp_1 | 40▕ $config = $team->getCrmConfiguration();
docker_lamp_1 | 41▕
docker_lamp_1 | ➜ 42▕ if ($config->getAttribute('sync_objects') === false) {
docker_lamp_1 | 43▕ continue;
docker_lamp_1 | 44▕ }
docker_lamp_1 | 45▕
docker_lamp_1 | 46▕ $this->info(
docker_lamp_1 |
docker_lamp_1 | 1 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:36
docker_lamp_1 | Jiminny\Console\Commands\Crm\SyncObjects::handle()
docker_lamp_1 |
docker_lamp_1 | 2 vendor/laravel/framework/src/Illuminate/Container/Util.php:43
docker_lamp_1 | Illuminate\Container\BoundMethod::Illuminate\Container\{closure}()
docker_lamp_1 |
docker_lamp_1 | 3 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:96
docker_lamp_1 | Illuminate\Container\Util::unwrapIfClosure(Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 4 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:35
docker_lamp_1 | Illuminate\Container\BoundMethod::callBoundMethod(Object(Illuminate\Foundation\Application), Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 5 vendor/laravel/framework/src/Illuminate/Container/Container.php:799
docker_lamp_1 | Illuminate\Container\BoundMethod::call(Object(Illuminate\Foundation\Application), [])
docker_lamp_1 |
docker_lamp_1 | 6 vendor/laravel/framework/src/Illuminate/Console/Command.php:211
docker_lamp_1 | Illuminate\Container\Container::call()
docker_lamp_1 |
docker_lamp_1 | 7 vendor/symfony/console/Command/Command.php:341
docker_lamp_1 | Illuminate\Console\Command::execute(Object(Symfony\Component\Console\Input\ArgvInput), Object(Illuminate\Console\OutputStyle))
docker_lamp_1 |
docker_lamp_1 | 8 vendor/laravel/framework/src/Illuminate/Console/Command.php:180
docker_lamp_1 | Symfony\Component\Console\Command\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Illuminate\Console\OutputStyle))
docker_lamp_1 |
docker_lamp_1 | 9 app/Console/Commands/Command.php:41
docker_lamp_1 | Illuminate\Console\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 10 vendor/symfony/console/Application.php:1117
docker_lamp_1 | Jiminny\Console\Commands\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 11 vendor/symfony/console/Application.php:356
docker_lamp_1 | Symfony\Component\Console\Application::doRunCommand(Object(Jiminny\Console\Commands\Crm\SyncObjects), Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 12 vendor/symfony/console/Application.php:195
docker_lamp_1 | Symfony\Component\Console\Application::doRun(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 13 vendor/laravel/framework/src/Illuminate/Foundation/Console/Kernel.php:198
docker_lamp_1 | Symfony\Component\Console\Application::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 14 vendor/laravel/framework/src/Illuminate/Foundation/Application.php:1235
docker_lamp_1 | Illuminate\Foundation\Console\Kernel::handle(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 15 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
docker_lamp_1 | [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:sync-objects > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 | 2026-04-20 06:44:15 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:15 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 98.77ms DONE
docker_lamp_1 | 2026-04-20 06:44:15 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:17 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-20 06:44:17 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:17 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 14.35ms DONE
docker_lamp_1 | 2026-04-20 06:44:17 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 14.13ms DONE
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 10.93ms DONE
docker_lamp_1 | 2026-04-20 06:44:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 13.44ms DONE
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 11.33ms DONE
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 10.38ms DONE
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 10.50ms DONE
docker_lamp_1 | 2026-04-20 06:44:19 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob RUNNING
docker_lamp_1 | 2026-04-20 06:44:20 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob 821.63ms DONE
docker_lamp_1 | 2026-04-20 06:44:20 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob RUNNING
docker_lamp_1 | 2026-04-20 06:44:20 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob 8.20ms DONE
docker_lamp_1 | 2026-04-20 06:44:25 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob RUNNING
docker_lamp_1 | 2026-04-20 06:44:25 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob 26.88ms DONE
docker_lamp_1 | 2026-04-20 06:44:30 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob RUNNING
docker_lamp_1 | 2026-04-20 06:44:30 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob 10.59ms DONE
docker_lamp_1 | 2026-04-20 06:44:35 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob RUNNING
docker_lamp_1 | 2026-04-20 06:44:35 Jiminny\Jobs\Crm\Salesforce\FetchSalesforceEntitiesJob 52.49ms DONE
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:45:06 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:45:13 Running ['artisan' dialers:monitor-activities] . 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:45:17 Running ['artisan' jiminny:monitor-social-accounts] 13s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:45:30 Running ['artisan' mailbox:skip-lists:refresh] 14s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:45:45 Running ['artisan' mailbox:batch:process --max-batches=15] 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:45:54 Running ['artisan' activity:purge-stale] ...... 10s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:purge-stale > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:05 Running ['artisan' mailbox:text-relay:sync] {
docker_lamp_1 | "error": "invalid_request",
docker_lamp_1 | "error_description": "Invalid impersonation \u0026quot;sub\u0026quot; field: @"
docker_lamp_1 | }
docker_lamp_1 | .... 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:text-relay:sync > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:14 Running ['artisan' conference:pre-meeting-notification] 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:pre-meeting-notification > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:18 Running ['artisan' conference:monitor:start] ... 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:start > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:27 Running ['artisan' conference:monitor:end] ..... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:end > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:29 Running ['artisan' jiminny:fix-hubspot-tokens] . 3s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:fix-hubspot-tokens > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' conference:pre-meeting-reminder] in background 1.51ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' conference:pre-meeting-reminder > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-805efb160ee8d9da02e60364ace7970eb2b35f31" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' hubspot:journal-poll --start] in background 3.11ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' hubspot:journal-poll --start > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-e26d77f915d2c55fe91ca4148a230e32eaa1865e" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:46:32 Running ['artisan' datadog:report:processing-sla-activities] 9s DONE
docker_lamp_1 | 🚀 Starting HubSpot journal polling service...
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' datadog:report:processing-sla-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:42 Running ['artisan' activity:sync --from='2026-04-20 06:29:00' --to='2026-04-20 06:45:00' --skipProviders='ringcentral' --skipProviders='avaya' --skipProviders='telus' --skipProviders='talkdesk'] 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:sync --from='2026-04-20 06:29:00' --to='2026-04-20 06:45:00' --skipProviders='ringcentral' --skipProviders='avaya' --skipProviders='telus' --skipProviders='talkdesk' > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:46:51 Running ['artisan' crm:bullhorn:ping --heartbeat] 2026-04-20 06:46:53 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:46:54 Jiminny\Jobs\Activity\SyncActivity ....... 959.91ms DONE
docker_lamp_1 | 2026-04-20 06:46:55 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:46:58 Jiminny\Notifications\Crm\UserAccountDisconnected RUNNING
docker_lamp_1 | 2026-04-20 06:46:59 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] 4s DONE
docker_lamp_1 | 2026-04-20 06:46:59 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:47:01 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] 2s DONE
docker_lamp_1 | 2026-04-20 06:47:01 Jiminny\Notifications\Crm\UserAccountDisconnected 2026-04-20 06:47:01 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 3s DONE
docker_lamp_1 | 2026-04-20 06:47:03 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-20 06:47:03 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:47:04 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-20 06:47:04 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-20 06:47:08 Jiminny\Jobs\Activity\SyncActivity [PASSWORD_DOTS] 4s DONE
docker_lamp_1 | 0 social account(s) to be processed ...
docker_lamp_1 |
docker_lamp_1 | Done!
docker_lamp_1 | 20s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:bullhorn:ping --heartbeat > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:47:11 Running ['artisan' nudges:send --silent] ....... 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' nudges:send --silent > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:47:17 Running ['artisan' jiminny:playlists:normalize-sort] 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:playlists:normalize-sort > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:48:06 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:48:11 Running ['artisan' dialers:monitor-activities] . 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:48:18 Running ['artisan' jiminny:monitor-social-accounts] 11s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:48:30 Running ['artisan' mailbox:skip-lists:refresh] 14s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:48:47,842Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][559] overhead, spent [265ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | 2026-04-20 06:48:45 Running ['artisan' mailbox:batch:process --max-batches=15] 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:48:50 Running ['artisan' conference:monitor:count] ... 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:48:58 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 14.85ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:48:58 Running ['artisan' crm:autolog-delayed] Dispatched autolog delayed jobs for all applicable teams:
docker_lamp_1 | ....... 10s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:autolog-delayed > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:50:07 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:50:13 Running ['artisan' dialers:monitor-activities] . 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:50:18 Running ['artisan' jiminny:monitor-social-accounts] 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:50:25 Running ['artisan' mailbox:skip-lists:refresh] 17s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:50:43 Running ['artisan' mailbox:batch:process --max-batches=15] 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:50:51 Running ['artisan' conference:monitor:count] .. 11s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:03 Running ['artisan' activity:purge-stale] ...... 12s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:purge-stale > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:15 Running ['artisan' mailbox:text-relay:sync] {
docker_lamp_1 | "error": "invalid_request",
docker_lamp_1 | "error_description": "Invalid impersonation \u0026quot;sub\u0026quot; field: @"
docker_lamp_1 | }
docker_lamp_1 | .... 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:text-relay:sync > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:23 Running ['artisan' conference:pre-meeting-notification] 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:pre-meeting-notification > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:30 Running ['artisan' conference:monitor:start] ... 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:start > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:39 Running ['artisan' conference:monitor:end] ..... 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:end > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:47 Running ['artisan' jiminny:fix-hubspot-tokens] . 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:fix-hubspot-tokens > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' conference:pre-meeting-reminder] in background 3.72ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' conference:pre-meeting-reminder > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-805efb160ee8d9da02e60364ace7970eb2b35f31" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' hubspot:journal-poll --start] in background 2.73ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' hubspot:journal-poll --start > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-e26d77f915d2c55fe91ca4148a230e32eaa1865e" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 | 2026-04-20 06:51:55 Running ['artisan' jiminny:transcription:retry-failed] No failed transcriptions found.
docker_lamp_1 | 🚀 Starting HubSpot journal polling service...
docker_lamp_1 | 12s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:transcription:retry-failed > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:52:08 Running ['artisan' crm:reset-governor] [PASSWORD_DOTS] 17s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:reset-governor > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:52:26 Running ['artisan' crm:bullhorn:ping --heartbeat] 0 social account(s) to be processed ...
docker_lamp_1 |
docker_lamp_1 | Done!
docker_lamp_1 | 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:bullhorn:ping --heartbeat > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:53:07 Running ['artisan' meeting-bot:schedule-bot] ... 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:53:15 Running ['artisan' dialers:monitor-activities] 11s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:53:26 Running ['artisan' jiminny:monitor-social-accounts] 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:53:35 Running ['artisan' mailbox:skip-lists:refresh] . 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:53:40 Running ['artisan' mailbox:batch:process --max-batches=15] 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:53:50 Running ['artisan' mailbox:batch:retry-failed --max-batches=15] in background 16.19ms DONE
docker_lamp_1 | ⇂ ('/usr/local/bin/php' 'artisan' mailbox:batch:retry-failed --max-batches=15 > '/proc/1/fd/1' 2>&1 ; '/usr/local/bin/php' 'artisan' schedule:finish "framework/schedule-390defd641effba0f73a895e426ded4cf2ba7f11" "$?") > '/dev/null' 2>&1 &
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 |
docker_lamp_1 | 2026-04-20 06:54:09 Running ['artisan' meeting-bot:schedule-bot] ... 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:15 Running ['artisan' dialers:monitor-activities] . 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:19 Running ['artisan' jiminny:monitor-social-accounts] 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:24 Running ['artisan' mailbox:skip-lists:refresh] . 4s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:28 Running ['artisan' mailbox:batch:process --max-batches=15] 12s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:40 Running ['artisan' conference:monitor:count] ... 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:46 Running ['artisan' activity:aircall:check-and-renew] 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:aircall:check-and-renew > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-20 06:54:53 Running ['artisan' track:retry-failed-downloads] 11s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' track:retry-failed-downloads > '/proc/1/fd/...
|
NULL
|
|
52462
|
1133
|
41
|
2026-04-20T07:04:07.291770+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776668647291_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
rirerox•• 0IY [SRD-6793] Les Mills activity typeJY rirerox•• 0IY [SRD-6793] Les Mills activity typeJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tabu Userpilot | EventsBookmarksQ Search bookmarksv x bookmarks loolba$ Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....v E Bookmarks Menu> → Mozilla FirefoxOther Bookmarksnny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909® For you• Recent# Starred0+ AppsQ Spaces+**Service-DeskJiminny (New)+ ***I 0 Platform TeamIID SE KanbanII Capture TeamŒ Enterprise Stability I…..Processing Team= More spaces= FiltersH DashboardsC÷ Operations*) Confluence88 Teams"= Customise sidebarspaces / Jiminny (New)Plaworm leam® Summary& TimelineE BacklogQ Search board00001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed_at periodCOST-EFFECTIVE AND FASTER NUDGESBacklod#JY-2048915 .000=Investiaate and fix why exceedMAINTENANCEReady Tor DevJY-205641 80 0000 =8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216.000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-20157I?eeee —Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00•=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909Іриммиу"$ JY-20577(BE] Add flag in AI Reports list for deleterightsG JY-20578 [FE] Add delete button$ JY-20579 [BE] Add new report type in filters options$ JY-20580 [FE] Rename column Shared9, JY-20581 [FEl Rework Shared Tooltio info9JY-20582 [BE+AI+Infra] Create new queue$ JY-20583 [BE] Add period to cron job$JY-20584 (BE) Change search report in cron job query$ JY-20585 [QA] Create test cases$ JY-20586 [QA] Manual Testing9 JY-20626 Intearation testingQ, JY-20678 [FEl Add custom Expiration table cell&JY-20679[BE] Add dash symbol between name andperiod in the title$ JY-20680 [FE] Add Info KB link in page header#*JY-20694when changina treauency or repont templa...& JY-20700Split Userpilot events for for AJ Reports andExec RevortsLinked work itemsis cloned hv# JY-20458 CRUD reports with Ask Jiminnyis tested bv= M= M= M— MI= M= MI= M=M= MI= M= MI= M= M= M• JY-20649 Generate Ask Jiminny Report and display it in AI Rep...JY-20656 Expired Ask Jiminny report template is automatically ….O JY-20660 Non-creator cannot delete generated Ask Jiminny re…O JY-20651 Generated Ask Jiminny Report PDF displays Jiminny I...DEPLOYED VDONEDONE VDONEDONEDONE VDONE VDONEDONEVDONEVDONE VIN DEV VDONEvDONE VIN DEVIN DEV VIn QADetailsAssigneeReporterDeveloomeniReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsOrganisatioPriorityFix versionsSprintDaysNeed QA$0+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 buildsA Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1Yes5 davs aqoOPEN]Dally - Platorm• Im len100% 52• Mon ZU Aor 10.:04.013 Ask RovoloymentsE Archived work items# PagesMore 3Complete sprintGroup: QueriesEPTANCEDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-206321∞ee—A.J Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet* JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedE JY-199671• •= 1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
NULL
|
-2656918481667093290
|
NULL
|
click
|
ocr
|
NULL
|
rirerox•• 0IY [SRD-6793] Les Mills activity typeJY rirerox•• 0IY [SRD-6793] Les Mills activity typeJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tabu Userpilot | EventsBookmarksQ Search bookmarksv x bookmarks loolba$ Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....v E Bookmarks Menu> → Mozilla FirefoxOther Bookmarksnny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909® For you• Recent# Starred0+ AppsQ Spaces+**Service-DeskJiminny (New)+ ***I 0 Platform TeamIID SE KanbanII Capture TeamŒ Enterprise Stability I…..Processing Team= More spaces= FiltersH DashboardsC÷ Operations*) Confluence88 Teams"= Customise sidebarspaces / Jiminny (New)Plaworm leam® Summary& TimelineE BacklogQ Search board00001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed_at periodCOST-EFFECTIVE AND FASTER NUDGESBacklod#JY-2048915 .000=Investiaate and fix why exceedMAINTENANCEReady Tor DevJY-205641 80 0000 =8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216.000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-20157I?eeee —Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00•=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909Іриммиу"$ JY-20577(BE] Add flag in AI Reports list for deleterightsG JY-20578 [FE] Add delete button$ JY-20579 [BE] Add new report type in filters options$ JY-20580 [FE] Rename column Shared9, JY-20581 [FEl Rework Shared Tooltio info9JY-20582 [BE+AI+Infra] Create new queue$ JY-20583 [BE] Add period to cron job$JY-20584 (BE) Change search report in cron job query$ JY-20585 [QA] Create test cases$ JY-20586 [QA] Manual Testing9 JY-20626 Intearation testingQ, JY-20678 [FEl Add custom Expiration table cell&JY-20679[BE] Add dash symbol between name andperiod in the title$ JY-20680 [FE] Add Info KB link in page header#*JY-20694when changina treauency or repont templa...& JY-20700Split Userpilot events for for AJ Reports andExec RevortsLinked work itemsis cloned hv# JY-20458 CRUD reports with Ask Jiminnyis tested bv= M= M= M— MI= M= MI= M=M= MI= M= MI= M= M= M• JY-20649 Generate Ask Jiminny Report and display it in AI Rep...JY-20656 Expired Ask Jiminny report template is automatically ….O JY-20660 Non-creator cannot delete generated Ask Jiminny re…O JY-20651 Generated Ask Jiminny Report PDF displays Jiminny I...DEPLOYED VDONEDONE VDONEDONEDONE VDONE VDONEDONEVDONEVDONE VIN DEV VDONEvDONE VIN DEVIN DEV VIn QADetailsAssigneeReporterDeveloomeniReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsOrganisatioPriorityFix versionsSprintDaysNeed QA$0+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 buildsA Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1Yes5 davs aqoOPEN]Dally - Platorm• Im len100% 52• Mon ZU Aor 10.:04.013 Ask RovoloymentsE Archived work items# PagesMore 3Complete sprintGroup: QueriesEPTANCEDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-206321∞ee—A.J Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet* JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedE JY-199671• •= 1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
52461
|
|
52478
|
1132
|
88
|
2026-04-20T07:05:13.946423+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776668713946_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
iTerm2ShellEditViewSessionScriptsProfilesWindowHel iTerm2ShellEditViewSessionScriptsProfilesWindowHelp§ Preparation for Refi... in 4 h 55 mEU (ssh)DOCKERAPP (-zsh)-zsh|DOCKER (-zsh)ugins"reporting"figuration."config"],"message" : "Found'server.host: \"O\"' in Kibana conincompatible with Reporting.enable Reporting to work,'xpack.reporting.kibanaServer.hostname: [IP_ADDRESS]' is beingautomatically to theconfiguration.canchange thesetting to'server.host: [IP_ADDRESS]'xpack.reporting.kibanaServer.hostname: [IP_ADDRESS]'in kibana.yml to prevent thiskibana:"log","@timestamp": "2026-04-20T06:40:04Z": ["warning", "plugins""reporting""config"],"message": "Chromiumsandbox provides an additionallayerprotection,not supported8.3.2011\n OS. Automaticallysetting"xpack.reporting.capture.browser.chromium.disableSandbox: true'kibana"@timestamp":"2026-04-20T06:40:04Z","tags" : ["warning", "plugins""actions"],'"pid" :7, "message" : "APIs aredisabled due to the Encrypted Saved Objectsplugin usingencryption key. Please set xpack.encryptedSaved0bjects.encryptionkey in kibana.yml."}kibana1 {"type" : "log""@timestamp": "2026-04-20T06:40:04Z", "tags" : ["warning", "plugins""alerts""plugins""alerting"],"pid":7,"message": "APIs are disabled due to the EncryptedSavedObjectsplugin usingephemeralencryption key. Please set xpack.encryptedSaved0bjects.encryptionkey in kibana.yml.kibana1 {"type" : "log","@timestamp" : "2026-04-20T06:40: 05Z","tags" : ["info", "plugins"monitoring""monitoring"], "pid" :7, "message" : "config sourced from: production clusteXIkibanaI {"type" : "log""@timestamp": "2026-04-20T06:40:05Z""tags" : ["info"objects-service"],"pid" :7,"message": "Waiting untilall Elasticsearch nodes are compatiblKibana before starting savedobjects migrations.kibana1 {"type" : "log""@timestamp": "2026-04-20T06:40:05Z","tags" : ["info","savedobjects-service"],"pid":7,*"message": "Starting savedobjects migrations"}kibana1 {"type" : "log""@timestamp" : "2026-04-20T06:40:06Z" , "tags" : ["info", "plugins-system"], "pid" :7, "message" : "Starting [96] plugins:[taskManager,licensing,globalSearch,globalSearchProviders, code,mapsLegacy, usageCollection, xpackLegacy, telemetryCollectionManager, telemetry, telemetryCollectionXpack, kibanaUsageCollection, securityOss, kibanaLegacy, newsfeed, translations, share, legacyExport, embeddable, uiActionsEnhanced, expressions, data,home, observability,cloud, console, consoleExtensions, apmOss, searchprofiler,painlessLab,grokdebugger,management, indexPatternManagement,advancedSettings,fileUpload, savedObjects, dashboard, visualizations, visTypeVega, visTypelimelion, timelion, features, upgradeAssistant, security,snapshotRestore, encryptedSavedObjects, ingestManager, indexManagement, remoteClusters,crossClusterReplication,indexLifecycleManagement, enterpriseSearch, dashboardMode,beatsManagement, transform, ingestPipelines,maps,licenseManagement,graph,dataEnhanced, visTypeTable, visTypeMarkdown, tileMap, regionMap, inputControlVis,visualize, esUiShared,bfetch, canvas, charts, lens, visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric, watcher,discover,discoverEnhanced, savedObjectsManagement, spaces, reporting, eventLog, actions, case,alerts,stackAlerts,triggersActionsUi,lists,ml, securitySolution, infra, monitoring, logstash, apm, uptime] "}kibana1 {"type" : "log", "@timestamp" : "2026-04-20T06:40:06Z"ns"taskManager","tags" : ["info", "plugi"taskManager"],"pid":7,"message" : "TaskManager is identified by the KibanaUUID: bf01f365-e094-4cde-940d-3e0db65fa22a"}100% [8• Mon 20 Apr 10:05:14181• 86-zsh|87•985'* Review screenpipe u...PROD (ssh)[URL_WITH_CREDENTIALS] 0New release'24.04.4 LTS' available.'do-release-upgrade' to upgrade to it.*** System restart required ***Last login: Fri Apr 17 05:56:312026 from 212.39.71.189lukas@jiminny-eu-bastion:~$T4STAGE (-zsh)Last login: Thu Apr 16 15:43:43 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ ]T5QA (-zsh)Last login: Thu Apr 16 15:43:43 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentsFE (-zsh)Last login: Thu Apr 16 15:48:07 on ttys004Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentsLukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ l17 EXT (-zsh)Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml fikas@Lukas-Kovaliks-MacBook-Pro-Jiminny=intsSTAGEFRONTENDEXTENSION...
|
NULL
|
-753498032432203396
|
NULL
|
visual_change
|
ocr
|
NULL
|
iTerm2ShellEditViewSessionScriptsProfilesWindowHel iTerm2ShellEditViewSessionScriptsProfilesWindowHelp§ Preparation for Refi... in 4 h 55 mEU (ssh)DOCKERAPP (-zsh)-zsh|DOCKER (-zsh)ugins"reporting"figuration."config"],"message" : "Found'server.host: \"O\"' in Kibana conincompatible with Reporting.enable Reporting to work,'xpack.reporting.kibanaServer.hostname: [IP_ADDRESS]' is beingautomatically to theconfiguration.canchange thesetting to'server.host: [IP_ADDRESS]'xpack.reporting.kibanaServer.hostname: [IP_ADDRESS]'in kibana.yml to prevent thiskibana:"log","@timestamp": "2026-04-20T06:40:04Z": ["warning", "plugins""reporting""config"],"message": "Chromiumsandbox provides an additionallayerprotection,not supported8.3.2011\n OS. Automaticallysetting"xpack.reporting.capture.browser.chromium.disableSandbox: true'kibana"@timestamp":"2026-04-20T06:40:04Z","tags" : ["warning", "plugins""actions"],'"pid" :7, "message" : "APIs aredisabled due to the Encrypted Saved Objectsplugin usingencryption key. Please set xpack.encryptedSaved0bjects.encryptionkey in kibana.yml."}kibana1 {"type" : "log""@timestamp": "2026-04-20T06:40:04Z", "tags" : ["warning", "plugins""alerts""plugins""alerting"],"pid":7,"message": "APIs are disabled due to the EncryptedSavedObjectsplugin usingephemeralencryption key. Please set xpack.encryptedSaved0bjects.encryptionkey in kibana.yml.kibana1 {"type" : "log","@timestamp" : "2026-04-20T06:40: 05Z","tags" : ["info", "plugins"monitoring""monitoring"], "pid" :7, "message" : "config sourced from: production clusteXIkibanaI {"type" : "log""@timestamp": "2026-04-20T06:40:05Z""tags" : ["info"objects-service"],"pid" :7,"message": "Waiting untilall Elasticsearch nodes are compatiblKibana before starting savedobjects migrations.kibana1 {"type" : "log""@timestamp": "2026-04-20T06:40:05Z","tags" : ["info","savedobjects-service"],"pid":7,*"message": "Starting savedobjects migrations"}kibana1 {"type" : "log""@timestamp" : "2026-04-20T06:40:06Z" , "tags" : ["info", "plugins-system"], "pid" :7, "message" : "Starting [96] plugins:[taskManager,licensing,globalSearch,globalSearchProviders, code,mapsLegacy, usageCollection, xpackLegacy, telemetryCollectionManager, telemetry, telemetryCollectionXpack, kibanaUsageCollection, securityOss, kibanaLegacy, newsfeed, translations, share, legacyExport, embeddable, uiActionsEnhanced, expressions, data,home, observability,cloud, console, consoleExtensions, apmOss, searchprofiler,painlessLab,grokdebugger,management, indexPatternManagement,advancedSettings,fileUpload, savedObjects, dashboard, visualizations, visTypeVega, visTypelimelion, timelion, features, upgradeAssistant, security,snapshotRestore, encryptedSavedObjects, ingestManager, indexManagement, remoteClusters,crossClusterReplication,indexLifecycleManagement, enterpriseSearch, dashboardMode,beatsManagement, transform, ingestPipelines,maps,licenseManagement,graph,dataEnhanced, visTypeTable, visTypeMarkdown, tileMap, regionMap, inputControlVis,visualize, esUiShared,bfetch, canvas, charts, lens, visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric, watcher,discover,discoverEnhanced, savedObjectsManagement, spaces, reporting, eventLog, actions, case,alerts,stackAlerts,triggersActionsUi,lists,ml, securitySolution, infra, monitoring, logstash, apm, uptime] "}kibana1 {"type" : "log", "@timestamp" : "2026-04-20T06:40:06Z"ns"taskManager","tags" : ["info", "plugi"taskManager"],"pid":7,"message" : "TaskManager is identified by the KibanaUUID: bf01f365-e094-4cde-940d-3e0db65fa22a"}100% [8• Mon 20 Apr 10:05:14181• 86-zsh|87•985'* Review screenpipe u...PROD (ssh)[URL_WITH_CREDENTIALS] 0New release'24.04.4 LTS' available.'do-release-upgrade' to upgrade to it.*** System restart required ***Last login: Fri Apr 17 05:56:312026 from 212.39.71.189lukas@jiminny-eu-bastion:~$T4STAGE (-zsh)Last login: Thu Apr 16 15:43:43 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ ]T5QA (-zsh)Last login: Thu Apr 16 15:43:43 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentsFE (-zsh)Last login: Thu Apr 16 15:48:07 on ttys004Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentsLukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ l17 EXT (-zsh)Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml fikas@Lukas-Kovaliks-MacBook-Pro-Jiminny=intsSTAGEFRONTENDEXTENSION...
|
52477
|
|
52479
|
NULL
|
0
|
2026-04-20T07:05:17.901416+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776668717901_m2.jpg...
|
iTerm2
|
screenpipe"
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
ggml_metal_init: use fusion = true
ggml_me ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
2026-04-18T14:43:07.918412Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks
2026-04-18T14:44:35.514401Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-8593116969702213352, trigger=visual_change)
2026-04-18T14:44:35.891413Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-8593116969702213352, trigger=visual_change)
2026-04-18T14:45:06.077184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=visual_change)
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
2026-04-18T14:45:09.423485Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks
2026-04-18T14:45:40.853509Z WARN screenpipe_db::db: ImmediateTx dropped without commit — rolling back
2026-04-18T14:45:45.829292Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=click)
2026-04-18T14:45:45.931557Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6130095137839516019, trigger=click)
zsh: terminated npx screenpipe@latest record --ignored-windows "Boosteroid"
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sp-start
detected hardware tier: Mid
warning: parakeet is not supported on this platform, using whisper-tiny instead
2026-04-20T09:09:02.916408Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store
checking permissions...
screen recording: ok
accessibility: ok
2026-04-20T09:09:03.042870Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6
2026-04-20T09:09:04.217747Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor
2026-04-20T09:09:04.219685Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)
2026-04-20T09:09:04.220577Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)
2026-04-20T09:09:04.314160Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true
2026-04-20T09:09:04.314251Z INFO screenpipe: meeting detector enabled — independent of transcription mode
2026-04-20T09:09:04.314473Z INFO screenpipe: API server listening on [IP_ADDRESS]:3030 (localhost only)
2026-04-20T09:09:04.314486Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key
2026-04-20T09:09:04.314428Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)
2026-04-20T09:09:04.314439Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)
2026-04-20T09:09:04.314563Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager
2026-04-20T09:09:04.370709Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))
2026-04-20T09:09:04.383079Z INFO screenpipe_core::pipes: loaded pipe: day-recap
2026-04-20T09:09:04.384136Z INFO screenpipe_core::pipes: loaded pipe: standup-update
2026-04-20T09:09:04.384745Z INFO screenpipe_core::pipes: loaded pipe: ai-habits
2026-04-20T09:09:04.385360Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown
2026-04-20T09:09:04.385876Z INFO screenpipe_core::pipes: loaded pipe: video-export
2026-04-20T09:09:04.386843Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary
2026-04-20T09:09:04.386950Z INFO screenpipe_core::pipes: loaded 6 pipes from "/Users/lukas/.screenpipe/pipes"
_
__________________ ___ ____ ____ (_____ ___
/ ___/ ___/ ___/ _ \/ _ \/ __ \ / __ \/ / __ \/ _ \
(__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/
/____/\___/_/ \___/\___/_/ /_/ / .___/_/ .___/\___/
/_/ /_/
power AI by everything you've seen, said or heard
open source | runs locally | developer friendly
┌────────────────────────┬────────────────────────────────────┐
│ setting │ value │
├────────────────────────┼────────────────────────────────────┤
│ audio chunk duration │ 30 seconds │
│ port │ 3030 │
│ audio disabled │ true │
│ vision disabled │ false │
│ pause on DRM content │ false │
│ audio engine │ Parakeet │
│ vad engine │ Silero │
│ data directory │ /Users/lukas/.screenpipe │
│ debug mode │ false │
│ telemetry │ true │
│ use pii removal │ true │
│ use all monitors │ true │
│ ignored windows │ ["Boosteroid"] │
│ included windows │ [] │
│ cloud sync │ disabled │
│ auto-destruct pid │ 0 │
│ deepgram key │ not set │
│ api auth │ enabled │
2026-04-20T09:09:04.390629Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)
2026-04-20T09:09:04.411414Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh
│ encrypt secrets │ disabled │
│ retention days │ 14 │
├────────────────────────┼────────────────────────────────────┤
│ languages │ │
│ │ all languages │
├────────────────────────┼────────────────────────────────────┤
│ monitors │ │
│ │ id: 1 │
│ │ id: 2 │
├────────────────────────┼────────────────────────────────────┤
│ audio devices │ │
│ │ disabled │
└────────────────────────┴────────────────────────────────────┘
you are using local processing. all your data stays on your computer.
warning: telemetry is enabled. only error-level data will be sent.
to disable, use the --disable-telemetry flag.
check latest changes here: https://github.com/screenpipe/screenpipe/releases
2026-04-20T09:09:04.444803Z INFO screenpipe: starting UI event capture
2026-04-20T09:09:04.482387Z INFO screenpipe_engine::ui_recorder: Starting UI event capture
2026-04-20T09:09:04.509116Z INFO screenpipe_engine::ui_recorder: UI recording session started: f0f4e024-5bd3-4a71-bf92-6feec89df1b0
2026-04-20T09:09:04.509231Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)
2026-04-20T09:09:04.509296Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 06:09:04.509295 UTC to 2026-04-20 06:09:04.509295 UTC)
2026-04-20T09:09:04.510396Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)
2026-04-20T09:09:04.511440Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 0 frame entries, coverage from 2026-04-19 06:09:04.509295 UTC
2026-04-20T09:09:04.519984Z INFO screenpipe_engine::server: Server listening on [IP_ADDRESS]:3030
2026-04-20T09:09:04.547289Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030
2026-04-20T09:09:04.965425Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)
2026-04-20T09:09:04.965463Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)
2026-04-20T09:09:04.965503Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)
2026-04-20T09:09:05.226982Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)
2026-04-20T09:09:05.227016Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)
2026-04-20T09:09:05.227136Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)
2026-04-20T09:09:05.227375Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)
2026-04-20T09:09:07.300820Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 3 excluded)
2026-04-20T09:09:07.600290Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=51507, dur=136ms
2026-04-20T09:09:07.683736Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 3 excluded)
2026-04-20T09:09:07.989000Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=51508, dur=185ms
2026-04-20T09:10:09.231456Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=197 elapsed=4.915444333s
2026-04-20T09:10:09.231772Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 197 eligible frames
2026-04-20T09:10:14.691743Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 83 frames, 16.3MB → 2.5MB (6.5x), 83 JPEGs deleted
2026-04-20T09:10:15.677833Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:17.628368Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:17.716440Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:21.386286Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 100 frames, 17.9MB → 6.8MB (2.6x), 100 JPEGs deleted
2026-04-20T09:10:22.685406Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 2.2MB → 0.9MB (2.4x), 14 JPEGs deleted
2026-04-20T09:10:23.507120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:42.391615Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:42.475795Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:44.750359Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:44.980679Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:53.686192Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:53.738064Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:56.291827Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:56.451083Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:12:46.023481Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)
2026-04-20T09:12:49.049656Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)
2026-04-20T09:13:23.025287Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=596503974090759224, trigger=click)
2026-04-20T09:13:28.322087Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3622643354436184557, trigger=visual_change)
2026-04-20T09:14:35.739274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3577748614007823836, trigger=click)
2026-04-20T09:15:24.523245Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=3 elapsed=1.840652833s
2026-04-20T09:15:24.523440Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 3 eligible frames
2026-04-20T09:15:37.319126Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7719336893453902381, trigger=click)
2026-04-20T09:16:25.021906Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2226463793404778776, trigger=visual_change)
2026-04-20T09:16:40.481038Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-2449473341097903678, trigger=visual_change)
2026-04-20T09:17:01.897197Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)
2026-04-20T09:17:11.727852Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:11.880765Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:12.886271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:12.961362Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:49.590890Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-5042772765556709364, trigger=visual_change)
2026-04-20T09:18:38.516980Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3873630181551472921, trigger=click)
2026-04-20T09:18:52.171689Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-3192006669261292210, trigger=click)
2026-04-20T09:18:52.339913Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)
2026-04-20T09:19:18.962835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)
2026-04-20T09:19:21.997050Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)
2026-04-20T09:20:27.568575Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=33 elapsed=3.045052625s
2026-04-20T09:20:27.568689Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 33 eligible frames
2026-04-20T09:20:28.589270Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 3.1MB → 0.7MB (4.6x), 14 JPEGs deleted
2026-04-20T09:20:30.840110Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 19 frames, 3.0MB → 1.1MB (2.8x), 19 JPEGs deleted
2026-04-20T09:20:37.786800Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"ggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\n2026-04-18T14:43:07.918412Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks\n2026-04-18T14:44:35.514401Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-8593116969702213352, trigger=visual_change)\n2026-04-18T14:44:35.891413Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-8593116969702213352, trigger=visual_change)\n2026-04-18T14:45:06.077184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=visual_change)\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\n2026-04-18T14:45:09.423485Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks\n2026-04-18T14:45:40.853509Z WARN screenpipe_db::db: ImmediateTx dropped without commit — rolling back\n2026-04-18T14:45:45.829292Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=click)\n2026-04-18T14:45:45.931557Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6130095137839516019, trigger=click)\nzsh: terminated npx screenpipe@latest record --ignored-windows \"Boosteroid\"\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sp-start\ndetected hardware tier: Mid\nwarning: parakeet is not supported on this platform, using whisper-tiny instead\n2026-04-20T09:09:02.916408Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store\nchecking permissions...\n screen recording: ok\n accessibility: ok\n2026-04-20T09:09:03.042870Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6\n2026-04-20T09:09:04.217747Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor\n2026-04-20T09:09:04.219685Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)\n2026-04-20T09:09:04.220577Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)\n2026-04-20T09:09:04.314160Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true\n2026-04-20T09:09:04.314251Z INFO screenpipe: meeting detector enabled — independent of transcription mode\n2026-04-20T09:09:04.314473Z INFO screenpipe: API server listening on 127.0.0.1:3030 (localhost only)\n2026-04-20T09:09:04.314486Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key\n2026-04-20T09:09:04.314428Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)\n2026-04-20T09:09:04.314439Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)\n2026-04-20T09:09:04.314563Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager\n2026-04-20T09:09:04.370709Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))\n2026-04-20T09:09:04.383079Z INFO screenpipe_core::pipes: loaded pipe: day-recap\n2026-04-20T09:09:04.384136Z INFO screenpipe_core::pipes: loaded pipe: standup-update\n2026-04-20T09:09:04.384745Z INFO screenpipe_core::pipes: loaded pipe: ai-habits\n2026-04-20T09:09:04.385360Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown\n2026-04-20T09:09:04.385876Z INFO screenpipe_core::pipes: loaded pipe: video-export\n2026-04-20T09:09:04.386843Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary\n2026-04-20T09:09:04.386950Z INFO screenpipe_core::pipes: loaded 6 pipes from \"/Users/lukas/.screenpipe/pipes\"\n\n\n\n _ \n __________________ ___ ____ ____ (_____ ___ \n / ___/ ___/ ___/ _ \\/ _ \\/ __ \\ / __ \\/ / __ \\/ _ \\\n (__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/\n/____/\\___/_/ \\___/\\___/_/ /_/ / .___/_/ .___/\\___/ \n /_/ /_/ \n\n\n\npower AI by everything you've seen, said or heard\nopen source | runs locally | developer friendly\n\n\n┌────────────────────────┬────────────────────────────────────┐\n│ setting │ value │\n├────────────────────────┼────────────────────────────────────┤\n│ audio chunk duration │ 30 seconds │\n│ port │ 3030 │\n│ audio disabled │ true │\n│ vision disabled │ false │\n│ pause on DRM content │ false │\n│ audio engine │ Parakeet │\n│ vad engine │ Silero │\n│ data directory │ /Users/lukas/.screenpipe │\n│ debug mode │ false │\n│ telemetry │ true │\n│ use pii removal │ true │\n│ use all monitors │ true │\n│ ignored windows │ [\"Boosteroid\"] │\n│ included windows │ [] │\n│ cloud sync │ disabled │\n│ auto-destruct pid │ 0 │\n│ deepgram key │ not set │\n│ api auth │ enabled │\n2026-04-20T09:09:04.390629Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)\n2026-04-20T09:09:04.411414Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh\n│ encrypt secrets │ disabled │\n│ retention days │ 14 │\n├────────────────────────┼────────────────────────────────────┤\n│ languages │ │\n│ │ all languages │\n├────────────────────────┼────────────────────────────────────┤\n│ monitors │ │\n│ │ id: 1 │\n│ │ id: 2 │\n├────────────────────────┼────────────────────────────────────┤\n│ audio devices │ │\n│ │ disabled │\n└────────────────────────┴────────────────────────────────────┘\nyou are using local processing. all your data stays on your computer.\n\nwarning: telemetry is enabled. only error-level data will be sent.\nto disable, use the --disable-telemetry flag.\n\ncheck latest changes here: https://github.com/screenpipe/screenpipe/releases\n2026-04-20T09:09:04.444803Z INFO screenpipe: starting UI event capture\n2026-04-20T09:09:04.482387Z INFO screenpipe_engine::ui_recorder: Starting UI event capture\n2026-04-20T09:09:04.509116Z INFO screenpipe_engine::ui_recorder: UI recording session started: f0f4e024-5bd3-4a71-bf92-6feec89df1b0\n2026-04-20T09:09:04.509231Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)\n2026-04-20T09:09:04.509296Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 06:09:04.509295 UTC to 2026-04-20 06:09:04.509295 UTC)\n2026-04-20T09:09:04.510396Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)\n2026-04-20T09:09:04.511440Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 0 frame entries, coverage from 2026-04-19 06:09:04.509295 UTC\n2026-04-20T09:09:04.519984Z INFO screenpipe_engine::server: Server listening on 127.0.0.1:3030\n2026-04-20T09:09:04.547289Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030\n2026-04-20T09:09:04.965425Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)\n2026-04-20T09:09:04.965463Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)\n2026-04-20T09:09:04.965503Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)\n2026-04-20T09:09:05.226982Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)\n2026-04-20T09:09:05.227016Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)\n2026-04-20T09:09:05.227136Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)\n2026-04-20T09:09:05.227375Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)\n2026-04-20T09:09:07.300820Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 3 excluded)\n2026-04-20T09:09:07.600290Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=51507, dur=136ms\n2026-04-20T09:09:07.683736Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 3 excluded)\n2026-04-20T09:09:07.989000Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=51508, dur=185ms\n2026-04-20T09:10:09.231456Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=197 elapsed=4.915444333s\n2026-04-20T09:10:09.231772Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 197 eligible frames\n2026-04-20T09:10:14.691743Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 83 frames, 16.3MB → 2.5MB (6.5x), 83 JPEGs deleted\n2026-04-20T09:10:15.677833Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:17.628368Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:17.716440Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:21.386286Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 100 frames, 17.9MB → 6.8MB (2.6x), 100 JPEGs deleted\n2026-04-20T09:10:22.685406Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 2.2MB → 0.9MB (2.4x), 14 JPEGs deleted\n2026-04-20T09:10:23.507120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:42.391615Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:42.475795Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:44.750359Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:44.980679Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:53.686192Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:53.738064Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:56.291827Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:56.451083Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:12:46.023481Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)\n2026-04-20T09:12:49.049656Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)\n2026-04-20T09:13:23.025287Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=596503974090759224, trigger=click)\n2026-04-20T09:13:28.322087Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3622643354436184557, trigger=visual_change)\n2026-04-20T09:14:35.739274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3577748614007823836, trigger=click)\n2026-04-20T09:15:24.523245Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=3 elapsed=1.840652833s\n2026-04-20T09:15:24.523440Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 3 eligible frames\n2026-04-20T09:15:37.319126Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7719336893453902381, trigger=click)\n2026-04-20T09:16:25.021906Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2226463793404778776, trigger=visual_change)\n2026-04-20T09:16:40.481038Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-2449473341097903678, trigger=visual_change)\n2026-04-20T09:17:01.897197Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)\n2026-04-20T09:17:11.727852Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:11.880765Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:12.886271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:12.961362Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:49.590890Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-5042772765556709364, trigger=visual_change)\n2026-04-20T09:18:38.516980Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3873630181551472921, trigger=click)\n2026-04-20T09:18:52.171689Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-3192006669261292210, trigger=click)\n2026-04-20T09:18:52.339913Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)\n2026-04-20T09:19:18.962835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)\n2026-04-20T09:19:21.997050Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)\n2026-04-20T09:20:27.568575Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=33 elapsed=3.045052625s\n2026-04-20T09:20:27.568689Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 33 eligible frames\n2026-04-20T09:20:28.589270Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 3.1MB → 0.7MB (4.6x), 14 JPEGs deleted\n2026-04-20T09:20:30.840110Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 19 frames, 3.0MB → 1.1MB (2.8x), 19 JPEGs deleted\n2026-04-20T09:20:37.786800Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2023204823691812477, trigger=visual_change)\n2026-04-20T09:20:41.129431Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-6072351155973619451, trigger=visual_change)\n2026-04-20T09:20:50.199285Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5706498223560949354, trigger=visual_change)\n2026-04-20T09:21:31.117184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:21:34.972211Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:35.170979Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:36.933389Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:37.021845Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:46.669822Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:21:48.854992Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:48.978898Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:50.309855Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:52.623328Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:53.747271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:22:30.747706Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=7611972069272730175, trigger=click)\n2026-04-20T09:23:30.996636Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:23:31.087562Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:23:48.889059Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4881264313375199900, trigger=visual_change)\n2026-04-20T09:24:06.522344Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:24:36.382066Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2367845838743006480, trigger=click)\n2026-04-20T09:24:36.620082Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=2367845838743006480, trigger=click)\n2026-04-20T09:25:35.337676Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5282847726719021120, trigger=visual_change)\n2026-04-20T09:25:35.422875Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=78 elapsed=4.581551458s\n2026-04-20T09:25:35.423240Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 78 eligible frames\n2026-04-20T09:25:37.400513Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 30 frames, 10.0MB → 2.2MB (4.6x), 30 JPEGs deleted\n2026-04-20T09:25:41.702147Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 46 frames, 6.5MB → 3.7MB (1.7x), 46 JPEGs deleted\n2026-04-20T09:26:39.541417Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5282847726719021120, trigger=click)\n2026-04-20T09:26:39.637408Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5282847726719021120, trigger=visual_change)\n2026-04-20T09:26:45.621450Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=4022234323396946320, trigger=click)\n2026-04-20T09:27:27.813009Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=8695496312677497071, trigger=visual_change)\n2026-04-20T09:27:56.174253Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-1474433992317029253, trigger=visual_change)\n2026-04-20T09:28:15.443669Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:15.559560Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:22.871037Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:22.998585Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:23.546566Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:27.616599Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:27.810327Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:28.805551Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:28.996759Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:38.226782Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:41.919454Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:43.104835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:43.291487Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:29:15.175914Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:29:15.288120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:30:44.693517Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=152 elapsed=2.991660041s\n2026-04-20T09:30:44.693996Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 152 eligible frames\n2026-04-20T09:30:49.454089Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 62 frames, 18.7MB → 3.2MB (5.8x), 62 JPEGs deleted\n2026-04-20T09:31:02.747279Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 88 frames, 12.9MB → 5.8MB (2.2x), 88 JPEGs deleted\n2026-04-20T09:32:40.896832Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2273900790310487349, trigger=visual_change)\n2026-04-20T09:34:10.145625Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=416962040610113261, trigger=visual_change)\n2026-04-20T09:36:04.111438Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=113 elapsed=1.365217541s\n2026-04-20T09:36:04.111554Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 113 eligible frames\n2026-04-20T09:36:07.290184Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 50 frames, 10.6MB → 1.1MB (9.7x), 50 JPEGs deleted\n2026-04-20T09:36:11.502821Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 61 frames, 9.4MB → 4.3MB (2.2x), 61 JPEGs deleted\n2026-04-20T09:36:23.826107Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:30.222596Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:30.344862Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:37.846785Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=visual_change)\n2026-04-20T09:36:39.497294Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:39.674063Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:43.099360Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:44.941648Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:48.648990Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:48.775433Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:59.935381Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5838015891328253158, trigger=click)\n2026-04-20T09:37:00.277881Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5838015891328253158, trigger=click)\n2026-04-20T09:37:07.535793Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:08.980062Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:09.230737Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:11.150616Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:11.367104Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:38:39.474470Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=click)\n2026-04-20T09:41:06.365928Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:13.008346Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:20.360309Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=114 elapsed=8.859599166s\n2026-04-20T09:41:20.362479Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 114 eligible frames\n2026-04-20T09:41:25.798397Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 52 frames, 9.9MB → 1.1MB (9.3x), 52 JPEGs deleted\n2026-04-20T09:41:41.703983Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 60 frames, 8.1MB → 3.4MB (2.4x), 60 JPEGs deleted\n2026-04-20T09:41:48.137009Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:50.361864Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:52.640333Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:58.785028Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:41:59.196006Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:02.302969Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:02.491579Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:05.398822Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-6926500999441679870, trigger=visual_change)\n2026-04-20T09:42:15.779232Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:24.922773Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:28.057485Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:28.270226Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:30.197825Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:32.654724Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:32.947917Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:50.661919Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:54.467651Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:06.075687Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:06.266678Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:13.645777Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:14.025201Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:17.704682Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:18.003229Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:23.818550Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:25.680139Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:26.624253Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:31.133945Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:31.650235Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:45:02.727558Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=4753636820005661432, trigger=visual_change)\n2026-04-20T09:46:50.876930Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=74 elapsed=9.106527083s\n2026-04-20T09:46:50.889948Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 74 eligible frames\n2026-04-20T09:46:55.120250Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 31 frames, 5.7MB → 0.3MB (19.2x), 31 JPEGs deleted\n2026-04-20T09:47:04.755567Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 41 frames, 6.0MB → 1.8MB (3.3x), 41 JPEGs deleted\n2026-04-20T09:47:04.770274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6010633228612407689, trigger=click)\n2026-04-20T09:47:32.689458Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-148021406869771861, trigger=visual_change)\n2026-04-20T09:47:35.766223Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-148021406869771861, trigger=visual_change)\n2026-04-20T09:47:58.935503Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-6169109282047323600, trigger=click)\n2026-04-20T09:52:12.454962Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=97 elapsed=7.696011375s\n2026-04-20T09:52:12.459566Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 97 eligible frames\n2026-04-20T09:52:23.116499Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 51 frames, 11.8MB → 2.4MB (4.9x), 51 JPEGs deleted\n2026-04-20T09:52:31.720397Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 44 frames, 7.6MB → 2.8MB (2.7x), 44 JPEGs deleted\n2026-04-20T09:52:59.219405Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1610157915853078934, trigger=visual_change)\n2026-04-20T09:53:45.292944Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5708228181494047783, trigger=visual_change)\n2026-04-20T09:53:48.323898Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5708228181494047783, trigger=visual_change)\n2026-04-20T09:55:24.761072Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-348703920195698856, trigger=visual_change)\n2026-04-20T09:55:36.910247Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7110021974046204241, trigger=visual_change)\n2026-04-20T09:56:10.545531Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=2028581362717921873, trigger=visual_change)\n2026-04-20T09:56:57.973992Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=8927321389811673919, trigger=click)\n2026-04-20T09:57:46.613875Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=104 elapsed=14.8827255s\n2026-04-20T09:57:46.614109Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 104 eligible frames\n2026-04-20T09:58:03.737408Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 61 frames, 10.5MB → 4.1MB (2.6x), 61 JPEGs deleted\n2026-04-20T09:58:09.508886Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 41 frames, 6.4MB → 1.8MB (3.5x), 41 JPEGs deleted\n2026-04-20T09:58:15.283826Z WARN sqlx::query: summary=\"COMMIT\" db.statement=\"\" rows_affected=1 rows_returned=0 elapsed=1.503935584s\n2026-04-20T09:58:46.569792Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5225052939832040319, trigger=visual_change)\n2026-04-20T10:00:52.594246Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=305670337277957061, trigger=visual_change)\n2026-04-20T10:01:23.329861Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5911983749499347001, trigger=visual_change)\n2026-04-20T10:02:18.294628Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:02:21.424720Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:02:24.423729Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:03:22.734042Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=82 elapsed=13.22261775s\n2026-04-20T10:03:22.753002Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 82 eligible frames\n2026-04-20T10:03:40.239737Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 48 frames, 8.6MB → 3.5MB (2.5x), 48 JPEGs deleted\n2026-04-20T10:03:45.878728Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 32 frames, 4.8MB → 1.1MB (4.3x), 32 JPEGs deleted","depth":4,"value":"ggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\n2026-04-18T14:43:07.918412Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks\n2026-04-18T14:44:35.514401Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-8593116969702213352, trigger=visual_change)\n2026-04-18T14:44:35.891413Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-8593116969702213352, trigger=visual_change)\n2026-04-18T14:45:06.077184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=visual_change)\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\n2026-04-18T14:45:09.423485Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks\n2026-04-18T14:45:40.853509Z WARN screenpipe_db::db: ImmediateTx dropped without commit — rolling back\n2026-04-18T14:45:45.829292Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=click)\n2026-04-18T14:45:45.931557Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6130095137839516019, trigger=click)\nzsh: terminated npx screenpipe@latest record --ignored-windows \"Boosteroid\"\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sp-start\ndetected hardware tier: Mid\nwarning: parakeet is not supported on this platform, using whisper-tiny instead\n2026-04-20T09:09:02.916408Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store\nchecking permissions...\n screen recording: ok\n accessibility: ok\n2026-04-20T09:09:03.042870Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6\n2026-04-20T09:09:04.217747Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor\n2026-04-20T09:09:04.219685Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)\n2026-04-20T09:09:04.220577Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)\n2026-04-20T09:09:04.314160Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true\n2026-04-20T09:09:04.314251Z INFO screenpipe: meeting detector enabled — independent of transcription mode\n2026-04-20T09:09:04.314473Z INFO screenpipe: API server listening on 127.0.0.1:3030 (localhost only)\n2026-04-20T09:09:04.314486Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key\n2026-04-20T09:09:04.314428Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)\n2026-04-20T09:09:04.314439Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)\n2026-04-20T09:09:04.314563Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager\n2026-04-20T09:09:04.370709Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))\n2026-04-20T09:09:04.383079Z INFO screenpipe_core::pipes: loaded pipe: day-recap\n2026-04-20T09:09:04.384136Z INFO screenpipe_core::pipes: loaded pipe: standup-update\n2026-04-20T09:09:04.384745Z INFO screenpipe_core::pipes: loaded pipe: ai-habits\n2026-04-20T09:09:04.385360Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown\n2026-04-20T09:09:04.385876Z INFO screenpipe_core::pipes: loaded pipe: video-export\n2026-04-20T09:09:04.386843Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary\n2026-04-20T09:09:04.386950Z INFO screenpipe_core::pipes: loaded 6 pipes from \"/Users/lukas/.screenpipe/pipes\"\n\n\n\n _ \n __________________ ___ ____ ____ (_____ ___ \n / ___/ ___/ ___/ _ \\/ _ \\/ __ \\ / __ \\/ / __ \\/ _ \\\n (__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/\n/____/\\___/_/ \\___/\\___/_/ /_/ / .___/_/ .___/\\___/ \n /_/ /_/ \n\n\n\npower AI by everything you've seen, said or heard\nopen source | runs locally | developer friendly\n\n\n┌────────────────────────┬────────────────────────────────────┐\n│ setting │ value │\n├────────────────────────┼────────────────────────────────────┤\n│ audio chunk duration │ 30 seconds │\n│ port │ 3030 │\n│ audio disabled │ true │\n│ vision disabled │ false │\n│ pause on DRM content │ false │\n│ audio engine │ Parakeet │\n│ vad engine │ Silero │\n│ data directory │ /Users/lukas/.screenpipe │\n│ debug mode │ false │\n│ telemetry │ true │\n│ use pii removal │ true │\n│ use all monitors │ true │\n│ ignored windows │ [\"Boosteroid\"] │\n│ included windows │ [] │\n│ cloud sync │ disabled │\n│ auto-destruct pid │ 0 │\n│ deepgram key │ not set │\n│ api auth │ enabled │\n2026-04-20T09:09:04.390629Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)\n2026-04-20T09:09:04.411414Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh\n│ encrypt secrets │ disabled │\n│ retention days │ 14 │\n├────────────────────────┼────────────────────────────────────┤\n│ languages │ │\n│ │ all languages │\n├────────────────────────┼────────────────────────────────────┤\n│ monitors │ │\n│ │ id: 1 │\n│ │ id: 2 │\n├────────────────────────┼────────────────────────────────────┤\n│ audio devices │ │\n│ │ disabled │\n└────────────────────────┴────────────────────────────────────┘\nyou are using local processing. all your data stays on your computer.\n\nwarning: telemetry is enabled. only error-level data will be sent.\nto disable, use the --disable-telemetry flag.\n\ncheck latest changes here: https://github.com/screenpipe/screenpipe/releases\n2026-04-20T09:09:04.444803Z INFO screenpipe: starting UI event capture\n2026-04-20T09:09:04.482387Z INFO screenpipe_engine::ui_recorder: Starting UI event capture\n2026-04-20T09:09:04.509116Z INFO screenpipe_engine::ui_recorder: UI recording session started: f0f4e024-5bd3-4a71-bf92-6feec89df1b0\n2026-04-20T09:09:04.509231Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)\n2026-04-20T09:09:04.509296Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 06:09:04.509295 UTC to 2026-04-20 06:09:04.509295 UTC)\n2026-04-20T09:09:04.510396Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)\n2026-04-20T09:09:04.511440Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 0 frame entries, coverage from 2026-04-19 06:09:04.509295 UTC\n2026-04-20T09:09:04.519984Z INFO screenpipe_engine::server: Server listening on 127.0.0.1:3030\n2026-04-20T09:09:04.547289Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030\n2026-04-20T09:09:04.965425Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)\n2026-04-20T09:09:04.965463Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)\n2026-04-20T09:09:04.965503Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)\n2026-04-20T09:09:05.226982Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)\n2026-04-20T09:09:05.227016Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)\n2026-04-20T09:09:05.227136Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)\n2026-04-20T09:09:05.227375Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)\n2026-04-20T09:09:07.300820Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 3 excluded)\n2026-04-20T09:09:07.600290Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=51507, dur=136ms\n2026-04-20T09:09:07.683736Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 3 excluded)\n2026-04-20T09:09:07.989000Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=51508, dur=185ms\n2026-04-20T09:10:09.231456Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=197 elapsed=4.915444333s\n2026-04-20T09:10:09.231772Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 197 eligible frames\n2026-04-20T09:10:14.691743Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 83 frames, 16.3MB → 2.5MB (6.5x), 83 JPEGs deleted\n2026-04-20T09:10:15.677833Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:17.628368Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:17.716440Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:21.386286Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 100 frames, 17.9MB → 6.8MB (2.6x), 100 JPEGs deleted\n2026-04-20T09:10:22.685406Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 2.2MB → 0.9MB (2.4x), 14 JPEGs deleted\n2026-04-20T09:10:23.507120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:42.391615Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:42.475795Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:44.750359Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:44.980679Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:53.686192Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:53.738064Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:56.291827Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:56.451083Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:12:46.023481Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)\n2026-04-20T09:12:49.049656Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)\n2026-04-20T09:13:23.025287Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=596503974090759224, trigger=click)\n2026-04-20T09:13:28.322087Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3622643354436184557, trigger=visual_change)\n2026-04-20T09:14:35.739274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3577748614007823836, trigger=click)\n2026-04-20T09:15:24.523245Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=3 elapsed=1.840652833s\n2026-04-20T09:15:24.523440Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 3 eligible frames\n2026-04-20T09:15:37.319126Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7719336893453902381, trigger=click)\n2026-04-20T09:16:25.021906Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2226463793404778776, trigger=visual_change)\n2026-04-20T09:16:40.481038Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-2449473341097903678, trigger=visual_change)\n2026-04-20T09:17:01.897197Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)\n2026-04-20T09:17:11.727852Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:11.880765Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:12.886271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:12.961362Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:49.590890Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-5042772765556709364, trigger=visual_change)\n2026-04-20T09:18:38.516980Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3873630181551472921, trigger=click)\n2026-04-20T09:18:52.171689Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-3192006669261292210, trigger=click)\n2026-04-20T09:18:52.339913Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)\n2026-04-20T09:19:18.962835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)\n2026-04-20T09:19:21.997050Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)\n2026-04-20T09:20:27.568575Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=33 elapsed=3.045052625s\n2026-04-20T09:20:27.568689Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 33 eligible frames\n2026-04-20T09:20:28.589270Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 3.1MB → 0.7MB (4.6x), 14 JPEGs deleted\n2026-04-20T09:20:30.840110Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 19 frames, 3.0MB → 1.1MB (2.8x), 19 JPEGs deleted\n2026-04-20T09:20:37.786800Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2023204823691812477, trigger=visual_change)\n2026-04-20T09:20:41.129431Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-6072351155973619451, trigger=visual_change)\n2026-04-20T09:20:50.199285Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5706498223560949354, trigger=visual_change)\n2026-04-20T09:21:31.117184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:21:34.972211Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:35.170979Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:36.933389Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:37.021845Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:46.669822Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:21:48.854992Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:48.978898Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:50.309855Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:52.623328Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:53.747271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:22:30.747706Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=7611972069272730175, trigger=click)\n2026-04-20T09:23:30.996636Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:23:31.087562Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:23:48.889059Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4881264313375199900, trigger=visual_change)\n2026-04-20T09:24:06.522344Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:24:36.382066Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2367845838743006480, trigger=click)\n2026-04-20T09:24:36.620082Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=2367845838743006480, trigger=click)\n2026-04-20T09:25:35.337676Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5282847726719021120, trigger=visual_change)\n2026-04-20T09:25:35.422875Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=78 elapsed=4.581551458s\n2026-04-20T09:25:35.423240Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 78 eligible frames\n2026-04-20T09:25:37.400513Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 30 frames, 10.0MB → 2.2MB (4.6x), 30 JPEGs deleted\n2026-04-20T09:25:41.702147Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 46 frames, 6.5MB → 3.7MB (1.7x), 46 JPEGs deleted\n2026-04-20T09:26:39.541417Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5282847726719021120, trigger=click)\n2026-04-20T09:26:39.637408Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5282847726719021120, trigger=visual_change)\n2026-04-20T09:26:45.621450Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=4022234323396946320, trigger=click)\n2026-04-20T09:27:27.813009Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=8695496312677497071, trigger=visual_change)\n2026-04-20T09:27:56.174253Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-1474433992317029253, trigger=visual_change)\n2026-04-20T09:28:15.443669Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:15.559560Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:22.871037Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:22.998585Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:23.546566Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:27.616599Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:27.810327Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:28.805551Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:28.996759Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:38.226782Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:41.919454Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:43.104835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:43.291487Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:29:15.175914Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:29:15.288120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:30:44.693517Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=152 elapsed=2.991660041s\n2026-04-20T09:30:44.693996Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 152 eligible frames\n2026-04-20T09:30:49.454089Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 62 frames, 18.7MB → 3.2MB (5.8x), 62 JPEGs deleted\n2026-04-20T09:31:02.747279Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 88 frames, 12.9MB → 5.8MB (2.2x), 88 JPEGs deleted\n2026-04-20T09:32:40.896832Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2273900790310487349, trigger=visual_change)\n2026-04-20T09:34:10.145625Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=416962040610113261, trigger=visual_change)\n2026-04-20T09:36:04.111438Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=113 elapsed=1.365217541s\n2026-04-20T09:36:04.111554Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 113 eligible frames\n2026-04-20T09:36:07.290184Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 50 frames, 10.6MB → 1.1MB (9.7x), 50 JPEGs deleted\n2026-04-20T09:36:11.502821Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 61 frames, 9.4MB → 4.3MB (2.2x), 61 JPEGs deleted\n2026-04-20T09:36:23.826107Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:30.222596Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:30.344862Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:37.846785Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=visual_change)\n2026-04-20T09:36:39.497294Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:39.674063Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:43.099360Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:44.941648Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:48.648990Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:48.775433Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:59.935381Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5838015891328253158, trigger=click)\n2026-04-20T09:37:00.277881Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5838015891328253158, trigger=click)\n2026-04-20T09:37:07.535793Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:08.980062Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:09.230737Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:11.150616Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:11.367104Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:38:39.474470Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=click)\n2026-04-20T09:41:06.365928Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:13.008346Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:20.360309Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=114 elapsed=8.859599166s\n2026-04-20T09:41:20.362479Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 114 eligible frames\n2026-04-20T09:41:25.798397Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 52 frames, 9.9MB → 1.1MB (9.3x), 52 JPEGs deleted\n2026-04-20T09:41:41.703983Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 60 frames, 8.1MB → 3.4MB (2.4x), 60 JPEGs deleted\n2026-04-20T09:41:48.137009Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:50.361864Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:52.640333Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:58.785028Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:41:59.196006Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:02.302969Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:02.491579Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:05.398822Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-6926500999441679870, trigger=visual_change)\n2026-04-20T09:42:15.779232Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:24.922773Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:28.057485Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:28.270226Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:30.197825Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:32.654724Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:32.947917Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:50.661919Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:54.467651Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:06.075687Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:06.266678Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:13.645777Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:14.025201Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:17.704682Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:18.003229Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:23.818550Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:25.680139Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:26.624253Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:31.133945Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:31.650235Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:45:02.727558Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=4753636820005661432, trigger=visual_change)\n2026-04-20T09:46:50.876930Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=74 elapsed=9.106527083s\n2026-04-20T09:46:50.889948Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 74 eligible frames\n2026-04-20T09:46:55.120250Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 31 frames, 5.7MB → 0.3MB (19.2x), 31 JPEGs deleted\n2026-04-20T09:47:04.755567Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 41 frames, 6.0MB → 1.8MB (3.3x), 41 JPEGs deleted\n2026-04-20T09:47:04.770274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6010633228612407689, trigger=click)\n2026-04-20T09:47:32.689458Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-148021406869771861, trigger=visual_change)\n2026-04-20T09:47:35.766223Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-148021406869771861, trigger=visual_change)\n2026-04-20T09:47:58.935503Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-6169109282047323600, trigger=click)\n2026-04-20T09:52:12.454962Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=97 elapsed=7.696011375s\n2026-04-20T09:52:12.459566Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 97 eligible frames\n2026-04-20T09:52:23.116499Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 51 frames, 11.8MB → 2.4MB (4.9x), 51 JPEGs deleted\n2026-04-20T09:52:31.720397Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 44 frames, 7.6MB → 2.8MB (2.7x), 44 JPEGs deleted\n2026-04-20T09:52:59.219405Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1610157915853078934, trigger=visual_change)\n2026-04-20T09:53:45.292944Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5708228181494047783, trigger=visual_change)\n2026-04-20T09:53:48.323898Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5708228181494047783, trigger=visual_change)\n2026-04-20T09:55:24.761072Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-348703920195698856, trigger=visual_change)\n2026-04-20T09:55:36.910247Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7110021974046204241, trigger=visual_change)\n2026-04-20T09:56:10.545531Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=2028581362717921873, trigger=visual_change)\n2026-04-20T09:56:57.973992Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=8927321389811673919, trigger=click)\n2026-04-20T09:57:46.613875Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=104 elapsed=14.8827255s\n2026-04-20T09:57:46.614109Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 104 eligible frames\n2026-04-20T09:58:03.737408Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 61 frames, 10.5MB → 4.1MB (2.6x), 61 JPEGs deleted\n2026-04-20T09:58:09.508886Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 41 frames, 6.4MB → 1.8MB (3.5x), 41 JPEGs deleted\n2026-04-20T09:58:15.283826Z WARN sqlx::query: summary=\"COMMIT\" db.statement=\"\" rows_affected=1 rows_returned=0 elapsed=1.503935584s\n2026-04-20T09:58:46.569792Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5225052939832040319, trigger=visual_change)\n2026-04-20T10:00:52.594246Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=305670337277957061, trigger=visual_change)\n2026-04-20T10:01:23.329861Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5911983749499347001, trigger=visual_change)\n2026-04-20T10:02:18.294628Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:02:21.424720Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:02:24.423729Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:03:22.734042Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=82 elapsed=13.22261775s\n2026-04-20T10:03:22.753002Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 82 eligible frames\n2026-04-20T10:03:40.239737Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 48 frames, 8.6MB → 3.5MB (2.5x), 48 JPEGs deleted\n2026-04-20T10:03:45.878728Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 32 frames, 4.8MB → 1.1MB (4.3x), 32 JPEGs deleted","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.27027926,"top":1.0,"width":0.06732048,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.27227393,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.33759972,"top":1.0,"width":0.06732048,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.33959442,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.40492022,"top":1.0,"width":0.06732048,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.4069149,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.4722407,"top":1.0,"width":0.06732048,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.4742354,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.53956115,"top":1.0,"width":0.06715426,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.5415558,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"✳ Review screenpipe usage and Boosteroid integration (claude)","depth":2,"bounds":{"left":0.60671544,"top":1.0,"width":0.06715426,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.6087101,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.67386967,"top":1.0,"width":0.06715426,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.67586434,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.7273936,"top":1.0,"width":0.01861702,"height":-0.023144484},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"screenpipe\"","depth":1,"bounds":{"left":0.4956782,"top":1.0,"width":0.027925532,"height":-0.02394259},"role_description":"text"}]...
|
-7539762675761127371
|
-1969116165549787679
|
click
|
accessibility
|
NULL
|
ggml_metal_init: use fusion = true
ggml_me ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
2026-04-18T14:43:07.918412Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks
2026-04-18T14:44:35.514401Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-8593116969702213352, trigger=visual_change)
2026-04-18T14:44:35.891413Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-8593116969702213352, trigger=visual_change)
2026-04-18T14:45:06.077184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=visual_change)
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
2026-04-18T14:45:09.423485Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks
2026-04-18T14:45:40.853509Z WARN screenpipe_db::db: ImmediateTx dropped without commit — rolling back
2026-04-18T14:45:45.829292Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=click)
2026-04-18T14:45:45.931557Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6130095137839516019, trigger=click)
zsh: terminated npx screenpipe@latest record --ignored-windows "Boosteroid"
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sp-start
detected hardware tier: Mid
warning: parakeet is not supported on this platform, using whisper-tiny instead
2026-04-20T09:09:02.916408Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store
checking permissions...
screen recording: ok
accessibility: ok
2026-04-20T09:09:03.042870Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6
2026-04-20T09:09:04.217747Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor
2026-04-20T09:09:04.219685Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)
2026-04-20T09:09:04.220577Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)
2026-04-20T09:09:04.314160Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true
2026-04-20T09:09:04.314251Z INFO screenpipe: meeting detector enabled — independent of transcription mode
2026-04-20T09:09:04.314473Z INFO screenpipe: API server listening on [IP_ADDRESS]:3030 (localhost only)
2026-04-20T09:09:04.314486Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key
2026-04-20T09:09:04.314428Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)
2026-04-20T09:09:04.314439Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)
2026-04-20T09:09:04.314563Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager
2026-04-20T09:09:04.370709Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))
2026-04-20T09:09:04.383079Z INFO screenpipe_core::pipes: loaded pipe: day-recap
2026-04-20T09:09:04.384136Z INFO screenpipe_core::pipes: loaded pipe: standup-update
2026-04-20T09:09:04.384745Z INFO screenpipe_core::pipes: loaded pipe: ai-habits
2026-04-20T09:09:04.385360Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown
2026-04-20T09:09:04.385876Z INFO screenpipe_core::pipes: loaded pipe: video-export
2026-04-20T09:09:04.386843Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary
2026-04-20T09:09:04.386950Z INFO screenpipe_core::pipes: loaded 6 pipes from "/Users/lukas/.screenpipe/pipes"
_
__________________ ___ ____ ____ (_____ ___
/ ___/ ___/ ___/ _ \/ _ \/ __ \ / __ \/ / __ \/ _ \
(__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/
/____/\___/_/ \___/\___/_/ /_/ / .___/_/ .___/\___/
/_/ /_/
power AI by everything you've seen, said or heard
open source | runs locally | developer friendly
┌────────────────────────┬────────────────────────────────────┐
│ setting │ value │
├────────────────────────┼────────────────────────────────────┤
│ audio chunk duration │ 30 seconds │
│ port │ 3030 │
│ audio disabled │ true │
│ vision disabled │ false │
│ pause on DRM content │ false │
│ audio engine │ Parakeet │
│ vad engine │ Silero │
│ data directory │ /Users/lukas/.screenpipe │
│ debug mode │ false │
│ telemetry │ true │
│ use pii removal │ true │
│ use all monitors │ true │
│ ignored windows │ ["Boosteroid"] │
│ included windows │ [] │
│ cloud sync │ disabled │
│ auto-destruct pid │ 0 │
│ deepgram key │ not set │
│ api auth │ enabled │
2026-04-20T09:09:04.390629Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)
2026-04-20T09:09:04.411414Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh
│ encrypt secrets │ disabled │
│ retention days │ 14 │
├────────────────────────┼────────────────────────────────────┤
│ languages │ │
│ │ all languages │
├────────────────────────┼────────────────────────────────────┤
│ monitors │ │
│ │ id: 1 │
│ │ id: 2 │
├────────────────────────┼────────────────────────────────────┤
│ audio devices │ │
│ │ disabled │
└────────────────────────┴────────────────────────────────────┘
you are using local processing. all your data stays on your computer.
warning: telemetry is enabled. only error-level data will be sent.
to disable, use the --disable-telemetry flag.
check latest changes here: https://github.com/screenpipe/screenpipe/releases
2026-04-20T09:09:04.444803Z INFO screenpipe: starting UI event capture
2026-04-20T09:09:04.482387Z INFO screenpipe_engine::ui_recorder: Starting UI event capture
2026-04-20T09:09:04.509116Z INFO screenpipe_engine::ui_recorder: UI recording session started: f0f4e024-5bd3-4a71-bf92-6feec89df1b0
2026-04-20T09:09:04.509231Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)
2026-04-20T09:09:04.509296Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 06:09:04.509295 UTC to 2026-04-20 06:09:04.509295 UTC)
2026-04-20T09:09:04.510396Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)
2026-04-20T09:09:04.511440Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 0 frame entries, coverage from 2026-04-19 06:09:04.509295 UTC
2026-04-20T09:09:04.519984Z INFO screenpipe_engine::server: Server listening on [IP_ADDRESS]:3030
2026-04-20T09:09:04.547289Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030
2026-04-20T09:09:04.965425Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)
2026-04-20T09:09:04.965463Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)
2026-04-20T09:09:04.965503Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)
2026-04-20T09:09:05.226982Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)
2026-04-20T09:09:05.227016Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)
2026-04-20T09:09:05.227136Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)
2026-04-20T09:09:05.227375Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)
2026-04-20T09:09:07.300820Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 3 excluded)
2026-04-20T09:09:07.600290Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=51507, dur=136ms
2026-04-20T09:09:07.683736Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 3 excluded)
2026-04-20T09:09:07.989000Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=51508, dur=185ms
2026-04-20T09:10:09.231456Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=197 elapsed=4.915444333s
2026-04-20T09:10:09.231772Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 197 eligible frames
2026-04-20T09:10:14.691743Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 83 frames, 16.3MB → 2.5MB (6.5x), 83 JPEGs deleted
2026-04-20T09:10:15.677833Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:17.628368Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:17.716440Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:21.386286Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 100 frames, 17.9MB → 6.8MB (2.6x), 100 JPEGs deleted
2026-04-20T09:10:22.685406Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 2.2MB → 0.9MB (2.4x), 14 JPEGs deleted
2026-04-20T09:10:23.507120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:42.391615Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:42.475795Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:44.750359Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:44.980679Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:53.686192Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:53.738064Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:56.291827Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:56.451083Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:12:46.023481Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)
2026-04-20T09:12:49.049656Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)
2026-04-20T09:13:23.025287Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=596503974090759224, trigger=click)
2026-04-20T09:13:28.322087Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3622643354436184557, trigger=visual_change)
2026-04-20T09:14:35.739274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3577748614007823836, trigger=click)
2026-04-20T09:15:24.523245Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=3 elapsed=1.840652833s
2026-04-20T09:15:24.523440Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 3 eligible frames
2026-04-20T09:15:37.319126Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7719336893453902381, trigger=click)
2026-04-20T09:16:25.021906Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2226463793404778776, trigger=visual_change)
2026-04-20T09:16:40.481038Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-2449473341097903678, trigger=visual_change)
2026-04-20T09:17:01.897197Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)
2026-04-20T09:17:11.727852Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:11.880765Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:12.886271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:12.961362Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:49.590890Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-5042772765556709364, trigger=visual_change)
2026-04-20T09:18:38.516980Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3873630181551472921, trigger=click)
2026-04-20T09:18:52.171689Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-3192006669261292210, trigger=click)
2026-04-20T09:18:52.339913Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)
2026-04-20T09:19:18.962835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)
2026-04-20T09:19:21.997050Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)
2026-04-20T09:20:27.568575Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=33 elapsed=3.045052625s
2026-04-20T09:20:27.568689Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 33 eligible frames
2026-04-20T09:20:28.589270Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 3.1MB → 0.7MB (4.6x), 14 JPEGs deleted
2026-04-20T09:20:30.840110Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 19 frames, 3.0MB → 1.1MB (2.8x), 19 JPEGs deleted
2026-04-20T09:20:37.786800Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (...
|
NULL
|
|
52480
|
NULL
|
0
|
2026-04-20T07:05:17.973307+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776668717973_m1.jpg...
|
iTerm2
|
screenpipe"
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
ggml_metal_init: use fusion = true
ggml_me ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
2026-04-18T14:43:07.918412Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks
2026-04-18T14:44:35.514401Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-8593116969702213352, trigger=visual_change)
2026-04-18T14:44:35.891413Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-8593116969702213352, trigger=visual_change)
2026-04-18T14:45:06.077184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=visual_change)
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
2026-04-18T14:45:09.423485Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks
2026-04-18T14:45:40.853509Z WARN screenpipe_db::db: ImmediateTx dropped without commit — rolling back
2026-04-18T14:45:45.829292Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=click)
2026-04-18T14:45:45.931557Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6130095137839516019, trigger=click)
zsh: terminated npx screenpipe@latest record --ignored-windows "Boosteroid"
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sp-start
detected hardware tier: Mid
warning: parakeet is not supported on this platform, using whisper-tiny instead
2026-04-20T09:09:02.916408Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store
checking permissions...
screen recording: ok
accessibility: ok
2026-04-20T09:09:03.042870Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6
2026-04-20T09:09:04.217747Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor
2026-04-20T09:09:04.219685Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)
2026-04-20T09:09:04.220577Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)
2026-04-20T09:09:04.314160Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true
2026-04-20T09:09:04.314251Z INFO screenpipe: meeting detector enabled — independent of transcription mode
2026-04-20T09:09:04.314473Z INFO screenpipe: API server listening on [IP_ADDRESS]:3030 (localhost only)
2026-04-20T09:09:04.314486Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key
2026-04-20T09:09:04.314428Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)
2026-04-20T09:09:04.314439Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)
2026-04-20T09:09:04.314563Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager
2026-04-20T09:09:04.370709Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))
2026-04-20T09:09:04.383079Z INFO screenpipe_core::pipes: loaded pipe: day-recap
2026-04-20T09:09:04.384136Z INFO screenpipe_core::pipes: loaded pipe: standup-update
2026-04-20T09:09:04.384745Z INFO screenpipe_core::pipes: loaded pipe: ai-habits
2026-04-20T09:09:04.385360Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown
2026-04-20T09:09:04.385876Z INFO screenpipe_core::pipes: loaded pipe: video-export
2026-04-20T09:09:04.386843Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary
2026-04-20T09:09:04.386950Z INFO screenpipe_core::pipes: loaded 6 pipes from "/Users/lukas/.screenpipe/pipes"
_
__________________ ___ ____ ____ (_____ ___
/ ___/ ___/ ___/ _ \/ _ \/ __ \ / __ \/ / __ \/ _ \
(__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/
/____/\___/_/ \___/\___/_/ /_/ / .___/_/ .___/\___/
/_/ /_/
power AI by everything you've seen, said or heard
open source | runs locally | developer friendly
┌────────────────────────┬────────────────────────────────────┐
│ setting │ value │
├────────────────────────┼────────────────────────────────────┤
│ audio chunk duration │ 30 seconds │
│ port │ 3030 │
│ audio disabled │ true │
│ vision disabled │ false │
│ pause on DRM content │ false │
│ audio engine │ Parakeet │
│ vad engine │ Silero │
│ data directory │ /Users/lukas/.screenpipe │
│ debug mode │ false │
│ telemetry │ true │
│ use pii removal │ true │
│ use all monitors │ true │
│ ignored windows │ ["Boosteroid"] │
│ included windows │ [] │
│ cloud sync │ disabled │
│ auto-destruct pid │ 0 │
│ deepgram key │ not set │
│ api auth │ enabled │
2026-04-20T09:09:04.390629Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)
2026-04-20T09:09:04.411414Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh
│ encrypt secrets │ disabled │
│ retention days │ 14 │
├────────────────────────┼────────────────────────────────────┤
│ languages │ │
│ │ all languages │
├────────────────────────┼────────────────────────────────────┤
│ monitors │ │
│ │ id: 1 │
│ │ id: 2 │
├────────────────────────┼────────────────────────────────────┤
│ audio devices │ │
│ │ disabled │
└────────────────────────┴────────────────────────────────────┘
you are using local processing. all your data stays on your computer.
warning: telemetry is enabled. only error-level data will be sent.
to disable, use the --disable-telemetry flag.
check latest changes here: https://github.com/screenpipe/screenpipe/releases
2026-04-20T09:09:04.444803Z INFO screenpipe: starting UI event capture
2026-04-20T09:09:04.482387Z INFO screenpipe_engine::ui_recorder: Starting UI event capture
2026-04-20T09:09:04.509116Z INFO screenpipe_engine::ui_recorder: UI recording session started: f0f4e024-5bd3-4a71-bf92-6feec89df1b0
2026-04-20T09:09:04.509231Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)
2026-04-20T09:09:04.509296Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 06:09:04.509295 UTC to 2026-04-20 06:09:04.509295 UTC)
2026-04-20T09:09:04.510396Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)
2026-04-20T09:09:04.511440Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 0 frame entries, coverage from 2026-04-19 06:09:04.509295 UTC
2026-04-20T09:09:04.519984Z INFO screenpipe_engine::server: Server listening on [IP_ADDRESS]:3030
2026-04-20T09:09:04.547289Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030
2026-04-20T09:09:04.965425Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)
2026-04-20T09:09:04.965463Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)
2026-04-20T09:09:04.965503Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)
2026-04-20T09:09:05.226982Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)
2026-04-20T09:09:05.227016Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)
2026-04-20T09:09:05.227136Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)
2026-04-20T09:09:05.227375Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)
2026-04-20T09:09:07.300820Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 3 excluded)
2026-04-20T09:09:07.600290Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=51507, dur=136ms
2026-04-20T09:09:07.683736Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 3 excluded)
2026-04-20T09:09:07.989000Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=51508, dur=185ms
2026-04-20T09:10:09.231456Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=197 elapsed=4.915444333s
2026-04-20T09:10:09.231772Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 197 eligible frames
2026-04-20T09:10:14.691743Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 83 frames, 16.3MB → 2.5MB (6.5x), 83 JPEGs deleted
2026-04-20T09:10:15.677833Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:17.628368Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:17.716440Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:21.386286Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 100 frames, 17.9MB → 6.8MB (2.6x), 100 JPEGs deleted
2026-04-20T09:10:22.685406Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 2.2MB → 0.9MB (2.4x), 14 JPEGs deleted
2026-04-20T09:10:23.507120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:42.391615Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:42.475795Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:44.750359Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:44.980679Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:53.686192Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:53.738064Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:56.291827Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:56.451083Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:12:46.023481Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)
2026-04-20T09:12:49.049656Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)
2026-04-20T09:13:23.025287Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=596503974090759224, trigger=click)
2026-04-20T09:13:28.322087Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3622643354436184557, trigger=visual_change)
2026-04-20T09:14:35.739274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3577748614007823836, trigger=click)
2026-04-20T09:15:24.523245Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=3 elapsed=1.840652833s
2026-04-20T09:15:24.523440Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 3 eligible frames
2026-04-20T09:15:37.319126Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7719336893453902381, trigger=click)
2026-04-20T09:16:25.021906Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2226463793404778776, trigger=visual_change)
2026-04-20T09:16:40.481038Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-2449473341097903678, trigger=visual_change)
2026-04-20T09:17:01.897197Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)
2026-04-20T09:17:11.727852Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:11.880765Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:12.886271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:12.961362Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:49.590890Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-5042772765556709364, trigger=visual_change)
2026-04-20T09:18:38.516980Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3873630181551472921, trigger=click)
2026-04-20T09:18:52.171689Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-3192006669261292210, trigger=click)
2026-04-20T09:18:52.339913Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)
2026-04-20T09:19:18.962835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)
2026-04-20T09:19:21.997050Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)
2026-04-20T09:20:27.568575Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=33 elapsed=3.045052625s
2026-04-20T09:20:27.568689Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 33 eligible frames
2026-04-20T09:20:28.589270Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 3.1MB → 0.7MB (4.6x), 14 JPEGs deleted
2026-04-20T09:20:30.840110Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 19 frames, 3.0MB → 1.1MB (2.8x), 19 JPEGs deleted
2026-04-20T09:20:37.786800Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"ggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\n2026-04-18T14:43:07.918412Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks\n2026-04-18T14:44:35.514401Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-8593116969702213352, trigger=visual_change)\n2026-04-18T14:44:35.891413Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-8593116969702213352, trigger=visual_change)\n2026-04-18T14:45:06.077184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=visual_change)\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\n2026-04-18T14:45:09.423485Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks\n2026-04-18T14:45:40.853509Z WARN screenpipe_db::db: ImmediateTx dropped without commit — rolling back\n2026-04-18T14:45:45.829292Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=click)\n2026-04-18T14:45:45.931557Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6130095137839516019, trigger=click)\nzsh: terminated npx screenpipe@latest record --ignored-windows \"Boosteroid\"\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sp-start\ndetected hardware tier: Mid\nwarning: parakeet is not supported on this platform, using whisper-tiny instead\n2026-04-20T09:09:02.916408Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store\nchecking permissions...\n screen recording: ok\n accessibility: ok\n2026-04-20T09:09:03.042870Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6\n2026-04-20T09:09:04.217747Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor\n2026-04-20T09:09:04.219685Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)\n2026-04-20T09:09:04.220577Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)\n2026-04-20T09:09:04.314160Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true\n2026-04-20T09:09:04.314251Z INFO screenpipe: meeting detector enabled — independent of transcription mode\n2026-04-20T09:09:04.314473Z INFO screenpipe: API server listening on 127.0.0.1:3030 (localhost only)\n2026-04-20T09:09:04.314486Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key\n2026-04-20T09:09:04.314428Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)\n2026-04-20T09:09:04.314439Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)\n2026-04-20T09:09:04.314563Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager\n2026-04-20T09:09:04.370709Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))\n2026-04-20T09:09:04.383079Z INFO screenpipe_core::pipes: loaded pipe: day-recap\n2026-04-20T09:09:04.384136Z INFO screenpipe_core::pipes: loaded pipe: standup-update\n2026-04-20T09:09:04.384745Z INFO screenpipe_core::pipes: loaded pipe: ai-habits\n2026-04-20T09:09:04.385360Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown\n2026-04-20T09:09:04.385876Z INFO screenpipe_core::pipes: loaded pipe: video-export\n2026-04-20T09:09:04.386843Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary\n2026-04-20T09:09:04.386950Z INFO screenpipe_core::pipes: loaded 6 pipes from \"/Users/lukas/.screenpipe/pipes\"\n\n\n\n _ \n __________________ ___ ____ ____ (_____ ___ \n / ___/ ___/ ___/ _ \\/ _ \\/ __ \\ / __ \\/ / __ \\/ _ \\\n (__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/\n/____/\\___/_/ \\___/\\___/_/ /_/ / .___/_/ .___/\\___/ \n /_/ /_/ \n\n\n\npower AI by everything you've seen, said or heard\nopen source | runs locally | developer friendly\n\n\n┌────────────────────────┬────────────────────────────────────┐\n│ setting │ value │\n├────────────────────────┼────────────────────────────────────┤\n│ audio chunk duration │ 30 seconds │\n│ port │ 3030 │\n│ audio disabled │ true │\n│ vision disabled │ false │\n│ pause on DRM content │ false │\n│ audio engine │ Parakeet │\n│ vad engine │ Silero │\n│ data directory │ /Users/lukas/.screenpipe │\n│ debug mode │ false │\n│ telemetry │ true │\n│ use pii removal │ true │\n│ use all monitors │ true │\n│ ignored windows │ [\"Boosteroid\"] │\n│ included windows │ [] │\n│ cloud sync │ disabled │\n│ auto-destruct pid │ 0 │\n│ deepgram key │ not set │\n│ api auth │ enabled │\n2026-04-20T09:09:04.390629Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)\n2026-04-20T09:09:04.411414Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh\n│ encrypt secrets │ disabled │\n│ retention days │ 14 │\n├────────────────────────┼────────────────────────────────────┤\n│ languages │ │\n│ │ all languages │\n├────────────────────────┼────────────────────────────────────┤\n│ monitors │ │\n│ │ id: 1 │\n│ │ id: 2 │\n├────────────────────────┼────────────────────────────────────┤\n│ audio devices │ │\n│ │ disabled │\n└────────────────────────┴────────────────────────────────────┘\nyou are using local processing. all your data stays on your computer.\n\nwarning: telemetry is enabled. only error-level data will be sent.\nto disable, use the --disable-telemetry flag.\n\ncheck latest changes here: https://github.com/screenpipe/screenpipe/releases\n2026-04-20T09:09:04.444803Z INFO screenpipe: starting UI event capture\n2026-04-20T09:09:04.482387Z INFO screenpipe_engine::ui_recorder: Starting UI event capture\n2026-04-20T09:09:04.509116Z INFO screenpipe_engine::ui_recorder: UI recording session started: f0f4e024-5bd3-4a71-bf92-6feec89df1b0\n2026-04-20T09:09:04.509231Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)\n2026-04-20T09:09:04.509296Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 06:09:04.509295 UTC to 2026-04-20 06:09:04.509295 UTC)\n2026-04-20T09:09:04.510396Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)\n2026-04-20T09:09:04.511440Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 0 frame entries, coverage from 2026-04-19 06:09:04.509295 UTC\n2026-04-20T09:09:04.519984Z INFO screenpipe_engine::server: Server listening on 127.0.0.1:3030\n2026-04-20T09:09:04.547289Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030\n2026-04-20T09:09:04.965425Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)\n2026-04-20T09:09:04.965463Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)\n2026-04-20T09:09:04.965503Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)\n2026-04-20T09:09:05.226982Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)\n2026-04-20T09:09:05.227016Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)\n2026-04-20T09:09:05.227136Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)\n2026-04-20T09:09:05.227375Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)\n2026-04-20T09:09:07.300820Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 3 excluded)\n2026-04-20T09:09:07.600290Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=51507, dur=136ms\n2026-04-20T09:09:07.683736Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 3 excluded)\n2026-04-20T09:09:07.989000Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=51508, dur=185ms\n2026-04-20T09:10:09.231456Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=197 elapsed=4.915444333s\n2026-04-20T09:10:09.231772Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 197 eligible frames\n2026-04-20T09:10:14.691743Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 83 frames, 16.3MB → 2.5MB (6.5x), 83 JPEGs deleted\n2026-04-20T09:10:15.677833Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:17.628368Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:17.716440Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:21.386286Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 100 frames, 17.9MB → 6.8MB (2.6x), 100 JPEGs deleted\n2026-04-20T09:10:22.685406Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 2.2MB → 0.9MB (2.4x), 14 JPEGs deleted\n2026-04-20T09:10:23.507120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:42.391615Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:42.475795Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:44.750359Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:44.980679Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:53.686192Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:53.738064Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:56.291827Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:56.451083Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:12:46.023481Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)\n2026-04-20T09:12:49.049656Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)\n2026-04-20T09:13:23.025287Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=596503974090759224, trigger=click)\n2026-04-20T09:13:28.322087Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3622643354436184557, trigger=visual_change)\n2026-04-20T09:14:35.739274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3577748614007823836, trigger=click)\n2026-04-20T09:15:24.523245Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=3 elapsed=1.840652833s\n2026-04-20T09:15:24.523440Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 3 eligible frames\n2026-04-20T09:15:37.319126Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7719336893453902381, trigger=click)\n2026-04-20T09:16:25.021906Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2226463793404778776, trigger=visual_change)\n2026-04-20T09:16:40.481038Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-2449473341097903678, trigger=visual_change)\n2026-04-20T09:17:01.897197Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)\n2026-04-20T09:17:11.727852Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:11.880765Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:12.886271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:12.961362Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:49.590890Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-5042772765556709364, trigger=visual_change)\n2026-04-20T09:18:38.516980Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3873630181551472921, trigger=click)\n2026-04-20T09:18:52.171689Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-3192006669261292210, trigger=click)\n2026-04-20T09:18:52.339913Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)\n2026-04-20T09:19:18.962835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)\n2026-04-20T09:19:21.997050Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)\n2026-04-20T09:20:27.568575Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=33 elapsed=3.045052625s\n2026-04-20T09:20:27.568689Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 33 eligible frames\n2026-04-20T09:20:28.589270Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 3.1MB → 0.7MB (4.6x), 14 JPEGs deleted\n2026-04-20T09:20:30.840110Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 19 frames, 3.0MB → 1.1MB (2.8x), 19 JPEGs deleted\n2026-04-20T09:20:37.786800Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2023204823691812477, trigger=visual_change)\n2026-04-20T09:20:41.129431Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-6072351155973619451, trigger=visual_change)\n2026-04-20T09:20:50.199285Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5706498223560949354, trigger=visual_change)\n2026-04-20T09:21:31.117184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:21:34.972211Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:35.170979Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:36.933389Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:37.021845Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:46.669822Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:21:48.854992Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:48.978898Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:50.309855Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:52.623328Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:53.747271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:22:30.747706Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=7611972069272730175, trigger=click)\n2026-04-20T09:23:30.996636Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:23:31.087562Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:23:48.889059Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4881264313375199900, trigger=visual_change)\n2026-04-20T09:24:06.522344Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:24:36.382066Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2367845838743006480, trigger=click)\n2026-04-20T09:24:36.620082Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=2367845838743006480, trigger=click)\n2026-04-20T09:25:35.337676Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5282847726719021120, trigger=visual_change)\n2026-04-20T09:25:35.422875Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=78 elapsed=4.581551458s\n2026-04-20T09:25:35.423240Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 78 eligible frames\n2026-04-20T09:25:37.400513Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 30 frames, 10.0MB → 2.2MB (4.6x), 30 JPEGs deleted\n2026-04-20T09:25:41.702147Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 46 frames, 6.5MB → 3.7MB (1.7x), 46 JPEGs deleted\n2026-04-20T09:26:39.541417Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5282847726719021120, trigger=click)\n2026-04-20T09:26:39.637408Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5282847726719021120, trigger=visual_change)\n2026-04-20T09:26:45.621450Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=4022234323396946320, trigger=click)\n2026-04-20T09:27:27.813009Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=8695496312677497071, trigger=visual_change)\n2026-04-20T09:27:56.174253Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-1474433992317029253, trigger=visual_change)\n2026-04-20T09:28:15.443669Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:15.559560Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:22.871037Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:22.998585Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:23.546566Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:27.616599Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:27.810327Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:28.805551Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:28.996759Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:38.226782Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:41.919454Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:43.104835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:43.291487Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:29:15.175914Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:29:15.288120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:30:44.693517Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=152 elapsed=2.991660041s\n2026-04-20T09:30:44.693996Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 152 eligible frames\n2026-04-20T09:30:49.454089Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 62 frames, 18.7MB → 3.2MB (5.8x), 62 JPEGs deleted\n2026-04-20T09:31:02.747279Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 88 frames, 12.9MB → 5.8MB (2.2x), 88 JPEGs deleted\n2026-04-20T09:32:40.896832Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2273900790310487349, trigger=visual_change)\n2026-04-20T09:34:10.145625Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=416962040610113261, trigger=visual_change)\n2026-04-20T09:36:04.111438Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=113 elapsed=1.365217541s\n2026-04-20T09:36:04.111554Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 113 eligible frames\n2026-04-20T09:36:07.290184Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 50 frames, 10.6MB → 1.1MB (9.7x), 50 JPEGs deleted\n2026-04-20T09:36:11.502821Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 61 frames, 9.4MB → 4.3MB (2.2x), 61 JPEGs deleted\n2026-04-20T09:36:23.826107Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:30.222596Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:30.344862Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:37.846785Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=visual_change)\n2026-04-20T09:36:39.497294Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:39.674063Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:43.099360Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:44.941648Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:48.648990Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:48.775433Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:59.935381Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5838015891328253158, trigger=click)\n2026-04-20T09:37:00.277881Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5838015891328253158, trigger=click)\n2026-04-20T09:37:07.535793Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:08.980062Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:09.230737Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:11.150616Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:11.367104Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:38:39.474470Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=click)\n2026-04-20T09:41:06.365928Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:13.008346Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:20.360309Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=114 elapsed=8.859599166s\n2026-04-20T09:41:20.362479Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 114 eligible frames\n2026-04-20T09:41:25.798397Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 52 frames, 9.9MB → 1.1MB (9.3x), 52 JPEGs deleted\n2026-04-20T09:41:41.703983Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 60 frames, 8.1MB → 3.4MB (2.4x), 60 JPEGs deleted\n2026-04-20T09:41:48.137009Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:50.361864Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:52.640333Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:58.785028Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:41:59.196006Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:02.302969Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:02.491579Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:05.398822Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-6926500999441679870, trigger=visual_change)\n2026-04-20T09:42:15.779232Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:24.922773Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:28.057485Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:28.270226Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:30.197825Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:32.654724Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:32.947917Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:50.661919Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:54.467651Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:06.075687Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:06.266678Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:13.645777Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:14.025201Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:17.704682Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:18.003229Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:23.818550Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:25.680139Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:26.624253Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:31.133945Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:31.650235Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:45:02.727558Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=4753636820005661432, trigger=visual_change)\n2026-04-20T09:46:50.876930Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=74 elapsed=9.106527083s\n2026-04-20T09:46:50.889948Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 74 eligible frames\n2026-04-20T09:46:55.120250Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 31 frames, 5.7MB → 0.3MB (19.2x), 31 JPEGs deleted\n2026-04-20T09:47:04.755567Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 41 frames, 6.0MB → 1.8MB (3.3x), 41 JPEGs deleted\n2026-04-20T09:47:04.770274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6010633228612407689, trigger=click)\n2026-04-20T09:47:32.689458Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-148021406869771861, trigger=visual_change)\n2026-04-20T09:47:35.766223Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-148021406869771861, trigger=visual_change)\n2026-04-20T09:47:58.935503Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-6169109282047323600, trigger=click)\n2026-04-20T09:52:12.454962Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=97 elapsed=7.696011375s\n2026-04-20T09:52:12.459566Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 97 eligible frames\n2026-04-20T09:52:23.116499Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 51 frames, 11.8MB → 2.4MB (4.9x), 51 JPEGs deleted\n2026-04-20T09:52:31.720397Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 44 frames, 7.6MB → 2.8MB (2.7x), 44 JPEGs deleted\n2026-04-20T09:52:59.219405Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1610157915853078934, trigger=visual_change)\n2026-04-20T09:53:45.292944Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5708228181494047783, trigger=visual_change)\n2026-04-20T09:53:48.323898Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5708228181494047783, trigger=visual_change)\n2026-04-20T09:55:24.761072Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-348703920195698856, trigger=visual_change)\n2026-04-20T09:55:36.910247Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7110021974046204241, trigger=visual_change)\n2026-04-20T09:56:10.545531Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=2028581362717921873, trigger=visual_change)\n2026-04-20T09:56:57.973992Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=8927321389811673919, trigger=click)\n2026-04-20T09:57:46.613875Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=104 elapsed=14.8827255s\n2026-04-20T09:57:46.614109Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 104 eligible frames\n2026-04-20T09:58:03.737408Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 61 frames, 10.5MB → 4.1MB (2.6x), 61 JPEGs deleted\n2026-04-20T09:58:09.508886Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 41 frames, 6.4MB → 1.8MB (3.5x), 41 JPEGs deleted\n2026-04-20T09:58:15.283826Z WARN sqlx::query: summary=\"COMMIT\" db.statement=\"\" rows_affected=1 rows_returned=0 elapsed=1.503935584s\n2026-04-20T09:58:46.569792Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5225052939832040319, trigger=visual_change)\n2026-04-20T10:00:52.594246Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=305670337277957061, trigger=visual_change)\n2026-04-20T10:01:23.329861Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5911983749499347001, trigger=visual_change)\n2026-04-20T10:02:18.294628Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:02:21.424720Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:02:24.423729Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:03:22.734042Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=82 elapsed=13.22261775s\n2026-04-20T10:03:22.753002Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 82 eligible frames\n2026-04-20T10:03:40.239737Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 48 frames, 8.6MB → 3.5MB (2.5x), 48 JPEGs deleted\n2026-04-20T10:03:45.878728Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 32 frames, 4.8MB → 1.1MB (4.3x), 32 JPEGs deleted","depth":4,"value":"ggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\n2026-04-18T14:43:07.918412Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks\n2026-04-18T14:44:35.514401Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-8593116969702213352, trigger=visual_change)\n2026-04-18T14:44:35.891413Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-8593116969702213352, trigger=visual_change)\n2026-04-18T14:45:06.077184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=visual_change)\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\nwhisper_backend_init_gpu: device 0: Metal (type: 1)\nwhisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)\nwhisper_backend_init_gpu: using Metal backend\nggml_metal_init: allocating\nggml_metal_init: found device: Apple M1\nggml_metal_init: picking default device: Apple M1\nggml_metal_init: use fusion = true\nggml_metal_init: use concurrency = true\nggml_metal_init: use graph optimize = true\nwhisper_backend_init: using BLAS backend\nwhisper_init_state: kv self size = 3.15 MB\nwhisper_init_state: kv cross size = 9.44 MB\nwhisper_init_state: kv pad size = 2.36 MB\nwhisper_init_state: compute buffer (conv) = 14.17 MB\nwhisper_init_state: compute buffer (encode) = 65.96 MB\nwhisper_init_state: compute buffer (cross) = 8.50 MB\nwhisper_init_state: compute buffer (decode) = 96.83 MB\nggml_metal_free: deallocating\n2026-04-18T14:45:09.423485Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks\n2026-04-18T14:45:40.853509Z WARN screenpipe_db::db: ImmediateTx dropped without commit — rolling back\n2026-04-18T14:45:45.829292Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=click)\n2026-04-18T14:45:45.931557Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6130095137839516019, trigger=click)\nzsh: terminated npx screenpipe@latest record --ignored-windows \"Boosteroid\"\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sp-start\ndetected hardware tier: Mid\nwarning: parakeet is not supported on this platform, using whisper-tiny instead\n2026-04-20T09:09:02.916408Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store\nchecking permissions...\n screen recording: ok\n accessibility: ok\n2026-04-20T09:09:03.042870Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6\n2026-04-20T09:09:04.217747Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor\n2026-04-20T09:09:04.219685Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)\n2026-04-20T09:09:04.220577Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)\n2026-04-20T09:09:04.314160Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true\n2026-04-20T09:09:04.314251Z INFO screenpipe: meeting detector enabled — independent of transcription mode\n2026-04-20T09:09:04.314473Z INFO screenpipe: API server listening on 127.0.0.1:3030 (localhost only)\n2026-04-20T09:09:04.314486Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key\n2026-04-20T09:09:04.314428Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)\n2026-04-20T09:09:04.314439Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)\n2026-04-20T09:09:04.314563Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager\n2026-04-20T09:09:04.370709Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))\n2026-04-20T09:09:04.383079Z INFO screenpipe_core::pipes: loaded pipe: day-recap\n2026-04-20T09:09:04.384136Z INFO screenpipe_core::pipes: loaded pipe: standup-update\n2026-04-20T09:09:04.384745Z INFO screenpipe_core::pipes: loaded pipe: ai-habits\n2026-04-20T09:09:04.385360Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown\n2026-04-20T09:09:04.385876Z INFO screenpipe_core::pipes: loaded pipe: video-export\n2026-04-20T09:09:04.386843Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary\n2026-04-20T09:09:04.386950Z INFO screenpipe_core::pipes: loaded 6 pipes from \"/Users/lukas/.screenpipe/pipes\"\n\n\n\n _ \n __________________ ___ ____ ____ (_____ ___ \n / ___/ ___/ ___/ _ \\/ _ \\/ __ \\ / __ \\/ / __ \\/ _ \\\n (__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/\n/____/\\___/_/ \\___/\\___/_/ /_/ / .___/_/ .___/\\___/ \n /_/ /_/ \n\n\n\npower AI by everything you've seen, said or heard\nopen source | runs locally | developer friendly\n\n\n┌────────────────────────┬────────────────────────────────────┐\n│ setting │ value │\n├────────────────────────┼────────────────────────────────────┤\n│ audio chunk duration │ 30 seconds │\n│ port │ 3030 │\n│ audio disabled │ true │\n│ vision disabled │ false │\n│ pause on DRM content │ false │\n│ audio engine │ Parakeet │\n│ vad engine │ Silero │\n│ data directory │ /Users/lukas/.screenpipe │\n│ debug mode │ false │\n│ telemetry │ true │\n│ use pii removal │ true │\n│ use all monitors │ true │\n│ ignored windows │ [\"Boosteroid\"] │\n│ included windows │ [] │\n│ cloud sync │ disabled │\n│ auto-destruct pid │ 0 │\n│ deepgram key │ not set │\n│ api auth │ enabled │\n2026-04-20T09:09:04.390629Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)\n2026-04-20T09:09:04.411414Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh\n│ encrypt secrets │ disabled │\n│ retention days │ 14 │\n├────────────────────────┼────────────────────────────────────┤\n│ languages │ │\n│ │ all languages │\n├────────────────────────┼────────────────────────────────────┤\n│ monitors │ │\n│ │ id: 1 │\n│ │ id: 2 │\n├────────────────────────┼────────────────────────────────────┤\n│ audio devices │ │\n│ │ disabled │\n└────────────────────────┴────────────────────────────────────┘\nyou are using local processing. all your data stays on your computer.\n\nwarning: telemetry is enabled. only error-level data will be sent.\nto disable, use the --disable-telemetry flag.\n\ncheck latest changes here: https://github.com/screenpipe/screenpipe/releases\n2026-04-20T09:09:04.444803Z INFO screenpipe: starting UI event capture\n2026-04-20T09:09:04.482387Z INFO screenpipe_engine::ui_recorder: Starting UI event capture\n2026-04-20T09:09:04.509116Z INFO screenpipe_engine::ui_recorder: UI recording session started: f0f4e024-5bd3-4a71-bf92-6feec89df1b0\n2026-04-20T09:09:04.509231Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)\n2026-04-20T09:09:04.509296Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 06:09:04.509295 UTC to 2026-04-20 06:09:04.509295 UTC)\n2026-04-20T09:09:04.510396Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)\n2026-04-20T09:09:04.511440Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 0 frame entries, coverage from 2026-04-19 06:09:04.509295 UTC\n2026-04-20T09:09:04.519984Z INFO screenpipe_engine::server: Server listening on 127.0.0.1:3030\n2026-04-20T09:09:04.547289Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030\n2026-04-20T09:09:04.965425Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)\n2026-04-20T09:09:04.965463Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)\n2026-04-20T09:09:04.965503Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)\n2026-04-20T09:09:05.226982Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)\n2026-04-20T09:09:05.227016Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)\n2026-04-20T09:09:05.227136Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)\n2026-04-20T09:09:05.227375Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)\n2026-04-20T09:09:07.300820Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 3 excluded)\n2026-04-20T09:09:07.600290Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=51507, dur=136ms\n2026-04-20T09:09:07.683736Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 3 excluded)\n2026-04-20T09:09:07.989000Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=51508, dur=185ms\n2026-04-20T09:10:09.231456Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=197 elapsed=4.915444333s\n2026-04-20T09:10:09.231772Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 197 eligible frames\n2026-04-20T09:10:14.691743Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 83 frames, 16.3MB → 2.5MB (6.5x), 83 JPEGs deleted\n2026-04-20T09:10:15.677833Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:17.628368Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:17.716440Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:10:21.386286Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 100 frames, 17.9MB → 6.8MB (2.6x), 100 JPEGs deleted\n2026-04-20T09:10:22.685406Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 2.2MB → 0.9MB (2.4x), 14 JPEGs deleted\n2026-04-20T09:10:23.507120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:42.391615Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:42.475795Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:44.750359Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:44.980679Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:53.686192Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:53.738064Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:56.291827Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:11:56.451083Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)\n2026-04-20T09:12:46.023481Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)\n2026-04-20T09:12:49.049656Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)\n2026-04-20T09:13:23.025287Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=596503974090759224, trigger=click)\n2026-04-20T09:13:28.322087Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3622643354436184557, trigger=visual_change)\n2026-04-20T09:14:35.739274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3577748614007823836, trigger=click)\n2026-04-20T09:15:24.523245Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=3 elapsed=1.840652833s\n2026-04-20T09:15:24.523440Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 3 eligible frames\n2026-04-20T09:15:37.319126Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7719336893453902381, trigger=click)\n2026-04-20T09:16:25.021906Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2226463793404778776, trigger=visual_change)\n2026-04-20T09:16:40.481038Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-2449473341097903678, trigger=visual_change)\n2026-04-20T09:17:01.897197Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)\n2026-04-20T09:17:11.727852Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:11.880765Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:12.886271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:12.961362Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)\n2026-04-20T09:17:49.590890Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-5042772765556709364, trigger=visual_change)\n2026-04-20T09:18:38.516980Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3873630181551472921, trigger=click)\n2026-04-20T09:18:52.171689Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-3192006669261292210, trigger=click)\n2026-04-20T09:18:52.339913Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)\n2026-04-20T09:19:18.962835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)\n2026-04-20T09:19:21.997050Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)\n2026-04-20T09:20:27.568575Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=33 elapsed=3.045052625s\n2026-04-20T09:20:27.568689Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 33 eligible frames\n2026-04-20T09:20:28.589270Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 3.1MB → 0.7MB (4.6x), 14 JPEGs deleted\n2026-04-20T09:20:30.840110Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 19 frames, 3.0MB → 1.1MB (2.8x), 19 JPEGs deleted\n2026-04-20T09:20:37.786800Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2023204823691812477, trigger=visual_change)\n2026-04-20T09:20:41.129431Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-6072351155973619451, trigger=visual_change)\n2026-04-20T09:20:50.199285Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5706498223560949354, trigger=visual_change)\n2026-04-20T09:21:31.117184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:21:34.972211Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:35.170979Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:36.933389Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:37.021845Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:46.669822Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:21:48.854992Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:48.978898Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:50.309855Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:52.623328Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1415600133454797157, trigger=click)\n2026-04-20T09:21:53.747271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1415600133454797157, trigger=visual_change)\n2026-04-20T09:22:30.747706Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=7611972069272730175, trigger=click)\n2026-04-20T09:23:30.996636Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:23:31.087562Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:23:48.889059Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4881264313375199900, trigger=visual_change)\n2026-04-20T09:24:06.522344Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:24:36.382066Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2367845838743006480, trigger=click)\n2026-04-20T09:24:36.620082Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=2367845838743006480, trigger=click)\n2026-04-20T09:25:35.337676Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5282847726719021120, trigger=visual_change)\n2026-04-20T09:25:35.422875Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=78 elapsed=4.581551458s\n2026-04-20T09:25:35.423240Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 78 eligible frames\n2026-04-20T09:25:37.400513Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 30 frames, 10.0MB → 2.2MB (4.6x), 30 JPEGs deleted\n2026-04-20T09:25:41.702147Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 46 frames, 6.5MB → 3.7MB (1.7x), 46 JPEGs deleted\n2026-04-20T09:26:39.541417Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5282847726719021120, trigger=click)\n2026-04-20T09:26:39.637408Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5282847726719021120, trigger=visual_change)\n2026-04-20T09:26:45.621450Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=4022234323396946320, trigger=click)\n2026-04-20T09:27:27.813009Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=8695496312677497071, trigger=visual_change)\n2026-04-20T09:27:56.174253Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-1474433992317029253, trigger=visual_change)\n2026-04-20T09:28:15.443669Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:15.559560Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:22.871037Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:22.998585Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:23.546566Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:27.616599Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:27.810327Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:28.805551Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:28.996759Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:38.226782Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:41.919454Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:43.104835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:28:43.291487Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:29:15.175914Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:29:15.288120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5262443228019302590, trigger=click)\n2026-04-20T09:30:44.693517Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=152 elapsed=2.991660041s\n2026-04-20T09:30:44.693996Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 152 eligible frames\n2026-04-20T09:30:49.454089Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 62 frames, 18.7MB → 3.2MB (5.8x), 62 JPEGs deleted\n2026-04-20T09:31:02.747279Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 88 frames, 12.9MB → 5.8MB (2.2x), 88 JPEGs deleted\n2026-04-20T09:32:40.896832Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2273900790310487349, trigger=visual_change)\n2026-04-20T09:34:10.145625Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=416962040610113261, trigger=visual_change)\n2026-04-20T09:36:04.111438Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=113 elapsed=1.365217541s\n2026-04-20T09:36:04.111554Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 113 eligible frames\n2026-04-20T09:36:07.290184Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 50 frames, 10.6MB → 1.1MB (9.7x), 50 JPEGs deleted\n2026-04-20T09:36:11.502821Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 61 frames, 9.4MB → 4.3MB (2.2x), 61 JPEGs deleted\n2026-04-20T09:36:23.826107Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:30.222596Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:30.344862Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:37.846785Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=visual_change)\n2026-04-20T09:36:39.497294Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:39.674063Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:43.099360Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1682368260798788554, trigger=click)\n2026-04-20T09:36:44.941648Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:48.648990Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:48.775433Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-843838009161543790, trigger=click)\n2026-04-20T09:36:59.935381Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5838015891328253158, trigger=click)\n2026-04-20T09:37:00.277881Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5838015891328253158, trigger=click)\n2026-04-20T09:37:07.535793Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:08.980062Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:09.230737Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:11.150616Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:37:11.367104Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5568544385744823258, trigger=click)\n2026-04-20T09:38:39.474470Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=click)\n2026-04-20T09:41:06.365928Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:13.008346Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:20.360309Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=114 elapsed=8.859599166s\n2026-04-20T09:41:20.362479Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 114 eligible frames\n2026-04-20T09:41:25.798397Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 52 frames, 9.9MB → 1.1MB (9.3x), 52 JPEGs deleted\n2026-04-20T09:41:41.703983Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 60 frames, 8.1MB → 3.4MB (2.4x), 60 JPEGs deleted\n2026-04-20T09:41:48.137009Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:50.361864Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:52.640333Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:41:58.785028Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:41:59.196006Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:02.302969Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:02.491579Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:05.398822Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-6926500999441679870, trigger=visual_change)\n2026-04-20T09:42:15.779232Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:24.922773Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:28.057485Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:28.270226Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:30.197825Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:42:32.654724Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:32.947917Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:50.661919Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:42:54.467651Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:06.075687Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:06.266678Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:13.645777Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:14.025201Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:17.704682Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:18.003229Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:23.818550Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:25.680139Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:26.624253Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=click)\n2026-04-20T09:43:31.133945Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:43:31.650235Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7700553420326176627, trigger=visual_change)\n2026-04-20T09:45:02.727558Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=4753636820005661432, trigger=visual_change)\n2026-04-20T09:46:50.876930Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=74 elapsed=9.106527083s\n2026-04-20T09:46:50.889948Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 74 eligible frames\n2026-04-20T09:46:55.120250Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 31 frames, 5.7MB → 0.3MB (19.2x), 31 JPEGs deleted\n2026-04-20T09:47:04.755567Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 41 frames, 6.0MB → 1.8MB (3.3x), 41 JPEGs deleted\n2026-04-20T09:47:04.770274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6010633228612407689, trigger=click)\n2026-04-20T09:47:32.689458Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-148021406869771861, trigger=visual_change)\n2026-04-20T09:47:35.766223Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-148021406869771861, trigger=visual_change)\n2026-04-20T09:47:58.935503Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-6169109282047323600, trigger=click)\n2026-04-20T09:52:12.454962Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=97 elapsed=7.696011375s\n2026-04-20T09:52:12.459566Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 97 eligible frames\n2026-04-20T09:52:23.116499Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 51 frames, 11.8MB → 2.4MB (4.9x), 51 JPEGs deleted\n2026-04-20T09:52:31.720397Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 44 frames, 7.6MB → 2.8MB (2.7x), 44 JPEGs deleted\n2026-04-20T09:52:59.219405Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=1610157915853078934, trigger=visual_change)\n2026-04-20T09:53:45.292944Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5708228181494047783, trigger=visual_change)\n2026-04-20T09:53:48.323898Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5708228181494047783, trigger=visual_change)\n2026-04-20T09:55:24.761072Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-348703920195698856, trigger=visual_change)\n2026-04-20T09:55:36.910247Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-7110021974046204241, trigger=visual_change)\n2026-04-20T09:56:10.545531Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=2028581362717921873, trigger=visual_change)\n2026-04-20T09:56:57.973992Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=8927321389811673919, trigger=click)\n2026-04-20T09:57:46.613875Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=104 elapsed=14.8827255s\n2026-04-20T09:57:46.614109Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 104 eligible frames\n2026-04-20T09:58:03.737408Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 61 frames, 10.5MB → 4.1MB (2.6x), 61 JPEGs deleted\n2026-04-20T09:58:09.508886Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 41 frames, 6.4MB → 1.8MB (3.5x), 41 JPEGs deleted\n2026-04-20T09:58:15.283826Z WARN sqlx::query: summary=\"COMMIT\" db.statement=\"\" rows_affected=1 rows_returned=0 elapsed=1.503935584s\n2026-04-20T09:58:46.569792Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=5225052939832040319, trigger=visual_change)\n2026-04-20T10:00:52.594246Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=305670337277957061, trigger=visual_change)\n2026-04-20T10:01:23.329861Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-5911983749499347001, trigger=visual_change)\n2026-04-20T10:02:18.294628Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:02:21.424720Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:02:24.423729Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4294515861318519702, trigger=visual_change)\n2026-04-20T10:03:22.734042Z WARN sqlx::query: summary=\"SELECT id, snapshot_path, device_name, …\" db.statement=\"\\n\\nSELECT\\n id,\\n snapshot_path,\\n device_name,\\n timestamp\\nFROM\\n frames\\nWHERE\\n snapshot_path IS NOT NULL\\n AND timestamp < ?1\\nORDER BY\\n device_name,\\n timestamp ASC\\nLIMIT\\n 5000\\n\" rows_affected=0 rows_returned=82 elapsed=13.22261775s\n2026-04-20T10:03:22.753002Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 82 eligible frames\n2026-04-20T10:03:40.239737Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 48 frames, 8.6MB → 3.5MB (2.5x), 48 JPEGs deleted\n2026-04-20T10:03:45.878728Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 32 frames, 4.8MB → 1.1MB (4.3x), 32 JPEGs deleted","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.140625,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.14479166,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.28125,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.28541666,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.421875,"top":0.05888889,"width":0.140625,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.42604166,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.5625,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.56666666,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"✳ Review screenpipe usage and Boosteroid integration (claude)","depth":2,"bounds":{"left":0.7027778,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.70694447,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.84305555,"top":0.05888889,"width":0.14027777,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.8472222,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.9548611,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"screenpipe\"","depth":1,"bounds":{"left":0.47083333,"top":0.033333335,"width":0.058333334,"height":0.017777778},"role_description":"text"}]...
|
-7539762675761127371
|
-1969116165549787679
|
click
|
accessibility
|
NULL
|
ggml_metal_init: use fusion = true
ggml_me ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
2026-04-18T14:43:07.918412Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks
2026-04-18T14:44:35.514401Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-8593116969702213352, trigger=visual_change)
2026-04-18T14:44:35.891413Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-8593116969702213352, trigger=visual_change)
2026-04-18T14:45:06.077184Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=visual_change)
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
whisper_backend_init_gpu: device 0: Metal (type: 1)
whisper_backend_init_gpu: found GPU device 0: Metal (type: 1, cnt: 0)
whisper_backend_init_gpu: using Metal backend
ggml_metal_init: allocating
ggml_metal_init: found device: Apple M1
ggml_metal_init: picking default device: Apple M1
ggml_metal_init: use fusion = true
ggml_metal_init: use concurrency = true
ggml_metal_init: use graph optimize = true
whisper_backend_init: using BLAS backend
whisper_init_state: kv self size = 3.15 MB
whisper_init_state: kv cross size = 9.44 MB
whisper_init_state: kv pad size = 2.36 MB
whisper_init_state: compute buffer (conv) = 14.17 MB
whisper_init_state: compute buffer (encode) = 65.96 MB
whisper_init_state: compute buffer (cross) = 8.50 MB
whisper_init_state: compute buffer (decode) = 96.83 MB
ggml_metal_free: deallocating
2026-04-18T14:45:09.423485Z INFO screenpipe_audio::audio_manager::manager: reconciliation: transcribed 50 orphaned chunks
2026-04-18T14:45:40.853509Z WARN screenpipe_db::db: ImmediateTx dropped without commit — rolling back
2026-04-18T14:45:45.829292Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6130095137839516019, trigger=click)
2026-04-18T14:45:45.931557Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6130095137839516019, trigger=click)
zsh: terminated npx screenpipe@latest record --ignored-windows "Boosteroid"
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sp-start
detected hardware tier: Mid
warning: parakeet is not supported on this platform, using whisper-tiny instead
2026-04-20T09:09:02.916408Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store
checking permissions...
screen recording: ok
accessibility: ok
2026-04-20T09:09:03.042870Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6
2026-04-20T09:09:04.217747Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor
2026-04-20T09:09:04.219685Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)
2026-04-20T09:09:04.220577Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)
2026-04-20T09:09:04.314160Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true
2026-04-20T09:09:04.314251Z INFO screenpipe: meeting detector enabled — independent of transcription mode
2026-04-20T09:09:04.314473Z INFO screenpipe: API server listening on [IP_ADDRESS]:3030 (localhost only)
2026-04-20T09:09:04.314486Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key
2026-04-20T09:09:04.314428Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)
2026-04-20T09:09:04.314439Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)
2026-04-20T09:09:04.314563Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager
2026-04-20T09:09:04.370709Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))
2026-04-20T09:09:04.383079Z INFO screenpipe_core::pipes: loaded pipe: day-recap
2026-04-20T09:09:04.384136Z INFO screenpipe_core::pipes: loaded pipe: standup-update
2026-04-20T09:09:04.384745Z INFO screenpipe_core::pipes: loaded pipe: ai-habits
2026-04-20T09:09:04.385360Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown
2026-04-20T09:09:04.385876Z INFO screenpipe_core::pipes: loaded pipe: video-export
2026-04-20T09:09:04.386843Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary
2026-04-20T09:09:04.386950Z INFO screenpipe_core::pipes: loaded 6 pipes from "/Users/lukas/.screenpipe/pipes"
_
__________________ ___ ____ ____ (_____ ___
/ ___/ ___/ ___/ _ \/ _ \/ __ \ / __ \/ / __ \/ _ \
(__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/
/____/\___/_/ \___/\___/_/ /_/ / .___/_/ .___/\___/
/_/ /_/
power AI by everything you've seen, said or heard
open source | runs locally | developer friendly
┌────────────────────────┬────────────────────────────────────┐
│ setting │ value │
├────────────────────────┼────────────────────────────────────┤
│ audio chunk duration │ 30 seconds │
│ port │ 3030 │
│ audio disabled │ true │
│ vision disabled │ false │
│ pause on DRM content │ false │
│ audio engine │ Parakeet │
│ vad engine │ Silero │
│ data directory │ /Users/lukas/.screenpipe │
│ debug mode │ false │
│ telemetry │ true │
│ use pii removal │ true │
│ use all monitors │ true │
│ ignored windows │ ["Boosteroid"] │
│ included windows │ [] │
│ cloud sync │ disabled │
│ auto-destruct pid │ 0 │
│ deepgram key │ not set │
│ api auth │ enabled │
2026-04-20T09:09:04.390629Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)
2026-04-20T09:09:04.411414Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh
│ encrypt secrets │ disabled │
│ retention days │ 14 │
├────────────────────────┼────────────────────────────────────┤
│ languages │ │
│ │ all languages │
├────────────────────────┼────────────────────────────────────┤
│ monitors │ │
│ │ id: 1 │
│ │ id: 2 │
├────────────────────────┼────────────────────────────────────┤
│ audio devices │ │
│ │ disabled │
└────────────────────────┴────────────────────────────────────┘
you are using local processing. all your data stays on your computer.
warning: telemetry is enabled. only error-level data will be sent.
to disable, use the --disable-telemetry flag.
check latest changes here: https://github.com/screenpipe/screenpipe/releases
2026-04-20T09:09:04.444803Z INFO screenpipe: starting UI event capture
2026-04-20T09:09:04.482387Z INFO screenpipe_engine::ui_recorder: Starting UI event capture
2026-04-20T09:09:04.509116Z INFO screenpipe_engine::ui_recorder: UI recording session started: f0f4e024-5bd3-4a71-bf92-6feec89df1b0
2026-04-20T09:09:04.509231Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)
2026-04-20T09:09:04.509296Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 06:09:04.509295 UTC to 2026-04-20 06:09:04.509295 UTC)
2026-04-20T09:09:04.510396Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)
2026-04-20T09:09:04.511440Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 0 frame entries, coverage from 2026-04-19 06:09:04.509295 UTC
2026-04-20T09:09:04.519984Z INFO screenpipe_engine::server: Server listening on [IP_ADDRESS]:3030
2026-04-20T09:09:04.547289Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030
2026-04-20T09:09:04.965425Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)
2026-04-20T09:09:04.965463Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)
2026-04-20T09:09:04.965503Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)
2026-04-20T09:09:05.226982Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)
2026-04-20T09:09:05.227016Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)
2026-04-20T09:09:05.227136Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)
2026-04-20T09:09:05.227375Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)
2026-04-20T09:09:07.300820Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 3 excluded)
2026-04-20T09:09:07.600290Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=51507, dur=136ms
2026-04-20T09:09:07.683736Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 3 excluded)
2026-04-20T09:09:07.989000Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=51508, dur=185ms
2026-04-20T09:10:09.231456Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=197 elapsed=4.915444333s
2026-04-20T09:10:09.231772Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 197 eligible frames
2026-04-20T09:10:14.691743Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 83 frames, 16.3MB → 2.5MB (6.5x), 83 JPEGs deleted
2026-04-20T09:10:15.677833Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:17.628368Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:17.716440Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:10:21.386286Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 100 frames, 17.9MB → 6.8MB (2.6x), 100 JPEGs deleted
2026-04-20T09:10:22.685406Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 2.2MB → 0.9MB (2.4x), 14 JPEGs deleted
2026-04-20T09:10:23.507120Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:42.391615Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:42.475795Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:44.750359Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:44.980679Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:53.686192Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:53.738064Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:56.291827Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=6544687314381512705, trigger=click)
2026-04-20T09:11:56.451083Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=6544687314381512705, trigger=click)
2026-04-20T09:12:46.023481Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)
2026-04-20T09:12:49.049656Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=1367434359766307217, trigger=visual_change)
2026-04-20T09:13:23.025287Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=596503974090759224, trigger=click)
2026-04-20T09:13:28.322087Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3622643354436184557, trigger=visual_change)
2026-04-20T09:14:35.739274Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3577748614007823836, trigger=click)
2026-04-20T09:15:24.523245Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=3 elapsed=1.840652833s
2026-04-20T09:15:24.523440Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 3 eligible frames
2026-04-20T09:15:37.319126Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-7719336893453902381, trigger=click)
2026-04-20T09:16:25.021906Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=2226463793404778776, trigger=visual_change)
2026-04-20T09:16:40.481038Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-2449473341097903678, trigger=visual_change)
2026-04-20T09:17:01.897197Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)
2026-04-20T09:17:11.727852Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:11.880765Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:12.886271Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:12.961362Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-4839184715325635051, trigger=click)
2026-04-20T09:17:49.590890Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-5042772765556709364, trigger=visual_change)
2026-04-20T09:18:38.516980Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=3873630181551472921, trigger=click)
2026-04-20T09:18:52.171689Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=-3192006669261292210, trigger=click)
2026-04-20T09:18:52.339913Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (hash=-3192006669261292210, trigger=visual_change)
2026-04-20T09:19:18.962835Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)
2026-04-20T09:19:21.997050Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 1 (hash=3035482053266865517, trigger=visual_change)
2026-04-20T09:20:27.568575Z WARN sqlx::query: summary="SELECT id, snapshot_path, device_name, …" db.statement="\n\nSELECT\n id,\n snapshot_path,\n device_name,\n timestamp\nFROM\n frames\nWHERE\n snapshot_path IS NOT NULL\n AND timestamp < ?1\nORDER BY\n device_name,\n timestamp ASC\nLIMIT\n 5000\n" rows_affected=0 rows_returned=33 elapsed=3.045052625s
2026-04-20T09:20:27.568689Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: found 33 eligible frames
2026-04-20T09:20:28.589270Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 14 frames, 3.1MB → 0.7MB (4.6x), 14 JPEGs deleted
2026-04-20T09:20:30.840110Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction: 19 frames, 3.0MB → 1.1MB (2.8x), 19 JPEGs deleted
2026-04-20T09:20:37.786800Z INFO screenpipe_engine::event_driven_capture: content dedup: skipping capture for monitor 2 (...
|
NULL
|
|
52481
|
1134
|
0
|
2026-04-20T07:17:09.786244+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669429786_m1.jpg...
|
iTerm2
|
screenpipe"
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys009
Poetry Last login: Mon Apr 20 10:16:41 on ttys009
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start
detected hardware tier: Mid
warning: parakeet is not supported on this platform, using whisper-tiny instead
2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store
checking permissions...
screen recording: ok
accessibility: ok
2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6
2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor
2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)
2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)
2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true
2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode
2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on [IP_ADDRESS]:3030 (localhost only)
2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key
2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)
2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager
2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)
2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))
2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap
2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update
2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits
2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown
2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export
2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary
2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from "/Users/lukas/.screenpipe/pipes"
_
__________________ ___ ____ ____ (_____ ___
/ ___/ ___/ ___/ _ \/ _ \/ __ \ / __ \/ / __ \/ _ \
(__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/
/____/\___/_/ \___/\___/_/ /_/ / .___/_/ .___/\___/
/_/ /_/
power AI by everything you've seen, said or heard
open source | runs locally | developer friendly
┌────────────────────────┬────────────────────────────────────┐
│ setting │ value │
├────────────────────────┼────────────────────────────────────┤
│ audio chunk duration │ 30 seconds │
│ port │ 3030 │
│ audio disabled │ true │
│ vision disabled │ false │
│ pause on DRM content │ false │
│ audio engine │ Parakeet │
│ vad engine │ Silero │
│ data directory │ /Users/lukas/.screenpipe │
│ debug mode │ false │
│ telemetry │ true │
│ use pii removal │ true │
│ use all monitors │ true │
2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)
│ ignored windows │ ["Boosteroid"] │
│ included windows │ [] │
│ cloud sync │ disabled │
│ auto-destruct pid │ 0 │
│ deepgram key │ not set │
│ api auth │ enabled │
2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh
│ encrypt secrets │ disabled │
│ retention days │ 14 │
├────────────────────────┼────────────────────────────────────┤
│ languages │ │
│ │ all languages │
├────────────────────────┼────────────────────────────────────┤
│ monitors │ │
│ │ id: 1 │
│ │ id: 2 │
├────────────────────────┼────────────────────────────────────┤
│ audio devices │ │
│ │ disabled │
└────────────────────────┴────────────────────────────────────┘
you are using local processing. all your data stays on your computer.
warning: telemetry is enabled. only error-level data will be sent.
to disable, use the --disable-telemetry flag.
check latest changes here: https://github.com/screenpipe/screenpipe/releases
2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture
2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture
2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)
2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e
2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)
2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)
2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on [IP_ADDRESS]:3030
2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030
2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)
2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)
2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)
2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
screenpipe"...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys009\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start \ndetected hardware tier: Mid\nwarning: parakeet is not supported on this platform, using whisper-tiny instead\n2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store\nchecking permissions...\n screen recording: ok\n accessibility: ok\n2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6\n2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor\n2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)\n2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)\n2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true\n2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode\n2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on 127.0.0.1:3030 (localhost only)\n2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key\n2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)\n2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager\n2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)\n2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))\n2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap\n2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update\n2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits\n2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown\n2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export\n2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary\n2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from \"/Users/lukas/.screenpipe/pipes\"\n\n\n\n _ \n __________________ ___ ____ ____ (_____ ___ \n / ___/ ___/ ___/ _ \\/ _ \\/ __ \\ / __ \\/ / __ \\/ _ \\\n (__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/\n/____/\\___/_/ \\___/\\___/_/ /_/ / .___/_/ .___/\\___/ \n /_/ /_/ \n\n\n\npower AI by everything you've seen, said or heard\nopen source | runs locally | developer friendly\n\n\n┌────────────────────────┬────────────────────────────────────┐\n│ setting │ value │\n├────────────────────────┼────────────────────────────────────┤\n│ audio chunk duration │ 30 seconds │\n│ port │ 3030 │\n│ audio disabled │ true │\n│ vision disabled │ false │\n│ pause on DRM content │ false │\n│ audio engine │ Parakeet │\n│ vad engine │ Silero │\n│ data directory │ /Users/lukas/.screenpipe │\n│ debug mode │ false │\n│ telemetry │ true │\n│ use pii removal │ true │\n│ use all monitors │ true │\n2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)\n│ ignored windows │ [\"Boosteroid\"] │\n│ included windows │ [] │\n│ cloud sync │ disabled │\n│ auto-destruct pid │ 0 │\n│ deepgram key │ not set │\n│ api auth │ enabled │\n2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh\n│ encrypt secrets │ disabled │\n│ retention days │ 14 │\n├────────────────────────┼────────────────────────────────────┤\n│ languages │ │\n│ │ all languages │\n├────────────────────────┼────────────────────────────────────┤\n│ monitors │ │\n│ │ id: 1 │\n│ │ id: 2 │\n├────────────────────────┼────────────────────────────────────┤\n│ audio devices │ │\n│ │ disabled │\n└────────────────────────┴────────────────────────────────────┘\nyou are using local processing. all your data stays on your computer.\n\nwarning: telemetry is enabled. only error-level data will be sent.\nto disable, use the --disable-telemetry flag.\n\ncheck latest changes here: https://github.com/screenpipe/screenpipe/releases\n2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture\n2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture\n2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)\n2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e\n2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)\n2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)\n2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on 127.0.0.1:3030\n2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030\n2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)\n2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)\n2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)\n2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)","depth":4,"value":"Last login: Mon Apr 20 10:16:41 on ttys009\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start \ndetected hardware tier: Mid\nwarning: parakeet is not supported on this platform, using whisper-tiny instead\n2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store\nchecking permissions...\n screen recording: ok\n accessibility: ok\n2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6\n2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor\n2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)\n2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)\n2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true\n2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode\n2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on 127.0.0.1:3030 (localhost only)\n2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key\n2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)\n2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager\n2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)\n2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))\n2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap\n2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update\n2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits\n2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown\n2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export\n2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary\n2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from \"/Users/lukas/.screenpipe/pipes\"\n\n\n\n _ \n __________________ ___ ____ ____ (_____ ___ \n / ___/ ___/ ___/ _ \\/ _ \\/ __ \\ / __ \\/ / __ \\/ _ \\\n (__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/\n/____/\\___/_/ \\___/\\___/_/ /_/ / .___/_/ .___/\\___/ \n /_/ /_/ \n\n\n\npower AI by everything you've seen, said or heard\nopen source | runs locally | developer friendly\n\n\n┌────────────────────────┬────────────────────────────────────┐\n│ setting │ value │\n├────────────────────────┼────────────────────────────────────┤\n│ audio chunk duration │ 30 seconds │\n│ port │ 3030 │\n│ audio disabled │ true │\n│ vision disabled │ false │\n│ pause on DRM content │ false │\n│ audio engine │ Parakeet │\n│ vad engine │ Silero │\n│ data directory │ /Users/lukas/.screenpipe │\n│ debug mode │ false │\n│ telemetry │ true │\n│ use pii removal │ true │\n│ use all monitors │ true │\n2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)\n│ ignored windows │ [\"Boosteroid\"] │\n│ included windows │ [] │\n│ cloud sync │ disabled │\n│ auto-destruct pid │ 0 │\n│ deepgram key │ not set │\n│ api auth │ enabled │\n2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh\n│ encrypt secrets │ disabled │\n│ retention days │ 14 │\n├────────────────────────┼────────────────────────────────────┤\n│ languages │ │\n│ │ all languages │\n├────────────────────────┼────────────────────────────────────┤\n│ monitors │ │\n│ │ id: 1 │\n│ │ id: 2 │\n├────────────────────────┼────────────────────────────────────┤\n│ audio devices │ │\n│ │ disabled │\n└────────────────────────┴────────────────────────────────────┘\nyou are using local processing. all your data stays on your computer.\n\nwarning: telemetry is enabled. only error-level data will be sent.\nto disable, use the --disable-telemetry flag.\n\ncheck latest changes here: https://github.com/screenpipe/screenpipe/releases\n2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture\n2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture\n2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)\n2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e\n2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)\n2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)\n2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on 127.0.0.1:3030\n2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030\n2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)\n2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)\n2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)\n2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.19722222,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.2013889,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.39444444,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.3986111,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.59166664,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.59583336,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.7888889,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.79305553,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.95763886,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"screenpipe\"","depth":1,"bounds":{"left":0.4722222,"top":0.033333335,"width":0.058333334,"height":0.017777778},"role_description":"text"}]...
|
-3311039590643839740
|
7761768586887812193
|
manual
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys009
Poetry Last login: Mon Apr 20 10:16:41 on ttys009
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start
detected hardware tier: Mid
warning: parakeet is not supported on this platform, using whisper-tiny instead
2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store
checking permissions...
screen recording: ok
accessibility: ok
2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6
2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor
2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)
2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)
2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true
2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode
2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on [IP_ADDRESS]:3030 (localhost only)
2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key
2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)
2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager
2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)
2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))
2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap
2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update
2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits
2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown
2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export
2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary
2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from "/Users/lukas/.screenpipe/pipes"
_
__________________ ___ ____ ____ (_____ ___
/ ___/ ___/ ___/ _ \/ _ \/ __ \ / __ \/ / __ \/ _ \
(__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/
/____/\___/_/ \___/\___/_/ /_/ / .___/_/ .___/\___/
/_/ /_/
power AI by everything you've seen, said or heard
open source | runs locally | developer friendly
┌────────────────────────┬────────────────────────────────────┐
│ setting │ value │
├────────────────────────┼────────────────────────────────────┤
│ audio chunk duration │ 30 seconds │
│ port │ 3030 │
│ audio disabled │ true │
│ vision disabled │ false │
│ pause on DRM content │ false │
│ audio engine │ Parakeet │
│ vad engine │ Silero │
│ data directory │ /Users/lukas/.screenpipe │
│ debug mode │ false │
│ telemetry │ true │
│ use pii removal │ true │
│ use all monitors │ true │
2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)
│ ignored windows │ ["Boosteroid"] │
│ included windows │ [] │
│ cloud sync │ disabled │
│ auto-destruct pid │ 0 │
│ deepgram key │ not set │
│ api auth │ enabled │
2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh
│ encrypt secrets │ disabled │
│ retention days │ 14 │
├────────────────────────┼────────────────────────────────────┤
│ languages │ │
│ │ all languages │
├────────────────────────┼────────────────────────────────────┤
│ monitors │ │
│ │ id: 1 │
│ │ id: 2 │
├────────────────────────┼────────────────────────────────────┤
│ audio devices │ │
│ │ disabled │
└────────────────────────┴────────────────────────────────────┘
you are using local processing. all your data stays on your computer.
warning: telemetry is enabled. only error-level data will be sent.
to disable, use the --disable-telemetry flag.
check latest changes here: https://github.com/screenpipe/screenpipe/releases
2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture
2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture
2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)
2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e
2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)
2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)
2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on [IP_ADDRESS]:3030
2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030
2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)
2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)
2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)
2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
screenpipe"...
|
NULL
|
|
52482
|
1135
|
0
|
2026-04-20T07:17:09.818609+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669429818_m2.jpg...
|
iTerm2
|
screenpipe"
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys009
Poetry Last login: Mon Apr 20 10:16:41 on ttys009
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start
detected hardware tier: Mid
warning: parakeet is not supported on this platform, using whisper-tiny instead
2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store
checking permissions...
screen recording: ok
accessibility: ok
2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6
2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor
2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)
2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)
2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true
2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode
2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on [IP_ADDRESS]:3030 (localhost only)
2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key
2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)
2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager
2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)
2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))
2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap
2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update
2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits
2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown
2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export
2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary
2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from "/Users/lukas/.screenpipe/pipes"
_
__________________ ___ ____ ____ (_____ ___
/ ___/ ___/ ___/ _ \/ _ \/ __ \ / __ \/ / __ \/ _ \
(__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/
/____/\___/_/ \___/\___/_/ /_/ / .___/_/ .___/\___/
/_/ /_/
power AI by everything you've seen, said or heard
open source | runs locally | developer friendly
┌────────────────────────┬────────────────────────────────────┐
│ setting │ value │
├────────────────────────┼────────────────────────────────────┤
│ audio chunk duration │ 30 seconds │
│ port │ 3030 │
│ audio disabled │ true │
│ vision disabled │ false │
│ pause on DRM content │ false │
│ audio engine │ Parakeet │
│ vad engine │ Silero │
│ data directory │ /Users/lukas/.screenpipe │
│ debug mode │ false │
│ telemetry │ true │
│ use pii removal │ true │
│ use all monitors │ true │
2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)
│ ignored windows │ ["Boosteroid"] │
│ included windows │ [] │
│ cloud sync │ disabled │
│ auto-destruct pid │ 0 │
│ deepgram key │ not set │
│ api auth │ enabled │
2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh
│ encrypt secrets │ disabled │
│ retention days │ 14 │
├────────────────────────┼────────────────────────────────────┤
│ languages │ │
│ │ all languages │
├────────────────────────┼────────────────────────────────────┤
│ monitors │ │
│ │ id: 1 │
│ │ id: 2 │
├────────────────────────┼────────────────────────────────────┤
│ audio devices │ │
│ │ disabled │
└────────────────────────┴────────────────────────────────────┘
you are using local processing. all your data stays on your computer.
warning: telemetry is enabled. only error-level data will be sent.
to disable, use the --disable-telemetry flag.
check latest changes here: https://github.com/screenpipe/screenpipe/releases
2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture
2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture
2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)
2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e
2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)
2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)
2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on [IP_ADDRESS]:3030
2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030
2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)
2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)
2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)
2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)
2026-04-20T10:17:10.050320Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 1 excluded)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
screenpipe"...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys009\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start \ndetected hardware tier: Mid\nwarning: parakeet is not supported on this platform, using whisper-tiny instead\n2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store\nchecking permissions...\n screen recording: ok\n accessibility: ok\n2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6\n2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor\n2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)\n2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)\n2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true\n2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode\n2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on 127.0.0.1:3030 (localhost only)\n2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key\n2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)\n2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager\n2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)\n2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))\n2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap\n2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update\n2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits\n2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown\n2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export\n2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary\n2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from \"/Users/lukas/.screenpipe/pipes\"\n\n\n\n _ \n __________________ ___ ____ ____ (_____ ___ \n / ___/ ___/ ___/ _ \\/ _ \\/ __ \\ / __ \\/ / __ \\/ _ \\\n (__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/\n/____/\\___/_/ \\___/\\___/_/ /_/ / .___/_/ .___/\\___/ \n /_/ /_/ \n\n\n\npower AI by everything you've seen, said or heard\nopen source | runs locally | developer friendly\n\n\n┌────────────────────────┬────────────────────────────────────┐\n│ setting │ value │\n├────────────────────────┼────────────────────────────────────┤\n│ audio chunk duration │ 30 seconds │\n│ port │ 3030 │\n│ audio disabled │ true │\n│ vision disabled │ false │\n│ pause on DRM content │ false │\n│ audio engine │ Parakeet │\n│ vad engine │ Silero │\n│ data directory │ /Users/lukas/.screenpipe │\n│ debug mode │ false │\n│ telemetry │ true │\n│ use pii removal │ true │\n│ use all monitors │ true │\n2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)\n│ ignored windows │ [\"Boosteroid\"] │\n│ included windows │ [] │\n│ cloud sync │ disabled │\n│ auto-destruct pid │ 0 │\n│ deepgram key │ not set │\n│ api auth │ enabled │\n2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh\n│ encrypt secrets │ disabled │\n│ retention days │ 14 │\n├────────────────────────┼────────────────────────────────────┤\n│ languages │ │\n│ │ all languages │\n├────────────────────────┼────────────────────────────────────┤\n│ monitors │ │\n│ │ id: 1 │\n│ │ id: 2 │\n├────────────────────────┼────────────────────────────────────┤\n│ audio devices │ │\n│ │ disabled │\n└────────────────────────┴────────────────────────────────────┘\nyou are using local processing. all your data stays on your computer.\n\nwarning: telemetry is enabled. only error-level data will be sent.\nto disable, use the --disable-telemetry flag.\n\ncheck latest changes here: https://github.com/screenpipe/screenpipe/releases\n2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture\n2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture\n2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)\n2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e\n2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)\n2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)\n2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on 127.0.0.1:3030\n2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030\n2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)\n2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)\n2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)\n2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)\n2026-04-20T10:17:10.050320Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 1 excluded)","depth":4,"bounds":{"left":0.27027926,"top":0.15802075,"width":0.4800532,"height":0.84197927},"value":"Last login: Mon Apr 20 10:16:41 on ttys009\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start \ndetected hardware tier: Mid\nwarning: parakeet is not supported on this platform, using whisper-tiny instead\n2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store\nchecking permissions...\n screen recording: ok\n accessibility: ok\n2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6\n2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor\n2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)\n2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)\n2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true\n2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode\n2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on 127.0.0.1:3030 (localhost only)\n2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key\n2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)\n2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager\n2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)\n2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))\n2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap\n2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update\n2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits\n2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown\n2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export\n2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary\n2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from \"/Users/lukas/.screenpipe/pipes\"\n\n\n\n _ \n __________________ ___ ____ ____ (_____ ___ \n / ___/ ___/ ___/ _ \\/ _ \\/ __ \\ / __ \\/ / __ \\/ _ \\\n (__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/\n/____/\\___/_/ \\___/\\___/_/ /_/ / .___/_/ .___/\\___/ \n /_/ /_/ \n\n\n\npower AI by everything you've seen, said or heard\nopen source | runs locally | developer friendly\n\n\n┌────────────────────────┬────────────────────────────────────┐\n│ setting │ value │\n├────────────────────────┼────────────────────────────────────┤\n│ audio chunk duration │ 30 seconds │\n│ port │ 3030 │\n│ audio disabled │ true │\n│ vision disabled │ false │\n│ pause on DRM content │ false │\n│ audio engine │ Parakeet │\n│ vad engine │ Silero │\n│ data directory │ /Users/lukas/.screenpipe │\n│ debug mode │ false │\n│ telemetry │ true │\n│ use pii removal │ true │\n│ use all monitors │ true │\n2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)\n│ ignored windows │ [\"Boosteroid\"] │\n│ included windows │ [] │\n│ cloud sync │ disabled │\n│ auto-destruct pid │ 0 │\n│ deepgram key │ not set │\n│ api auth │ enabled │\n2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh\n│ encrypt secrets │ disabled │\n│ retention days │ 14 │\n├────────────────────────┼────────────────────────────────────┤\n│ languages │ │\n│ │ all languages │\n├────────────────────────┼────────────────────────────────────┤\n│ monitors │ │\n│ │ id: 1 │\n│ │ id: 2 │\n├────────────────────────┼────────────────────────────────────┤\n│ audio devices │ │\n│ │ disabled │\n└────────────────────────┴────────────────────────────────────┘\nyou are using local processing. all your data stays on your computer.\n\nwarning: telemetry is enabled. only error-level data will be sent.\nto disable, use the --disable-telemetry flag.\n\ncheck latest changes here: https://github.com/screenpipe/screenpipe/releases\n2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture\n2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture\n2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)\n2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e\n2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)\n2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)\n2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on 127.0.0.1:3030\n2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030\n2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)\n2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)\n2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)\n2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)\n2026-04-20T10:17:10.050320Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 1 excluded)","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.27027926,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.27227393,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.36469415,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.36668882,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.45910904,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.46110374,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.55352396,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.5555186,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.64793885,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.6499335,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.7287234,"top":1.0,"width":0.01861702,"height":-0.023144484},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"screenpipe\"","depth":1,"bounds":{"left":0.49634308,"top":1.0,"width":0.027925532,"height":-0.02394259},"role_description":"text"}]...
|
6514894978562218628
|
7185307697145435233
|
manual
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys009
Poetry Last login: Mon Apr 20 10:16:41 on ttys009
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start
detected hardware tier: Mid
warning: parakeet is not supported on this platform, using whisper-tiny instead
2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store
checking permissions...
screen recording: ok
accessibility: ok
2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6
2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor
2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)
2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)
2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true
2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode
2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on [IP_ADDRESS]:3030 (localhost only)
2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key
2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)
2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager
2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)
2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))
2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap
2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update
2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits
2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown
2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export
2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary
2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from "/Users/lukas/.screenpipe/pipes"
_
__________________ ___ ____ ____ (_____ ___
/ ___/ ___/ ___/ _ \/ _ \/ __ \ / __ \/ / __ \/ _ \
(__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/
/____/\___/_/ \___/\___/_/ /_/ / .___/_/ .___/\___/
/_/ /_/
power AI by everything you've seen, said or heard
open source | runs locally | developer friendly
┌────────────────────────┬────────────────────────────────────┐
│ setting │ value │
├────────────────────────┼────────────────────────────────────┤
│ audio chunk duration │ 30 seconds │
│ port │ 3030 │
│ audio disabled │ true │
│ vision disabled │ false │
│ pause on DRM content │ false │
│ audio engine │ Parakeet │
│ vad engine │ Silero │
│ data directory │ /Users/lukas/.screenpipe │
│ debug mode │ false │
│ telemetry │ true │
│ use pii removal │ true │
│ use all monitors │ true │
2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)
│ ignored windows │ ["Boosteroid"] │
│ included windows │ [] │
│ cloud sync │ disabled │
│ auto-destruct pid │ 0 │
│ deepgram key │ not set │
│ api auth │ enabled │
2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh
│ encrypt secrets │ disabled │
│ retention days │ 14 │
├────────────────────────┼────────────────────────────────────┤
│ languages │ │
│ │ all languages │
├────────────────────────┼────────────────────────────────────┤
│ monitors │ │
│ │ id: 1 │
│ │ id: 2 │
├────────────────────────┼────────────────────────────────────┤
│ audio devices │ │
│ │ disabled │
└────────────────────────┴────────────────────────────────────┘
you are using local processing. all your data stays on your computer.
warning: telemetry is enabled. only error-level data will be sent.
to disable, use the --disable-telemetry flag.
check latest changes here: https://github.com/screenpipe/screenpipe/releases
2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture
2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture
2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)
2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e
2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)
2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)
2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on [IP_ADDRESS]:3030
2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030
2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)
2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)
2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)
2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)
2026-04-20T10:17:10.050320Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 1 excluded)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
screenpipe"...
|
NULL
|
|
52483
|
1134
|
1
|
2026-04-20T07:17:12.406158+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669432406_m1.jpg...
|
iTerm2
|
screenpipe"
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys009
Poetry Last login: Mon Apr 20 10:16:41 on ttys009
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start
detected hardware tier: Mid
warning: parakeet is not supported on this platform, using whisper-tiny instead
2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store
checking permissions...
screen recording: ok
accessibility: ok
2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6
2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor
2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)
2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)
2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true
2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode
2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on [IP_ADDRESS]:3030 (localhost only)
2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key
2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)
2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager
2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)
2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))
2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap
2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update
2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits
2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown
2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export
2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary
2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from "/Users/lukas/.screenpipe/pipes"
_
__________________ ___ ____ ____ (_____ ___
/ ___/ ___/ ___/ _ \/ _ \/ __ \ / __ \/ / __ \/ _ \
(__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/
/____/\___/_/ \___/\___/_/ /_/ / .___/_/ .___/\___/
/_/ /_/
power AI by everything you've seen, said or heard
open source | runs locally | developer friendly
┌────────────────────────┬────────────────────────────────────┐
│ setting │ value │
├────────────────────────┼────────────────────────────────────┤
│ audio chunk duration │ 30 seconds │
│ port │ 3030 │
│ audio disabled │ true │
│ vision disabled │ false │
│ pause on DRM content │ false │
│ audio engine │ Parakeet │
│ vad engine │ Silero │
│ data directory │ /Users/lukas/.screenpipe │
│ debug mode │ false │
│ telemetry │ true │
│ use pii removal │ true │
│ use all monitors │ true │
2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)
│ ignored windows │ ["Boosteroid"] │
│ included windows │ [] │
│ cloud sync │ disabled │
│ auto-destruct pid │ 0 │
│ deepgram key │ not set │
│ api auth │ enabled │
2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh
│ encrypt secrets │ disabled │
│ retention days │ 14 │
├────────────────────────┼────────────────────────────────────┤
│ languages │ │
│ │ all languages │
├────────────────────────┼────────────────────────────────────┤
│ monitors │ │
│ │ id: 1 │
│ │ id: 2 │
├────────────────────────┼────────────────────────────────────┤
│ audio devices │ │
│ │ disabled │
└────────────────────────┴────────────────────────────────────┘
you are using local processing. all your data stays on your computer.
warning: telemetry is enabled. only error-level data will be sent.
to disable, use the --disable-telemetry flag.
check latest changes here: https://github.com/screenpipe/screenpipe/releases
2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture
2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture
2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)
2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e
2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)
2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)
2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on [IP_ADDRESS]:3030
2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030
2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)
2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)
2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)
2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)
2026-04-20T10:17:10.050320Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 1 excluded)
2026-04-20T10:17:10.109022Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=52481, dur=72ms
2026-04-20T10:17:10.197530Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=52482, dur=98ms
2026-04-20T10:17:10.926185Z WARN sqlx::query: summary="SELECT f.id, f.timestamp, f.offset_index, …" db.statement="\n\nSELECT\n f.id,\n f.timestamp,\n f.offset_index,\n COALESCE(\n SUBSTR(f.full_text, 1, 200),\n SUBSTR(f.accessibility_text, 1, 200),\n (\n SELECT\n SUBSTR(ot.text, 1, 200)\n FROM\n ocr_text ot\n WHERE\n ot.frame_id = f.id\n LIMIT\n 1\n )\n ) as text,\n COALESCE(\n f.app_name,\n (\n SELECT\n ot.app_name\n FROM\n ocr_text ot\n WHERE\n ot.frame_id = f.id\n LIMIT\n 1\n )\n ) as app_name,\n COALESCE(\n f.window_name,\n (\n SELECT\n ot.window_name\n FROM\n ocr_text ot\n WHERE\n ot.frame_id = f.id\n LIMIT\n 1\n )\n ) as window_name,\n COALESCE(vc.device_name, f.device_name) as screen_device,\n COALESCE(vc.file_path, f.snapshot_path) as video_path,\n COALESCE(vc.fps, 0.033) as chunk_fps,\n f.browser_url,\n f.machine_id\nFROM\n frames f\n LEFT JOIN video_chunks vc ON f.video_chunk_id = vc.id\nWHERE\n f.timestamp >= ?1\n AND f.timestamp <= ?2\n AND COALESCE(vc.file_path, f.snapshot_path, '') NOT LIKE 'cloud://%'\nORDER BY\n f.timestamp DESC,\n f.offset_index DESC\nLIMIT\n 10000\n" rows_affected=0 rows_returned=974 elapsed=1.657379083s
2026-04-20T10:17:10.928810Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 974 frame entries, coverage from 2026-04-19 07:17:09.268169 UTC
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
screenpipe"...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys009\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start \ndetected hardware tier: Mid\nwarning: parakeet is not supported on this platform, using whisper-tiny instead\n2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store\nchecking permissions...\n screen recording: ok\n accessibility: ok\n2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6\n2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor\n2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)\n2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)\n2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true\n2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode\n2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on 127.0.0.1:3030 (localhost only)\n2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key\n2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)\n2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager\n2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)\n2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))\n2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap\n2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update\n2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits\n2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown\n2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export\n2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary\n2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from \"/Users/lukas/.screenpipe/pipes\"\n\n\n\n _ \n __________________ ___ ____ ____ (_____ ___ \n / ___/ ___/ ___/ _ \\/ _ \\/ __ \\ / __ \\/ / __ \\/ _ \\\n (__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/\n/____/\\___/_/ \\___/\\___/_/ /_/ / .___/_/ .___/\\___/ \n /_/ /_/ \n\n\n\npower AI by everything you've seen, said or heard\nopen source | runs locally | developer friendly\n\n\n┌────────────────────────┬────────────────────────────────────┐\n│ setting │ value │\n├────────────────────────┼────────────────────────────────────┤\n│ audio chunk duration │ 30 seconds │\n│ port │ 3030 │\n│ audio disabled │ true │\n│ vision disabled │ false │\n│ pause on DRM content │ false │\n│ audio engine │ Parakeet │\n│ vad engine │ Silero │\n│ data directory │ /Users/lukas/.screenpipe │\n│ debug mode │ false │\n│ telemetry │ true │\n│ use pii removal │ true │\n│ use all monitors │ true │\n2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)\n│ ignored windows │ [\"Boosteroid\"] │\n│ included windows │ [] │\n│ cloud sync │ disabled │\n│ auto-destruct pid │ 0 │\n│ deepgram key │ not set │\n│ api auth │ enabled │\n2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh\n│ encrypt secrets │ disabled │\n│ retention days │ 14 │\n├────────────────────────┼────────────────────────────────────┤\n│ languages │ │\n│ │ all languages │\n├────────────────────────┼────────────────────────────────────┤\n│ monitors │ │\n│ │ id: 1 │\n│ │ id: 2 │\n├────────────────────────┼────────────────────────────────────┤\n│ audio devices │ │\n│ │ disabled │\n└────────────────────────┴────────────────────────────────────┘\nyou are using local processing. all your data stays on your computer.\n\nwarning: telemetry is enabled. only error-level data will be sent.\nto disable, use the --disable-telemetry flag.\n\ncheck latest changes here: https://github.com/screenpipe/screenpipe/releases\n2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture\n2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture\n2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)\n2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e\n2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)\n2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)\n2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on 127.0.0.1:3030\n2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030\n2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)\n2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)\n2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)\n2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)\n2026-04-20T10:17:10.050320Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 1 excluded)\n2026-04-20T10:17:10.109022Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=52481, dur=72ms\n2026-04-20T10:17:10.197530Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=52482, dur=98ms\n2026-04-20T10:17:10.926185Z WARN sqlx::query: summary=\"SELECT f.id, f.timestamp, f.offset_index, …\" db.statement=\"\\n\\nSELECT\\n f.id,\\n f.timestamp,\\n f.offset_index,\\n COALESCE(\\n SUBSTR(f.full_text, 1, 200),\\n SUBSTR(f.accessibility_text, 1, 200),\\n (\\n SELECT\\n SUBSTR(ot.text, 1, 200)\\n FROM\\n ocr_text ot\\n WHERE\\n ot.frame_id = f.id\\n LIMIT\\n 1\\n )\\n ) as text,\\n COALESCE(\\n f.app_name,\\n (\\n SELECT\\n ot.app_name\\n FROM\\n ocr_text ot\\n WHERE\\n ot.frame_id = f.id\\n LIMIT\\n 1\\n )\\n ) as app_name,\\n COALESCE(\\n f.window_name,\\n (\\n SELECT\\n ot.window_name\\n FROM\\n ocr_text ot\\n WHERE\\n ot.frame_id = f.id\\n LIMIT\\n 1\\n )\\n ) as window_name,\\n COALESCE(vc.device_name, f.device_name) as screen_device,\\n COALESCE(vc.file_path, f.snapshot_path) as video_path,\\n COALESCE(vc.fps, 0.033) as chunk_fps,\\n f.browser_url,\\n f.machine_id\\nFROM\\n frames f\\n LEFT JOIN video_chunks vc ON f.video_chunk_id = vc.id\\nWHERE\\n f.timestamp >= ?1\\n AND f.timestamp <= ?2\\n AND COALESCE(vc.file_path, f.snapshot_path, '') NOT LIKE 'cloud://%'\\nORDER BY\\n f.timestamp DESC,\\n f.offset_index DESC\\nLIMIT\\n 10000\\n\" rows_affected=0 rows_returned=974 elapsed=1.657379083s\n2026-04-20T10:17:10.928810Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 974 frame entries, coverage from 2026-04-19 07:17:09.268169 UTC","depth":4,"value":"Last login: Mon Apr 20 10:16:41 on ttys009\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start \ndetected hardware tier: Mid\nwarning: parakeet is not supported on this platform, using whisper-tiny instead\n2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store\nchecking permissions...\n screen recording: ok\n accessibility: ok\n2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6\n2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor\n2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)\n2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)\n2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true\n2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode\n2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on 127.0.0.1:3030 (localhost only)\n2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key\n2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)\n2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager\n2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)\n2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))\n2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap\n2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update\n2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits\n2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown\n2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export\n2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary\n2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from \"/Users/lukas/.screenpipe/pipes\"\n\n\n\n _ \n __________________ ___ ____ ____ (_____ ___ \n / ___/ ___/ ___/ _ \\/ _ \\/ __ \\ / __ \\/ / __ \\/ _ \\\n (__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/\n/____/\\___/_/ \\___/\\___/_/ /_/ / .___/_/ .___/\\___/ \n /_/ /_/ \n\n\n\npower AI by everything you've seen, said or heard\nopen source | runs locally | developer friendly\n\n\n┌────────────────────────┬────────────────────────────────────┐\n│ setting │ value │\n├────────────────────────┼────────────────────────────────────┤\n│ audio chunk duration │ 30 seconds │\n│ port │ 3030 │\n│ audio disabled │ true │\n│ vision disabled │ false │\n│ pause on DRM content │ false │\n│ audio engine │ Parakeet │\n│ vad engine │ Silero │\n│ data directory │ /Users/lukas/.screenpipe │\n│ debug mode │ false │\n│ telemetry │ true │\n│ use pii removal │ true │\n│ use all monitors │ true │\n2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)\n│ ignored windows │ [\"Boosteroid\"] │\n│ included windows │ [] │\n│ cloud sync │ disabled │\n│ auto-destruct pid │ 0 │\n│ deepgram key │ not set │\n│ api auth │ enabled │\n2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh\n│ encrypt secrets │ disabled │\n│ retention days │ 14 │\n├────────────────────────┼────────────────────────────────────┤\n│ languages │ │\n│ │ all languages │\n├────────────────────────┼────────────────────────────────────┤\n│ monitors │ │\n│ │ id: 1 │\n│ │ id: 2 │\n├────────────────────────┼────────────────────────────────────┤\n│ audio devices │ │\n│ │ disabled │\n└────────────────────────┴────────────────────────────────────┘\nyou are using local processing. all your data stays on your computer.\n\nwarning: telemetry is enabled. only error-level data will be sent.\nto disable, use the --disable-telemetry flag.\n\ncheck latest changes here: https://github.com/screenpipe/screenpipe/releases\n2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture\n2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture\n2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)\n2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e\n2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)\n2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)\n2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on 127.0.0.1:3030\n2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030\n2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)\n2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)\n2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)\n2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)\n2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)\n2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)\n2026-04-20T10:17:10.050320Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 1 excluded)\n2026-04-20T10:17:10.109022Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=52481, dur=72ms\n2026-04-20T10:17:10.197530Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=52482, dur=98ms\n2026-04-20T10:17:10.926185Z WARN sqlx::query: summary=\"SELECT f.id, f.timestamp, f.offset_index, …\" db.statement=\"\\n\\nSELECT\\n f.id,\\n f.timestamp,\\n f.offset_index,\\n COALESCE(\\n SUBSTR(f.full_text, 1, 200),\\n SUBSTR(f.accessibility_text, 1, 200),\\n (\\n SELECT\\n SUBSTR(ot.text, 1, 200)\\n FROM\\n ocr_text ot\\n WHERE\\n ot.frame_id = f.id\\n LIMIT\\n 1\\n )\\n ) as text,\\n COALESCE(\\n f.app_name,\\n (\\n SELECT\\n ot.app_name\\n FROM\\n ocr_text ot\\n WHERE\\n ot.frame_id = f.id\\n LIMIT\\n 1\\n )\\n ) as app_name,\\n COALESCE(\\n f.window_name,\\n (\\n SELECT\\n ot.window_name\\n FROM\\n ocr_text ot\\n WHERE\\n ot.frame_id = f.id\\n LIMIT\\n 1\\n )\\n ) as window_name,\\n COALESCE(vc.device_name, f.device_name) as screen_device,\\n COALESCE(vc.file_path, f.snapshot_path) as video_path,\\n COALESCE(vc.fps, 0.033) as chunk_fps,\\n f.browser_url,\\n f.machine_id\\nFROM\\n frames f\\n LEFT JOIN video_chunks vc ON f.video_chunk_id = vc.id\\nWHERE\\n f.timestamp >= ?1\\n AND f.timestamp <= ?2\\n AND COALESCE(vc.file_path, f.snapshot_path, '') NOT LIKE 'cloud://%'\\nORDER BY\\n f.timestamp DESC,\\n f.offset_index DESC\\nLIMIT\\n 10000\\n\" rows_affected=0 rows_returned=974 elapsed=1.657379083s\n2026-04-20T10:17:10.928810Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 974 frame entries, coverage from 2026-04-19 07:17:09.268169 UTC","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.19722222,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.2013889,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.39444444,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.3986111,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.59166664,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.59583336,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.7888889,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.79305553,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.95763886,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"screenpipe\"","depth":1,"bounds":{"left":0.4722222,"top":0.033333335,"width":0.058333334,"height":0.017777778},"role_description":"text"}]...
|
5556097941235749601
|
7234847430214931553
|
visual_change
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys009
Poetry Last login: Mon Apr 20 10:16:41 on ttys009
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ sp-start
detected hardware tier: Mid
warning: parakeet is not supported on this platform, using whisper-tiny instead
2026-04-20T10:17:08.672567Z INFO screenpipe_engine::auth_key: api auth: key resolved via secret store
checking permissions...
screen recording: ok
accessibility: ok
2026-04-20T10:17:08.735636Z INFO screenpipe_screen::monitor::macos_version: Detected macOS version: 14.6
2026-04-20T10:17:09.168849Z INFO screenpipe_engine::sleep_monitor: Starting macOS sleep/wake monitor
2026-04-20T10:17:09.170474Z INFO screenpipe_engine::sleep_monitor: Screen lock/unlock observers registered (CFNotificationCenter)
2026-04-20T10:17:09.170918Z INFO screenpipe_engine::sleep_monitor: Display reconfiguration watcher registered (CGDisplayRegisterReconfigurationCallback)
2026-04-20T10:17:09.203704Z INFO screenpipe_engine::permission_monitor: permission monitor started screen=true mic=true accessibility=true keychain=true
2026-04-20T10:17:09.203780Z INFO screenpipe: meeting detector enabled — independent of transcription mode
2026-04-20T10:17:09.203962Z INFO screenpipe: API server listening on [IP_ADDRESS]:3030 (localhost only)
2026-04-20T10:17:09.203976Z INFO screenpipe: API auth enabled — run `screenpipe auth token` to view your key
2026-04-20T10:17:09.203934Z INFO screenpipe_engine::power::manager: power manager started (poll interval: 10s)
2026-04-20T10:17:09.204030Z INFO screenpipe_engine::vision_manager::manager: Starting VisionManager
2026-04-20T10:17:09.203954Z INFO screenpipe_engine::snapshot_compaction: snapshot compaction worker started (min_age=600s, poll=300s)
2026-04-20T10:17:09.215020Z INFO screenpipe_engine::power::manager: initial power profile: Performance (on_ac=true, battery=Some(100))
2026-04-20T10:17:09.220127Z INFO screenpipe_core::pipes: loaded pipe: day-recap
2026-04-20T10:17:09.220306Z INFO screenpipe_core::pipes: loaded pipe: standup-update
2026-04-20T10:17:09.220863Z INFO screenpipe_core::pipes: loaded pipe: ai-habits
2026-04-20T10:17:09.221034Z INFO screenpipe_core::pipes: loaded pipe: time-breakdown
2026-04-20T10:17:09.221177Z INFO screenpipe_core::pipes: loaded pipe: video-export
2026-04-20T10:17:09.221820Z INFO screenpipe_core::pipes: loaded pipe: meeting-summary
2026-04-20T10:17:09.221839Z INFO screenpipe_core::pipes: loaded 6 pipes from "/Users/lukas/.screenpipe/pipes"
_
__________________ ___ ____ ____ (_____ ___
/ ___/ ___/ ___/ _ \/ _ \/ __ \ / __ \/ / __ \/ _ \
(__ / /__/ / / __/ __/ / / / / /_/ / / /_/ / __/
/____/\___/_/ \___/\___/_/ /_/ / .___/_/ .___/\___/
/_/ /_/
power AI by everything you've seen, said or heard
open source | runs locally | developer friendly
┌────────────────────────┬────────────────────────────────────┐
│ setting │ value │
├────────────────────────┼────────────────────────────────────┤
│ audio chunk duration │ 30 seconds │
│ port │ 3030 │
│ audio disabled │ true │
│ vision disabled │ false │
│ pause on DRM content │ false │
│ audio engine │ Parakeet │
│ vad engine │ Silero │
│ data directory │ /Users/lukas/.screenpipe │
│ debug mode │ false │
│ telemetry │ true │
│ use pii removal │ true │
│ use all monitors │ true │
2026-04-20T10:17:09.223448Z INFO screenpipe_core::pipes: pipe scheduler started (generation 2)
│ ignored windows │ ["Boosteroid"] │
│ included windows │ [] │
│ cloud sync │ disabled │
│ auto-destruct pid │ 0 │
│ deepgram key │ not set │
│ api auth │ enabled │
2026-04-20T10:17:09.227535Z WARN screenpipe: pi agent install failed: bun not found — install from https://bun.sh
│ encrypt secrets │ disabled │
│ retention days │ 14 │
├────────────────────────┼────────────────────────────────────┤
│ languages │ │
│ │ all languages │
├────────────────────────┼────────────────────────────────────┤
│ monitors │ │
│ │ id: 1 │
│ │ id: 2 │
├────────────────────────┼────────────────────────────────────┤
│ audio devices │ │
│ │ disabled │
└────────────────────────┴────────────────────────────────────┘
you are using local processing. all your data stays on your computer.
warning: telemetry is enabled. only error-level data will be sent.
to disable, use the --disable-telemetry flag.
check latest changes here: https://github.com/screenpipe/screenpipe/releases
2026-04-20T10:17:09.240088Z INFO screenpipe: starting UI event capture
2026-04-20T10:17:09.253668Z INFO screenpipe_engine::ui_recorder: Starting UI event capture
2026-04-20T10:17:09.267973Z INFO screenpipe_engine::calendar_speaker_id: speaker identification: started (user_name=<not set>)
2026-04-20T10:17:09.268072Z INFO screenpipe_engine::ui_recorder: UI recording session started: f196d677-2a10-4dec-93e7-9dd9048a1d8e
2026-04-20T10:17:09.268173Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warming from DB (2026-04-19 07:17:09.268169 UTC to 2026-04-20 07:17:09.268169 UTC)
2026-04-20T10:17:09.268665Z INFO screenpipe_engine::meeting_detector: meeting v2: detection loop started (base_interval=5s, profiles=12)
2026-04-20T10:17:09.275204Z INFO screenpipe_engine::server: Server listening on [IP_ADDRESS]:3030
2026-04-20T10:17:09.280547Z INFO screenpipe_connect::mdns: mdns: advertising screenpipe on port 3030
2026-04-20T10:17:09.284876Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 1 (1440x900)
2026-04-20T10:17:09.284906Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.284942Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 1 (device: monitor_1)
2026-04-20T10:17:09.316523Z INFO screenpipe_engine::vision_manager::manager: Starting vision recording for monitor 2 (3008x1253)
2026-04-20T10:17:09.316557Z INFO screenpipe_engine::vision_manager::manager: Starting event-driven capture for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.316583Z INFO screenpipe_engine::vision_manager::monitor_watcher: Starting monitor watcher (event-driven via CGDisplayRegisterReconfigurationCallback, 60s backstop poll)
2026-04-20T10:17:09.316603Z INFO screenpipe_engine::event_driven_capture: event-driven capture started for monitor 2 (device: monitor_2)
2026-04-20T10:17:09.975374Z INFO sck_rs::stream_manager: persistent SCK stream started for display 1 (1440x900, 2fps, 1 excluded)
2026-04-20T10:17:10.050320Z INFO sck_rs::stream_manager: persistent SCK stream started for display 2 (3008x1253, 2fps, 1 excluded)
2026-04-20T10:17:10.109022Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 1: frame_id=52481, dur=72ms
2026-04-20T10:17:10.197530Z INFO screenpipe_engine::event_driven_capture: startup capture for monitor 2: frame_id=52482, dur=98ms
2026-04-20T10:17:10.926185Z WARN sqlx::query: summary="SELECT f.id, f.timestamp, f.offset_index, …" db.statement="\n\nSELECT\n f.id,\n f.timestamp,\n f.offset_index,\n COALESCE(\n SUBSTR(f.full_text, 1, 200),\n SUBSTR(f.accessibility_text, 1, 200),\n (\n SELECT\n SUBSTR(ot.text, 1, 200)\n FROM\n ocr_text ot\n WHERE\n ot.frame_id = f.id\n LIMIT\n 1\n )\n ) as text,\n COALESCE(\n f.app_name,\n (\n SELECT\n ot.app_name\n FROM\n ocr_text ot\n WHERE\n ot.frame_id = f.id\n LIMIT\n 1\n )\n ) as app_name,\n COALESCE(\n f.window_name,\n (\n SELECT\n ot.window_name\n FROM\n ocr_text ot\n WHERE\n ot.frame_id = f.id\n LIMIT\n 1\n )\n ) as window_name,\n COALESCE(vc.device_name, f.device_name) as screen_device,\n COALESCE(vc.file_path, f.snapshot_path) as video_path,\n COALESCE(vc.fps, 0.033) as chunk_fps,\n f.browser_url,\n f.machine_id\nFROM\n frames f\n LEFT JOIN video_chunks vc ON f.video_chunk_id = vc.id\nWHERE\n f.timestamp >= ?1\n AND f.timestamp <= ?2\n AND COALESCE(vc.file_path, f.snapshot_path, '') NOT LIKE 'cloud://%'\nORDER BY\n f.timestamp DESC,\n f.offset_index DESC\nLIMIT\n 10000\n" rows_affected=0 rows_returned=974 elapsed=1.657379083s
2026-04-20T10:17:10.928810Z INFO screenpipe_engine::hot_frame_cache: hot_frame_cache: warmed with 974 frame entries, coverage from 2026-04-19 07:17:09.268169 UTC
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
screenpipe"...
|
NULL
|
|
52492
|
NULL
|
0
|
2026-04-20T07:18:04.548284+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669484548_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Notion••• <›File+EditViewHistoryWindowHelp‹ ≥ 0 Notion••• <›File+EditViewHistoryWindowHelp‹ ≥ 0 hl100% <7*|8 Mon 20 Apr 10:18:04...
|
NULL
|
7131459381894416072
|
NULL
|
visual_change
|
ocr
|
NULL
|
Notion••• <›File+EditViewHistoryWindowHelp‹ ≥ 0 Notion••• <›File+EditViewHistoryWindowHelp‹ ≥ 0 hl100% <7*|8 Mon 20 Apr 10:18:04...
|
NULL
|
|
52497
|
1136
|
1
|
2026-04-20T07:18:45.600972+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669525600_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
NotionFile+EditViewHistoryWindowHelp< →0 ll100% NotionFile+EditViewHistoryWindowHelp< →0 ll100% [8 Mon 20 Apr 10:18:45...
|
NULL
|
9008871063989310405
|
NULL
|
click
|
ocr
|
NULL
|
NotionFile+EditViewHistoryWindowHelp< →0 ll100% NotionFile+EditViewHistoryWindowHelp< →0 ll100% [8 Mon 20 Apr 10:18:45...
|
NULL
|
|
52498
|
1137
|
1
|
2026-04-20T07:18:45.601116+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669525601_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
rirerox••0!IY [SRD-6793] Les Mills activity typeJY rirerox••0!IY [SRD-6793] Les Mills activity typeJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tabu Userpilot | Eventsiminny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksJIMINNYQ Search bookmarks® For you@ Recentv la bookmarks loolba• Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Spacesv E Bookmarks MenuJiminny (New)> → Mozilla FirefoxOther Bookmarkslu Plauorm leamil SE KanbanI Capture TeamI Enterprise Stability I...M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations* Confluence88 TeamsU= Customise sidebarSpaces / Jiminny (New)Platform Team&® Summary& TimelineE BacklogQ Search board0001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed at neriodCOST-EFFECTIVE AND FASTER NUDGESBacklodTJ1Y-2048915 •.0=Investiaate and fix why exceedFontawesome package limitsMAINTENANCEReady Tor DevJY-20564180 0000=8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216 .000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-201572 .000=€Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyDescriptionWe want to allow our users to automate the execution of their AJA prompts in order to save time and have themready when they need them.Create the reports:• admins and managers should be able to automate reports based on their Panorama prompts and savedsearches• the report should be generated in a pdf - use a lightly branded one this time- [URL_WITH_CREDENTIALS] ContIn QADetailsAssigneeReporterDevelopmentReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsosganisatioPriorityFix versionsSprintDaysNeed QA+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 builds5 davs aqoOPEN]A Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1YesloymentsEPTANCE# Archived work itemsComplete sprint100% 52MOn ZU AOr 10:10.433 Ask Rovo# PagesMore 3Group: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-20632A.l Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
NULL
|
2867404871886203379
|
NULL
|
click
|
ocr
|
NULL
|
rirerox••0!IY [SRD-6793] Les Mills activity typeJY rirerox••0!IY [SRD-6793] Les Mills activity typeJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tabu Userpilot | Eventsiminny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksJIMINNYQ Search bookmarks® For you@ Recentv la bookmarks loolba• Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Spacesv E Bookmarks MenuJiminny (New)> → Mozilla FirefoxOther Bookmarkslu Plauorm leamil SE KanbanI Capture TeamI Enterprise Stability I...M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations* Confluence88 TeamsU= Customise sidebarSpaces / Jiminny (New)Platform Team&® Summary& TimelineE BacklogQ Search board0001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed at neriodCOST-EFFECTIVE AND FASTER NUDGESBacklodTJ1Y-2048915 •.0=Investiaate and fix why exceedFontawesome package limitsMAINTENANCEReady Tor DevJY-20564180 0000=8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216 .000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-201572 .000=€Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyDescriptionWe want to allow our users to automate the execution of their AJA prompts in order to save time and have themready when they need them.Create the reports:• admins and managers should be able to automate reports based on their Panorama prompts and savedsearches• the report should be generated in a pdf - use a lightly branded one this time- [URL_WITH_CREDENTIALS] ContIn QADetailsAssigneeReporterDevelopmentReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsosganisatioPriorityFix versionsSprintDaysNeed QA+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 builds5 davs aqoOPEN]A Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1YesloymentsEPTANCE# Archived work itemsComplete sprint100% 52MOn ZU AOr 10:10.433 Ask Rovo# PagesMore 3Group: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-20632A.l Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
NULL
|
|
52499
|
1137
|
2
|
2026-04-20T07:18:47.387350+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669527387_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
rirerox••0!IY [SRD-6793] Les Mills activity typeJY rirerox••0!IY [SRD-6793] Les Mills activity typeJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tabu Userpilot | EventsBookmarksQ Search bookmarksv a bookmarks loolba• Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....v E Bookmarks Menu> → Mozilla FirefoxOther Bookmarksards/37?selected|ssue=JY-18909® For you@ Recent# Starred0+ AppsQ SpacesJiminny (New)lu Plauorm leamil SE KanbanI Capture TeamI Enterprise Stability I...M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations* Confluence88 TeamsU= Customise sidebarSpaces / Jiminny (New)Platform Team &() Summary& TimelineE BacklogQ Search board0001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed at neriodCOST-EFFECTIVE AND FASTER NUDGESBacklodTJY-2048915 •.0=Investiaate and fix why exceedFontawesome package limitsMAINTENANCEReady Tor DevJY-20564180 0000=8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216 .000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-201572 .000=€Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyDescriptionWe want to allow our users to automate the execution of their AJA prompts in order to save time and have themready when they need them.Create the reports:• admins and managers should be able to automate reports based on their Panorama prompts and savedsearches• the report should be generated in a pdf - use a lightly branded one this time- [URL_WITH_CREDENTIALS] ContIn QADetailsAssigneeReporterDevelopmentReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsosganisatioPriorityFix versionsSprintDaysNeed QA+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 builds5 davs aqoOPEN]A Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1Yes>0.loymentsEPTANCE# Archived work itemsComplete sprint100% 52Mon ZU AOr 10.10.413 Ask Rovo# PagesMore 3Group: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-20632A.l Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
NULL
|
7371332409296257910
|
NULL
|
visual_change
|
ocr
|
NULL
|
rirerox••0!IY [SRD-6793] Les Mills activity typeJY rirerox••0!IY [SRD-6793] Les Mills activity typeJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tabu Userpilot | EventsBookmarksQ Search bookmarksv a bookmarks loolba• Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....v E Bookmarks Menu> → Mozilla FirefoxOther Bookmarksards/37?selected|ssue=JY-18909® For you@ Recent# Starred0+ AppsQ SpacesJiminny (New)lu Plauorm leamil SE KanbanI Capture TeamI Enterprise Stability I...M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations* Confluence88 TeamsU= Customise sidebarSpaces / Jiminny (New)Platform Team &() Summary& TimelineE BacklogQ Search board0001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed at neriodCOST-EFFECTIVE AND FASTER NUDGESBacklodTJY-2048915 •.0=Investiaate and fix why exceedFontawesome package limitsMAINTENANCEReady Tor DevJY-20564180 0000=8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216 .000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-201572 .000=€Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyDescriptionWe want to allow our users to automate the execution of their AJA prompts in order to save time and have themready when they need them.Create the reports:• admins and managers should be able to automate reports based on their Panorama prompts and savedsearches• the report should be generated in a pdf - use a lightly branded one this time- [URL_WITH_CREDENTIALS] ContIn QADetailsAssigneeReporterDevelopmentReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsosganisatioPriorityFix versionsSprintDaysNeed QA+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 builds5 davs aqoOPEN]A Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1Yes>0.loymentsEPTANCE# Archived work itemsComplete sprint100% 52Mon ZU AOr 10.10.413 Ask Rovo# PagesMore 3Group: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-20632A.l Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
52498
|
|
52500
|
1136
|
2
|
2026-04-20T07:18:47.631950+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669527631_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
NotionFile+EditViewHistoryWindowHelp‹ >0 ldl100 NotionFile+EditViewHistoryWindowHelp‹ >0 ldl100% <4*Mon 20 Apr 10:18:47...
|
NULL
|
558290945860217492
|
NULL
|
click
|
ocr
|
NULL
|
NotionFile+EditViewHistoryWindowHelp‹ >0 ldl100 NotionFile+EditViewHistoryWindowHelp‹ >0 ldl100% <4*Mon 20 Apr 10:18:47...
|
52497
|
|
52501
|
1136
|
3
|
2026-04-20T07:18:50.000058+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669530000_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
NotionFileEditViewHistoryWindowDOCKERO ₴1power Al NotionFileEditViewHistoryWindowDOCKERO ₴1power Al by everything you've seen, said or heardopen source I runs locally | developer friendlyHelpDEV (-zsh)<lahl100% <47O $82screenpipe"APP (-zsh)• *3-zsh• *4screenpipe"Mon 20 Apr 10:18:491881*5settingvalueaudio chunk durationportaudio disabledvision disabledpause on DRM contentaudio enginevad enginedata directorydebug modetelemetryuse pii removaluse all monitors2026-04-20110:17:09.2234482ignored windowsincluded windowscloud syncauto-destruct piddeepgram keyapi auth2026-04-20T10:17:09.227535Zencrypt secretsretention days30 seconds3030truefalsefalseParakeetSilero/Users/lukas/.screenpipefalsetruetruetrueINFO screenpipe_core::pipes: pipe scheduler started (generation["Boosteroid"]disabled0not setenabledWARN screenpipe: pi agent install failed: bun not found - install from https://bun.shdisabled14Languagesall languagesmonitorsid: 1id: 2audio devicesdisabledyou are using local processing. all your data stays on your computer.warning: telemetry is enabled. only error-level data will be sent.to disable, use the --disable-telemetry flag....
|
NULL
|
6031726032390726378
|
NULL
|
visual_change
|
ocr
|
NULL
|
NotionFileEditViewHistoryWindowDOCKERO ₴1power Al NotionFileEditViewHistoryWindowDOCKERO ₴1power Al by everything you've seen, said or heardopen source I runs locally | developer friendlyHelpDEV (-zsh)<lahl100% <47O $82screenpipe"APP (-zsh)• *3-zsh• *4screenpipe"Mon 20 Apr 10:18:491881*5settingvalueaudio chunk durationportaudio disabledvision disabledpause on DRM contentaudio enginevad enginedata directorydebug modetelemetryuse pii removaluse all monitors2026-04-20110:17:09.2234482ignored windowsincluded windowscloud syncauto-destruct piddeepgram keyapi auth2026-04-20T10:17:09.227535Zencrypt secretsretention days30 seconds3030truefalsefalseParakeetSilero/Users/lukas/.screenpipefalsetruetruetrueINFO screenpipe_core::pipes: pipe scheduler started (generation["Boosteroid"]disabled0not setenabledWARN screenpipe: pi agent install failed: bun not found - install from https://bun.shdisabled14Languagesall languagesmonitorsid: 1id: 2audio devicesdisabledyou are using local processing. all your data stays on your computer.warning: telemetry is enabled. only error-level data will be sent.to disable, use the --disable-telemetry flag....
|
NULL
|
|
52502
|
1137
|
3
|
2026-04-20T07:18:50.359073+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669530359_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
NotioncaltViewmistomWindowheltny.atlassian.net/jir NotioncaltViewmistomWindowheltny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksI JIMINNYIY [SRD-6793] Les Mills activity typeQ Search bookmarks® For youJy-20698 handle failed field sync• RecentSpaces / Jiminny (New)Platform Team&₩ Summarv& Timeline E Backlog@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Trackingv a bookmarks loolbalSprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Search boardQ Spaces00001(JY-18909] (Part2) Automated rer•• dv E Bookmarks Menu@ Ask Jiminny Reports by nikolav-vaJiminny (New)• _ Mozilla FirefoxOther Bookmarkslll Plauorm leam€ New TabI1 SE KanbanIID Capture Teamu Userpilot | EventsI Enterprise Stability I.M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations2 Confluence88 Teams"= Customise sidebarQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask Jiminnyowner user ld is nuliPLATFORM STABILITYBacklogIn QA v+ CreateC S* Improve Storyn GeorgievDimitrovach VS Code5 davs aqoOPEN]enlovmentsE QArint 2 Q2 +1Need QAYesloymentsEPTANCE$0100% LzMon 20 Apr 10:18:50Ask Rovo# Archived work items© PagesMore 3Complete sprintGroup: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-206321∞ee—A.l Panorama> Don't show internalASK ANYTHING ON ANYTHINGDeployedProphet* JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedE JY-199671• •= 1CLONE - (Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedY.IV-206810.5 %2 .00 gTech Navl Imnrove NenendabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
NULL
|
4836158373186621124
|
NULL
|
visual_change
|
ocr
|
NULL
|
NotioncaltViewmistomWindowheltny.atlassian.net/jir NotioncaltViewmistomWindowheltny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksI JIMINNYIY [SRD-6793] Les Mills activity typeQ Search bookmarks® For youJy-20698 handle failed field sync• RecentSpaces / Jiminny (New)Platform Team&₩ Summarv& Timeline E Backlog@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Trackingv a bookmarks loolbalSprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Search boardQ Spaces00001(JY-18909] (Part2) Automated rer•• dv E Bookmarks Menu@ Ask Jiminny Reports by nikolav-vaJiminny (New)• _ Mozilla FirefoxOther Bookmarkslll Plauorm leam€ New TabI1 SE KanbanIID Capture Teamu Userpilot | EventsI Enterprise Stability I.M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations2 Confluence88 Teams"= Customise sidebarQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask Jiminnyowner user ld is nuliPLATFORM STABILITYBacklogIn QA v+ CreateC S* Improve Storyn GeorgievDimitrovach VS Code5 davs aqoOPEN]enlovmentsE QArint 2 Q2 +1Need QAYesloymentsEPTANCE$0100% LzMon 20 Apr 10:18:50Ask Rovo# Archived work items© PagesMore 3Complete sprintGroup: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-206321∞ee—A.l Panorama> Don't show internalASK ANYTHING ON ANYTHINGDeployedProphet* JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedE JY-199671• •= 1CLONE - (Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedY.IV-206810.5 %2 .00 gTech Navl Imnrove NenendabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
NULL
|
|
52503
|
1136
|
4
|
2026-04-20T07:18:53.094627+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669533094_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
NotionFileEditViewHistoryWindowDOCKERO ₴1power Al NotionFileEditViewHistoryWindowDOCKERO ₴1power Al by everything you've seen, said or heardopen source I runs locally | developer friendlyHelpADEV (-zsh)O $82screenpipe"APP (-zsh)• *3-zsh• ₴4100% C8Mon 20 Apr 10:18:52T81screenpipe™*5settingvalueaudio chunk durationportaudio disabledvision disabledpause on DRM contentaudio enginevad enginedata directorydebug modetelemetryuse pii removaluse all monitors2026-04-20110:17:09.2234482ignored windowsincluded windowscloud syncauto-destruct piddeepgram keyapi auth2026-04-20T10:17:09.227535Zencrypt secretsretention days30 seconds3030truefalsefalseParakeetSilero/Users/lukas/.screenpipefalsetruetruetrueINFO screenpipe_core::pipes: pipe scheduler started (generation["Boosteroid"]disabled0not setenabledWARN screenpipe: pi agent install failed: bun not found - install from https://bun.shdisabled14Languagesall languagesmonitorsid: 1id: 2audio devicesdisabledyou are using local processing. all your data stays on your computer.warning: telemetry is enabled. only error-level data will be sent.to disable, use the --disable-telemetry flag....
|
NULL
|
2017569825927650571
|
NULL
|
click
|
ocr
|
NULL
|
NotionFileEditViewHistoryWindowDOCKERO ₴1power Al NotionFileEditViewHistoryWindowDOCKERO ₴1power Al by everything you've seen, said or heardopen source I runs locally | developer friendlyHelpADEV (-zsh)O $82screenpipe"APP (-zsh)• *3-zsh• ₴4100% C8Mon 20 Apr 10:18:52T81screenpipe™*5settingvalueaudio chunk durationportaudio disabledvision disabledpause on DRM contentaudio enginevad enginedata directorydebug modetelemetryuse pii removaluse all monitors2026-04-20110:17:09.2234482ignored windowsincluded windowscloud syncauto-destruct piddeepgram keyapi auth2026-04-20T10:17:09.227535Zencrypt secretsretention days30 seconds3030truefalsefalseParakeetSilero/Users/lukas/.screenpipefalsetruetruetrueINFO screenpipe_core::pipes: pipe scheduler started (generation["Boosteroid"]disabled0not setenabledWARN screenpipe: pi agent install failed: bun not found - install from https://bun.shdisabled14Languagesall languagesmonitorsid: 1id: 2audio devicesdisabledyou are using local processing. all your data stays on your computer.warning: telemetry is enabled. only error-level data will be sent.to disable, use the --disable-telemetry flag....
|
52501
|
|
52504
|
1137
|
4
|
2026-04-20T07:18:53.042077+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669533042_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
NotioncaltViewmistomWindowhelt•• 0ny.atlassian.net NotioncaltViewmistomWindowhelt•• 0ny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksI JIMINNYIY [SRD-6793] Les Mills activity typeQ Search bookmarks® For youJy-20698 handle failed field sync@ RecentSpaces / Jiminny (New)Platform Team&() Summary& Timeline E Backlog@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Trackingv a bookmarks loolbarSprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Search boardQ Spaces00001(JY-18909] (Part2) Automated rerv E Bookmarks Menu@ Ask Jiminny Reports by nikolav-vaJiminny (New)• _ Mozilla FirefoxOther Bookmarkslll Plauorm leam€ New Tab11 SE Kanbanu Userpilot | EventsI Capture TeamI Enterprise Stability I.M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations2 Confluence88 Teams"= Customise sidebarQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyIn QA v+ CreateC S* Improve Storyn GeorgievDimitrovach VS Code5 davs aqoOPEN]enlovmentsE QArint 2 Q2 +1YesloymentsEPTANCE$0100% LzMon 20 Apr 10:18:52Ask Rovo# Archived work items© PagesMore3Complete sprintGroup: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-20632A.l Panorama> Don't show internalASK ANYTHING ON ANYTHINGDeployedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - (Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedY.IV-206810.5 12 000 gTech Navl Imnrove NenendabotBot & Experiment with GithubActionsClosedJ JY-206960 =8Need QAowner user ld is nuliPLATFORM STABILITYBacklog...
|
NULL
|
7408541228819871360
|
NULL
|
click
|
ocr
|
NULL
|
NotioncaltViewmistomWindowhelt•• 0ny.atlassian.net NotioncaltViewmistomWindowhelt•• 0ny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksI JIMINNYIY [SRD-6793] Les Mills activity typeQ Search bookmarks® For youJy-20698 handle failed field sync@ RecentSpaces / Jiminny (New)Platform Team&() Summary& Timeline E Backlog@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Trackingv a bookmarks loolbarSprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Search boardQ Spaces00001(JY-18909] (Part2) Automated rerv E Bookmarks Menu@ Ask Jiminny Reports by nikolav-vaJiminny (New)• _ Mozilla FirefoxOther Bookmarkslll Plauorm leam€ New Tab11 SE Kanbanu Userpilot | EventsI Capture TeamI Enterprise Stability I.M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations2 Confluence88 Teams"= Customise sidebarQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyIn QA v+ CreateC S* Improve Storyn GeorgievDimitrovach VS Code5 davs aqoOPEN]enlovmentsE QArint 2 Q2 +1YesloymentsEPTANCE$0100% LzMon 20 Apr 10:18:52Ask Rovo# Archived work items© PagesMore3Complete sprintGroup: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-20632A.l Panorama> Don't show internalASK ANYTHING ON ANYTHINGDeployedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - (Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedY.IV-206810.5 12 000 gTech Navl Imnrove NenendabotBot & Experiment with GithubActionsClosedJ JY-206960 =8Need QAowner user ld is nuliPLATFORM STABILITYBacklog...
|
52502
|
|
52505
|
1137
|
5
|
2026-04-20T07:18:53.817957+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669533817_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
NotionVIeWmistomWindowheltny.atlassian.net/jira/so NotionVIeWmistomWindowheltny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksI JIMINNYIY [SRD-6793] Les Mills activity typeQ Search bookmarks® For you@ RecentJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Trackingv la bookmarks loolba• Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Spaces(JY-18909] (Part2) Automated rerv E Bookmarks Menu@ Ask Jiminny Reports by nikolav-vaJiminny (New)> → Mozilla FirefoxOther Bookmarkslu Plauorm leam€ New Tabil SE KanbanI Capture Teamu Userpilot | EventsI Enterprise Stability I...M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations* Confluence88 TeamsU= Customise sidebarSpaces / Jiminny (New)Platform Team&() Summary& TimelineE BacklogQ Search board0001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed_at periodCOST-EFFECTIVE AND FASTER NUDGESBacklodTJY-2048915 •.0=Investiaate and fix why exceedMAINTENANCEReady Tor DevJY-205641 80 0000 =8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216.000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-201572 .000=€Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00•=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyDescriptionWe want to allow our users to automate the execution of their AJA prompts in order to save time and have themready when they need them.Create the reports:• admins and managers should be able to automate reports based on their Panorama prompts and savedsearches• the report should be generated in a pdf - use a lightly branded one this time - f [URL_WITH_CREDENTIALS] ConIn QADetailsAssigneeReporterDevelopmentReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsosganisatioPriorityFix versionsSprintDaysNeed QA+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 builds5 davs aqoOPEN]A Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1YesloymentsEPTANCE# Archived work itemsComplete sprint100% LzMon 20 Apr 10:18:533 Ask Rovo# PagesMore 3Group: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-206321œee—A.l Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
NULL
|
297725783583553971
|
NULL
|
visual_change
|
ocr
|
NULL
|
NotionVIeWmistomWindowheltny.atlassian.net/jira/so NotionVIeWmistomWindowheltny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksI JIMINNYIY [SRD-6793] Les Mills activity typeQ Search bookmarks® For you@ RecentJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Trackingv la bookmarks loolba• Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Spaces(JY-18909] (Part2) Automated rerv E Bookmarks Menu@ Ask Jiminny Reports by nikolav-vaJiminny (New)> → Mozilla FirefoxOther Bookmarkslu Plauorm leam€ New Tabil SE KanbanI Capture Teamu Userpilot | EventsI Enterprise Stability I...M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations* Confluence88 TeamsU= Customise sidebarSpaces / Jiminny (New)Platform Team&() Summary& TimelineE BacklogQ Search board0001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed_at periodCOST-EFFECTIVE AND FASTER NUDGESBacklodTJY-2048915 •.0=Investiaate and fix why exceedMAINTENANCEReady Tor DevJY-205641 80 0000 =8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216.000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-201572 .000=€Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00•=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyDescriptionWe want to allow our users to automate the execution of their AJA prompts in order to save time and have themready when they need them.Create the reports:• admins and managers should be able to automate reports based on their Panorama prompts and savedsearches• the report should be generated in a pdf - use a lightly branded one this time - f [URL_WITH_CREDENTIALS] ConIn QADetailsAssigneeReporterDevelopmentReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsosganisatioPriorityFix versionsSprintDaysNeed QA+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 builds5 davs aqoOPEN]A Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1YesloymentsEPTANCE# Archived work itemsComplete sprint100% LzMon 20 Apr 10:18:533 Ask Rovo# PagesMore 3Group: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-206321œee—A.l Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
NULL
|
|
52506
|
1136
|
5
|
2026-04-20T07:18:55.416936+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669535416_m1.jpg...
|
iTerm2
|
DOCKER (-zsh)
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:15:14 on console
Poetry Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $
Menu
⌥1 DOCKER (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥2 PROD (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥3 EU (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥4 STAGE (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥5 QA (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥6 FE (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥7 EXT (-zsh)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DOCKER (-zsh)...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $","depth":4,"bounds":{"left":0.0,"top":0.112222224,"width":0.50069445,"height":0.8877778},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $","is_focused":true},{"role":"AXButton","text":"Menu","depth":3,"bounds":{"left":0.4861111,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥1 DOCKER (-zsh)","depth":3,"bounds":{"left":0.01875,"top":0.09,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.112222224,"width":0.4986111,"height":0.12222222},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥2 PROD (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.09,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.26222223,"width":0.4986111,"height":0.14222223},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.23944445,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥3 EU (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.24,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.43222222,"width":0.4986111,"height":0.12222222},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.40944445,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥4 STAGE (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.41,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.5822222,"width":0.4986111,"height":0.12222222},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.5594444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥5 QA (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.56,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.7322222,"width":0.4986111,"height":0.12222222},"value":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.70944446,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥6 FE (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.71,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.8622222,"width":0.4986111,"height":0.13777778},"value":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.85944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥7 EXT (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.86,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.19722222,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.2013889,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.39444444,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.3986111,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.59166664,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.59583336,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.7888889,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.79305553,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.95763886,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"DOCKER (-zsh)","depth":1,"bounds":{"left":0.4652778,"top":0.033333335,"width":0.07152778,"height":0.017777778},"role_description":"text"}]...
|
1399546062189663050
|
8197420140589008921
|
click
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:15:14 on console
Poetry Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $
Menu
⌥1 DOCKER (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥2 PROD (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥3 EU (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥4 STAGE (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥5 QA (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥6 FE (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥7 EXT (-zsh)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DOCKER (-zsh)...
|
NULL
|
|
52507
|
1137
|
6
|
2026-04-20T07:18:55.351432+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669535351_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
NotionVIeWmistomWindowheltny.atlassian.net/jira/so NotionVIeWmistomWindowheltny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksI JIMINNYIY [SRD-6793] Les Mills activity typeQ Search bookmarks® For you@ RecentJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Trackingv la bookmarks loolba• Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Spaces(JY-18909] (Part2) Automated rerv E Bookmarks Menu@ Ask Jiminny Reports by nikolav-vaJiminny (New)> → Mozilla FirefoxOther Bookmarkslu Plauorm leam€ New Tabil SE KanbanI Capture Teamu Userpilot | EventsI Enterprise Stability I...M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations* Confluence88 TeamsU= Customise sidebarSpaces / Jiminny (New)Platform Team&() Summary& TimelineE BacklogQ Search board0001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed_at periodCOST-EFFECTIVE AND FASTER NUDGESBacklodTJY-2048915 •.0=Investiaate and fix why exceedMAINTENANCEReady Tor DevJY-205641 80 0000 =8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216.000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-201572 .000=€Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00•=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyDescriptionWe want to allow our users to automate the execution of their AJA prompts in order to save time and have themready when they need them.Create the reports:• admins and managers should be able to automate reports based on their Panorama prompts and savedsearches• the report should be generated in a pdf - use a lightly branded one this time- [URL_WITH_CREDENTIALS] ContIn QADetailsAssigneeReporterDevelopmentReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsosganisatioPriorityFix versionsSprintDaysNeed QA+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 builds5 davs aqoOPEN]A Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1YesloymentsEPTANCE# Archived work itemsComplete sprint100% LzMon 20 Apr 10:18:553 Ask Rovo# PagesMore 3Group: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-206321œee—A.l Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
NULL
|
-878560113160976809
|
NULL
|
click
|
ocr
|
NULL
|
NotionVIeWmistomWindowheltny.atlassian.net/jira/so NotionVIeWmistomWindowheltny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksI JIMINNYIY [SRD-6793] Les Mills activity typeQ Search bookmarks® For you@ RecentJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Trackingv la bookmarks loolba• Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Spaces(JY-18909] (Part2) Automated rerv E Bookmarks Menu@ Ask Jiminny Reports by nikolav-vaJiminny (New)> → Mozilla FirefoxOther Bookmarkslu Plauorm leam€ New Tabil SE KanbanI Capture Teamu Userpilot | EventsI Enterprise Stability I...M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations* Confluence88 TeamsU= Customise sidebarSpaces / Jiminny (New)Platform Team&() Summary& TimelineE BacklogQ Search board0001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed_at periodCOST-EFFECTIVE AND FASTER NUDGESBacklodTJY-2048915 •.0=Investiaate and fix why exceedMAINTENANCEReady Tor DevJY-205641 80 0000 =8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216.000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-201572 .000=€Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00•=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyDescriptionWe want to allow our users to automate the execution of their AJA prompts in order to save time and have themready when they need them.Create the reports:• admins and managers should be able to automate reports based on their Panorama prompts and savedsearches• the report should be generated in a pdf - use a lightly branded one this time- [URL_WITH_CREDENTIALS] ContIn QADetailsAssigneeReporterDevelopmentReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsosganisatioPriorityFix versionsSprintDaysNeed QA+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 builds5 davs aqoOPEN]A Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1YesloymentsEPTANCE# Archived work itemsComplete sprint100% LzMon 20 Apr 10:18:553 Ask Rovo# PagesMore 3Group: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-206321œee—A.l Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
52505
|
|
52508
|
1137
|
7
|
2026-04-20T07:18:56.885779+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669536885_m2.jpg...
|
iTerm2
|
DOCKER (-zsh)
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:15:14 on console
Poetry Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $
Menu
⌥1 DOCKER (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥2 PROD (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥3 EU (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥4 STAGE (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥5 QA (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥6 FE (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥7 EXT (-zsh)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DOCKER (-zsh)...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $","depth":4,"bounds":{"left":0.27027926,"top":1.0,"width":0.23969415,"height":-0.08060658},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $","is_focused":true},{"role":"AXButton","text":"Menu","depth":3,"bounds":{"left":0.50299203,"top":1.0,"width":0.004986702,"height":-0.06424582},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥1 DOCKER (-zsh)","depth":3,"bounds":{"left":0.27925533,"top":1.0,"width":0.22207446,"height":-0.06464481},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5103058,"top":1.0,"width":0.2400266,"height":-0.08060658},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.74335104,"top":1.0,"width":0.004986702,"height":-0.06424582},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥2 PROD (-zsh)","depth":4,"bounds":{"left":0.5192819,"top":1.0,"width":0.22240691,"height":-0.06464481},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥3 EU (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥4 STAGE (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥5 QA (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥6 FE (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥7 EXT (-zsh)","depth":4,"role_description":"text"},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.27027926,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.27227393,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.36469415,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.36668882,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.45910904,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.46110374,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.55352396,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.5555186,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.64793885,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.6499335,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.7287234,"top":1.0,"width":0.01861702,"height":-0.023144484},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"DOCKER (-zsh)","depth":1,"bounds":{"left":0.49301863,"top":1.0,"width":0.034242023,"height":-0.02394259},"role_description":"text"}]...
|
1399546062189663050
|
8197420140589008921
|
click
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:15:14 on console
Poetry Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $
Menu
⌥1 DOCKER (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥2 PROD (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥3 EU (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥4 STAGE (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥5 QA (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥6 FE (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥7 EXT (-zsh)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DOCKER (-zsh)...
|
NULL
|
|
52510
|
1137
|
8
|
2026-04-20T07:18:59.915182+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669539915_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
NOLIOnIY [SRD-6793] Les Mills activity typeJY-2069 NOLIOnIY [SRD-6793] Les Mills activity typeJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tabu Userpilot | Eventsminny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksI JIMINNYQ Search bookmarks® For you@ Recentv la bookmarks loolba• Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Spacesv E Bookmarks MenuJiminny (New)> → Mozilla FirefoxOther Bookmarkslu Plauorm leamil SE KanbanI Capture TeamI Enterprise Stability I...M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations* Confluence88 TeamsU= Customise sidebarSpaces / Jiminny (New)Platform Team&() Summary& TimelineE BacklogQ Search board0001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed_at periodCOST-EFFECTIVE AND FASTER NUDGESBacklodTJY-2048915 •.0=Investiaate and fix why exceedMAINTENANCEReady Tor DevJY-205641 80 0000 =8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216.000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-201572 .000=€Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00•=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyDescriptionWe want to allow our users to automate the execution of their AJA prompts in order to save time and have themready when they need them.Create the reports:• admins and managers should be able to automate reports based on their Panorama prompts and savedsearches• the report should be generated in a pdf - use a lightly branded one this time - f [URL_WITH_CREDENTIALS] ContIn QADetailsAssigneeReporterDevelopmentReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsosganisatioPriorityFix versionsSprintDaysNeed QA+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 builds5 davs aqoOPEN]A Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1Yes>0.loymentsEPTANCE# Archived work itemsComplete sprint100% 52MOn ZU AOr 10:10:033 Ask Rovo# PagesMore 3Group: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-206321œee—A.l Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
NULL
|
-6915544251133292144
|
NULL
|
click
|
ocr
|
NULL
|
NOLIOnIY [SRD-6793] Les Mills activity typeJY-2069 NOLIOnIY [SRD-6793] Les Mills activity typeJY-20698 handle failed field sync@ JY-20692 change confirmation pa(UY-20543) AJ Reports > Tracking(JY-18909] (Part2) Automated rer@ Ask Jiminny Reports by nikolav-va€ New Tabu Userpilot | Eventsminny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-18909BookmarksI JIMINNYQ Search bookmarks® For you@ Recentv la bookmarks loolba• Sprint Board# SRD QueueGithubJiminny DEVAsk liminnv Renorts bv nikolav-vankov • Pulli....# Starred0+ AppsQ Spacesv E Bookmarks MenuJiminny (New)> → Mozilla FirefoxOther Bookmarkslu Plauorm leamil SE KanbanI Capture TeamI Enterprise Stability I...M Processina TeamService-Desk= More spaces= FiltersH DashboardsC÷ Operations* Confluence88 TeamsU= Customise sidebarSpaces / Jiminny (New)Platform Team&() Summary& TimelineE BacklogQ Search board0001READY FOR DEV 7Rework Nudaes - Phase 2 -change Nudges to use theindexed_at periodCOST-EFFECTIVE AND FASTER NUDGESBacklodTJY-2048915 •.0=Investiaate and fix why exceedMAINTENANCEReady Tor DevJY-205641 80 0000 =8Al Reports > Empty page designA.J REPORTSBacklogA JY-2037216.000=8Send email notification when thereport is not generatedAJ REPORTSBacklogO JY-201572 .000=€Notify a user before the AJ ReportexoiresAJ REPORTSBacklod0JY-205091 •00•=0Sync opportunities without a localowner user id is nuli)PLATFORM STABILITYQ Search& JY-19240 / [ JY-18909[Part2] Automated reports with Ask JiminnyDescriptionWe want to allow our users to automate the execution of their AJA prompts in order to save time and have themready when they need them.Create the reports:• admins and managers should be able to automate reports based on their Panorama prompts and savedsearches• the report should be generated in a pdf - use a lightly branded one this time - f [URL_WITH_CREDENTIALS] ContIn QADetailsAssigneeReporterDevelopmentReleasesComponentsSub-ProductlabelsStory pointestimateStory PointsosganisatioPriorityFix versionsSprintDaysNeed QA+ CreateC S* Improve StorySteliyan GeorgievAssign to meGalya Dimitrova• Onen with VS Code4 branches42 commits1 pull request5 builds5 davs aqoOPEN]A Production7 See all deplovmentsPlatformAdd ootionsiAI) (BE) (FE) QA)NoneNonel= MediumNonePlatform Sprint 2 Q2 +1Yes>0.loymentsEPTANCE# Archived work itemsComplete sprint100% 52MOn ZU AOr 10:10:033 Ask Rovo# PagesMore 3Group: QueriesDEPLOY 5Prevare tallback with email torSSO for 'persistent'name id tormatREDUCE CHURNClosedNJY-206321œee—A.l Panorama> Don't show internalerrors to customersASK ANYTHING ON ANYTHINGDeplovedProphet** JY-202781@ ••=TUpgrade Python and libraries - AprMAINTENANCEDeployedJY-199671• •=1CLONE - [Team insights] Filtergets reset automaticallySUPPORT TICKETSIDeolovedX* JY-206810.5 %2 000 g[Tech Day] Improve DependabotBot & Experiment with GithubActionsClosedJ JY-206960 =8...
|
52508
|
|
52511
|
1136
|
7
|
2026-04-20T07:19:02.123255+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669542123_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Hidden Bar••DOCKER (-zsh)APP (-zsh)DOCKER081X MDOC Hidden Bar••DOCKER (-zsh)APP (-zsh)DOCKER081X MDOCKER (-zsh)Last login: Mon Apr 20 10:15:14 on consoleDEV (-zsh)O $2Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentsPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $+*lahl100% <47• *3-zshPROD (-zsh)Last login: Mon Apr 20 10:15:14 on console• ₴4screenpipe"Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ |X 13EU (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parents@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ I|T4STAGE (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-JiminnyX T5 QA (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentsX T6 FE (-zsh)Last login: Mon Apr 20 10:16:41on ttys004Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ IX Y7 EXT (-zsh)Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentskas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ I|Mon 20 Apr 10:19:02181• *5PRODSTAGEFRONTENDEXTENSION...
|
NULL
|
-8973749310877740996
|
NULL
|
visual_change
|
ocr
|
NULL
|
Hidden Bar••DOCKER (-zsh)APP (-zsh)DOCKER081X MDOC Hidden Bar••DOCKER (-zsh)APP (-zsh)DOCKER081X MDOCKER (-zsh)Last login: Mon Apr 20 10:15:14 on consoleDEV (-zsh)O $2Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentsPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $+*lahl100% <47• *3-zshPROD (-zsh)Last login: Mon Apr 20 10:15:14 on console• ₴4screenpipe"Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ |X 13EU (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parents@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ I|T4STAGE (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-JiminnyX T5 QA (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentsX T6 FE (-zsh)Last login: Mon Apr 20 10:16:41on ttys004Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ IX Y7 EXT (-zsh)Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentskas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ I|Mon 20 Apr 10:19:02181• *5PRODSTAGEFRONTENDEXTENSION...
|
NULL
|
|
52512
|
1136
|
8
|
2026-04-20T07:19:05.096253+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669545096_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Hidden Bar+*lallDOCKER (-zsh)APP (-zsh)A• *4100% & Hidden Bar+*lallDOCKER (-zsh)APP (-zsh)A• *4100% <47Mon 20 Apr 10:19:04181• *5DOCKER081X MDOCKER (-zsh)Last login: Mon Apr 20 10:15:14 on consoleDEV (-zsh)O $82Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentsPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $• *3-zshPROD (-zsh)Last login: Mon Apr 20 10:15:14 on consolescreenpipe"Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ |X t3EU (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parents@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ I|T4STAGE (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-JiminnyX T5 QA (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentsX T6 FE (-zsh)Last login: Mon Apr 20 10:16:41on ttys004Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ IX Y7 EXT (-zsh)Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ [|PRODSTAGEFRONTENDEXTENSION...
|
NULL
|
-1390754861312897825
|
NULL
|
visual_change
|
ocr
|
NULL
|
Hidden Bar+*lallDOCKER (-zsh)APP (-zsh)A• *4100% & Hidden Bar+*lallDOCKER (-zsh)APP (-zsh)A• *4100% <47Mon 20 Apr 10:19:04181• *5DOCKER081X MDOCKER (-zsh)Last login: Mon Apr 20 10:15:14 on consoleDEV (-zsh)O $82Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentsPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $• *3-zshPROD (-zsh)Last login: Mon Apr 20 10:15:14 on consolescreenpipe"Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ |X t3EU (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parents@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ I|T4STAGE (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-JiminnyX T5 QA (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentsX T6 FE (-zsh)Last login: Mon Apr 20 10:16:41on ttys004Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ IX Y7 EXT (-zsh)Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ [|PRODSTAGEFRONTENDEXTENSION...
|
52511
|
|
52513
|
1137
|
9
|
2026-04-20T07:19:08.922063+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669548922_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
Docker DeskiopJiminny ...00 muaaiesDrafts & se Docker DeskiopJiminny ...00 muaaiesDrafts & sent® DirectoriesAh External connections.Activity# Starred& jiminny-x-integrati...8 platform-inner-team© Channels4 al-chapten# alerts# backend# confusion-clinic# curiosity_lab… engineering# frontend# general# infra-changes# jiminny-bg# platform-tickets# product_launches# random# releases# support# thank-yous# the_people_of jimi...o? Direct messagesPa Stoyan Tomov. Aneliya Angelova3 Aneliya Angelova,..P. Nikolay NikolovP. Galya Dimitrova&. Stoyan TanevVasil VasilevP. Nikolay IvanovP. Vesi: Apps• Toastf Jira Cloud& platform-inner-team• MessagesChannel OverviewRehnementsr Files& Pins• BookmarksRetro Action Itemsможе ли едно ревю на PR-a за репортите?https://github.com/jiminny/app/pull/11894Galya Dimitrova 4:38 PMсторита за рефаинмьнт за понеделникFriday, April 17th ~• https://minnv.atlassian.net/browse/Jy-200/o• https://minnv.atlassian.net/browse/Jy-20015• https://uminnv.atlassian.net/browse/Jy-20013• https://jiminny.atlassian.net/browse/JY-20500• nttos:///uminnv.atlassian.net/orowse/y-20507JY-20676 Notify the user if a Panorama prompts is deleted but is used in AJ ...Status. RackloeType: Story® Assignee: Unassigned1 Priority: MediumSync thread*AISummariseAdded by Jira CloudJY-20615 Notify the user if a SS is deleted but is used in AJ ReportStatus: BacklogI Type: StoryAssignee: UnassignedT Priority: MediumSync thread+ Al SummariseAdded by lira CloudlJY-0613 Allow owner's role to be selected when setung uo a triallStatus: BacklogType: StoryAssignee: UnassignedT Priority: MediumSync thread*+ Al SummariseAdded bv lira CloudJY-20500 Batch initial svnc for SalesforceStatus: BacklogType: Story®)Assignee: Unassisned1 Prioritv: MediumSync thread*+ Al SummariseAdded hy lim CloudJY-20501 Batch initial sync for HubspotStatus: BacklogType: StoryAssignee: Unassigned1 Prioritv: MediumSunc thread*Al SummariseAdded by Jira CloudLukas Kovalik 10.08 AMнещо ми изключи audio изцяло, края не съм чувал нищо, какво да погледна аз от тезиMessage Anlatform-inner-team+ Аa& 10o Deleted hlevith Ask Jiminny• ••ily docker.desktop PERSONALIn QA+ CreateC S* Improve StoryQ SearchStarting the Docker Engine...Docker Enaine is the underlvina technoloav that runs containersEngine startingRAM 0.00 GB CPU 0.00% Disk: -.- GB used (imit -.-. -B)delete it - deleting it will delete only this specific pdfoption to the Report type filter so users can filter the list forow who shared it in the 'Shared' column - & httos://www.fiamDroiect_Dhoenix?node-id-14360-400782t-WeZZfvQ>T||fHuXVIFix versionsSprintDaysNeed QANonePlatform Sprint 2 Q2 +1Yesloyments# Archived work itemsComplete sprintSign in›_ O Update available100% LzMon 20 Apr 10:19:08Ask Rovo© PagesGroup: QueriesEPLOY 5revare tallback with email torSO for 'persistentame id tormatJ Panorama> Don't show internalASK ANYTHING ON ANYTHINGveplovedProphet]B JY-202781@ ••=TIpgrade Python and libraries - Apreployed2 JY-199671• •=1CLONE - [Team insights) Filterets reset automaticallyUPPORT TICKETSDeolovedPY.IV-2068105"000(Tech Day] Improve DependabotBot & Experiment with GithubActionsClosed...
|
NULL
|
-2923538429991339160
|
NULL
|
visual_change
|
ocr
|
NULL
|
Docker DeskiopJiminny ...00 muaaiesDrafts & se Docker DeskiopJiminny ...00 muaaiesDrafts & sent® DirectoriesAh External connections.Activity# Starred& jiminny-x-integrati...8 platform-inner-team© Channels4 al-chapten# alerts# backend# confusion-clinic# curiosity_lab… engineering# frontend# general# infra-changes# jiminny-bg# platform-tickets# product_launches# random# releases# support# thank-yous# the_people_of jimi...o? Direct messagesPa Stoyan Tomov. Aneliya Angelova3 Aneliya Angelova,..P. Nikolay NikolovP. Galya Dimitrova&. Stoyan TanevVasil VasilevP. Nikolay IvanovP. Vesi: Apps• Toastf Jira Cloud& platform-inner-team• MessagesChannel OverviewRehnementsr Files& Pins• BookmarksRetro Action Itemsможе ли едно ревю на PR-a за репортите?https://github.com/jiminny/app/pull/11894Galya Dimitrova 4:38 PMсторита за рефаинмьнт за понеделникFriday, April 17th ~• https://minnv.atlassian.net/browse/Jy-200/o• https://minnv.atlassian.net/browse/Jy-20015• https://uminnv.atlassian.net/browse/Jy-20013• https://jiminny.atlassian.net/browse/JY-20500• nttos:///uminnv.atlassian.net/orowse/y-20507JY-20676 Notify the user if a Panorama prompts is deleted but is used in AJ ...Status. RackloeType: Story® Assignee: Unassigned1 Priority: MediumSync thread*AISummariseAdded by Jira CloudJY-20615 Notify the user if a SS is deleted but is used in AJ ReportStatus: BacklogI Type: StoryAssignee: UnassignedT Priority: MediumSync thread+ Al SummariseAdded by lira CloudlJY-0613 Allow owner's role to be selected when setung uo a triallStatus: BacklogType: StoryAssignee: UnassignedT Priority: MediumSync thread*+ Al SummariseAdded bv lira CloudJY-20500 Batch initial svnc for SalesforceStatus: BacklogType: Story®)Assignee: Unassisned1 Prioritv: MediumSync thread*+ Al SummariseAdded hy lim CloudJY-20501 Batch initial sync for HubspotStatus: BacklogType: StoryAssignee: Unassigned1 Prioritv: MediumSunc thread*Al SummariseAdded by Jira CloudLukas Kovalik 10.08 AMнещо ми изключи audio изцяло, края не съм чувал нищо, какво да погледна аз от тезиMessage Anlatform-inner-team+ Аa& 10o Deleted hlevith Ask Jiminny• ••ily docker.desktop PERSONALIn QA+ CreateC S* Improve StoryQ SearchStarting the Docker Engine...Docker Enaine is the underlvina technoloav that runs containersEngine startingRAM 0.00 GB CPU 0.00% Disk: -.- GB used (imit -.-. -B)delete it - deleting it will delete only this specific pdfoption to the Report type filter so users can filter the list forow who shared it in the 'Shared' column - & httos://www.fiamDroiect_Dhoenix?node-id-14360-400782t-WeZZfvQ>T||fHuXVIFix versionsSprintDaysNeed QANonePlatform Sprint 2 Q2 +1Yesloyments# Archived work itemsComplete sprintSign in›_ O Update available100% LzMon 20 Apr 10:19:08Ask Rovo© PagesGroup: QueriesEPLOY 5revare tallback with email torSO for 'persistentame id tormatJ Panorama> Don't show internalASK ANYTHING ON ANYTHINGveplovedProphet]B JY-202781@ ••=TIpgrade Python and libraries - Apreployed2 JY-199671• •=1CLONE - [Team insights) Filterets reset automaticallyUPPORT TICKETSDeolovedPY.IV-2068105"000(Tech Day] Improve DependabotBot & Experiment with GithubActionsClosed...
|
NULL
|
|
52516
|
1136
|
10
|
2026-04-20T07:19:12.068225+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669552068_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
iTerm2Shell|EditViewSessionScriptsProfilesWindowHe iTerm2Shell|EditViewSessionScriptsProfilesWindowHelp• 0DOCKER081DEV (-zsh)O $82X MDOCKER (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentsPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $‹$0lall100% C8Mon 20 Apr 10:19:12DOCKER (-zsh)181APP (-zsh)• &3-zsh• ₴4X 12PROD (-zsh)Last login: Mon Apr 20 10:15:14 on consolescreenpipe"• ₴5|+Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPRODPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ |X 13EU (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parents@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ I|X T4STAGE (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsSTAGEPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-JiminnyX T5 QA (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentsX T6 FE (-zsh)Last login: Mon Apr 20 10:16:41on ttys004Poetry could not find a pyproject.toml file in /Users/lukas or its parentsFRONTENDPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ IX Y7 EXT (-zsh)Poetry could not find a pyproject.toml file in /Users/lukas or its parentsEXTENSIONPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ [|...
|
NULL
|
5379777780053802160
|
NULL
|
click
|
ocr
|
NULL
|
iTerm2Shell|EditViewSessionScriptsProfilesWindowHe iTerm2Shell|EditViewSessionScriptsProfilesWindowHelp• 0DOCKER081DEV (-zsh)O $82X MDOCKER (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentsPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $‹$0lall100% C8Mon 20 Apr 10:19:12DOCKER (-zsh)181APP (-zsh)• &3-zsh• ₴4X 12PROD (-zsh)Last login: Mon Apr 20 10:15:14 on consolescreenpipe"• ₴5|+Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPRODPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ |X 13EU (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.toml file in /Users/lukas or its parents@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ I|X T4STAGE (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsSTAGEPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-JiminnyX T5 QA (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentsX T6 FE (-zsh)Last login: Mon Apr 20 10:16:41on ttys004Poetry could not find a pyproject.toml file in /Users/lukas or its parentsFRONTENDPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ IX Y7 EXT (-zsh)Poetry could not find a pyproject.toml file in /Users/lukas or its parentsEXTENSIONPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ [|...
|
52514
|
|
52517
|
1136
|
11
|
2026-04-20T07:19:16.720542+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669556720_m1.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
iTerm2ShelllEditViewSessionScriptsProfilesWindowHe iTerm2ShelllEditViewSessionScriptsProfilesWindowHelpDOCKER081DEV (-zsh)O $82X MDOCKER (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentsPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $lall100%8Mon 20 Apr 10:19:16DOCKER (-zsh)181APP (-zsh)• &3-zsh• 84screenpipe"• *5+X 12PROD (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPRODPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ |X L3 EU (-zsh)Last loain: Mon Apr20 10:15:14 on consolefile in /Users/lukas or its parentsfile in /Users/lukas or its parents~$IPhpStorm2026.1olefile in /Users/lukas or its parentsfilein /Users/lukas or its parentsSTAGEO JETBRAINS IDEsolefile in /Users/lukas or its parentsfilein /Users/lukas or its parents$Last login: Mon Apr 20 10:16:41on ttys004Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ IX Y7 EXT (-zsh)Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ [|FRONTENDEXTENSION...
|
NULL
|
5395977789359968414
|
NULL
|
click
|
ocr
|
NULL
|
iTerm2ShelllEditViewSessionScriptsProfilesWindowHe iTerm2ShelllEditViewSessionScriptsProfilesWindowHelpDOCKER081DEV (-zsh)O $82X MDOCKER (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentsPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $lall100%8Mon 20 Apr 10:19:16DOCKER (-zsh)181APP (-zsh)• &3-zsh• 84screenpipe"• *5+X 12PROD (-zsh)Last login: Mon Apr 20 10:15:14 on consolePoetry could not find a pyproject.toml file in /Users/lukas or its parentsPRODPoetry could not find a pyproject.toml file in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ |X L3 EU (-zsh)Last loain: Mon Apr20 10:15:14 on consolefile in /Users/lukas or its parentsfile in /Users/lukas or its parents~$IPhpStorm2026.1olefile in /Users/lukas or its parentsfilein /Users/lukas or its parentsSTAGEO JETBRAINS IDEsolefile in /Users/lukas or its parentsfilein /Users/lukas or its parents$Last login: Mon Apr 20 10:16:41on ttys004Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ IX Y7 EXT (-zsh)Poetry could not find a pyproject.toml file in /Users/lukas or its parentsPoetry could not find a pyproject.tomlfile in /Users/lukas or its parentslukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $ [|FRONTENDEXTENSION...
|
NULL
|
|
52518
|
1137
|
11
|
2026-04-20T07:19:16.622747+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669556622_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
PhpStormActivityJiminny ...od huaalesDrafts & PhpStormActivityJiminny ...od huaalesDrafts & sent® DirectoriesAb External connections* Starred8 jiminny-x-integrati...• platform-inner-team© Channelsac al-chanter# alerts# backend# confusion-clinic# curiosity_lab# engineering# frontend# general# infra-changes# jiminny-bgi nlatform-nickets# product_launchesac random# releases# support# thank-yous# the_people_of_jimi..•? Direct messagesRP. Stoyan Tomovf. Aneliya Angelova3 Aneliya Angelova,...P. Nikolay Nikolov. Galya Dimitrova2. Stoyan TanevVasil Vasilev.. Nikolay Ivanove Ves#: Apps9 ToastS Jira Cloud& platform-inner-team• MessagesChannel OverviewRehnementsr Files& Pins• BookmarksRetro Action Itemsможе ли едно ревю на PR-a за репортите?https://github.com/jiminny/app/pull/11894Galya Dimitrova 4:38 PMсторита за рефаинмьнт за понеделникFriday. April 17thv• https://minnv.atlassian.net/browse/Jy-200/o• https://minnv.atlassian.net/browse/Jy-20015• https://uminnv.atlassian.net/browse/Jy-20013• https://jiminny.atlassian.net/browse/JY-20500• nttos:///uminnv.atlassian.net/orowse/y-20501JY-20676 Notify the user if a Panorama prompts is deleted but is used in AJ ...Status. RackloeType: Story® Assignee: Unassigned1 Priority: MediumSync thread*AISummariseAdded by Jira CloudJY-20615 Notify the user if a SS is deleted but is used in AJ ReportStatus: BacklogI Type: StoryAssignee: UnassignedT Priority: MediumSync thread+ Al SummariseAdded by lira CloudhJY-0613 Allow owner's role to be selected when setung uo a triallStatus: BacklogType: StoryAssignee: UnassignedT Priority: MediumSync thread*+ Al SummariseAdded bv lira CloudJY-20500 Batch initial svnc for Salesforce.Status: BacklogType: Story®)Assignee: Unassisned1 Prioritv: MediumSync thread*+ Al SummariseAdded hy lim Cloud1Y.20501 Ratch initial svnc for HubsnotStatus: BacklogType: StoryAssignee: Unassigned1 Prioritv: MediumSunc thread*Al SummariseAdded by Jira CloudLukas Kovallk 10-08 AMнещо ми изключи audio изцяло, края не съм чувал нищо, какво да погледна аз от тезиMessage & platform-inner-team+ Аa& 10o Deleted hlevith Ask JiminnyIn QAild docker.desktop PERSONALAsk Gordon BETAContainers Give feedbackContainersImagesContainer CPU usageVolumesNo containers are runnina.KubernetesSearchBuildsContainer IDModelsMCP Toolkit BETA1229ffe7ed37Docker HubkibanaД А 7с3ес7911304Docker ScoutngrokAAN 00a86edb2-8dEytencionehlacktire.121A65207054Manaaejiminny_ext-1 587546c8d3e0elasticsearch e802ad473a4f1- Resource usagedaladod-l0727543fa332mariado-b7e64079c3c3RAM 0.61 GB CPU 4.35%Disk: -.- GB used (limit -.- GB)delete it - deleting it will delete only this specific pdfFix versionsontion to the Revort tvoe tilter so users can tilter the list torSprintow who shared it in the 'Shared' column - & httos://www.fiamDroiect_Dhoenix?node-id-14360-400782t-WeZZfvQ>T||fHuXVIDaysNeed QA+ Create* Improve StoryQ SearchContainer memory usageNo containers are runnina.Only show running containersImagePort(s)reals.o6379:6379kibana/kibana:7.10.2 5601:5601werniaht/narok4040:4040hlacktire/blacktire'1: 8707-8707noceo. lrealpheeaskcseacmeasile0200•0200|Show all ports )oaladoglagento.lCPU(% ActionsN/A DN/ADN/AN/AN/A DN/A DN/A Dmariado:11.4.5You are signed outSian in to share images andcollaborate with your teamNonePlatform Sprint 2 Q2 +1Yes100% LzMon ZUAO 10.19.103 Ask Rovoloyments# Archived work items© PagesMore 3Show chartsSign inGroup: QueriesEPLOY 5revare tallback with email torSO for 'persistentame id tormatJY-206321œee—J Panorama> Don't show internalASK ANYTHING ON ANYTHINGveplovedProphet]B JY-202781@ ••=TIpgrade Python and libraries - Apreployed2 JY-199671@eo —CLONE - [Team insights) Filterets reset automaticallyUPPORT TICKETSleolovedP.IV-2068105"000›_ O Update availableTech Navl Imnrove NenendahotActionsClosed#JY-20696...
|
NULL
|
-1994999543308681824
|
NULL
|
click
|
ocr
|
NULL
|
PhpStormActivityJiminny ...od huaalesDrafts & PhpStormActivityJiminny ...od huaalesDrafts & sent® DirectoriesAb External connections* Starred8 jiminny-x-integrati...• platform-inner-team© Channelsac al-chanter# alerts# backend# confusion-clinic# curiosity_lab# engineering# frontend# general# infra-changes# jiminny-bgi nlatform-nickets# product_launchesac random# releases# support# thank-yous# the_people_of_jimi..•? Direct messagesRP. Stoyan Tomovf. Aneliya Angelova3 Aneliya Angelova,...P. Nikolay Nikolov. Galya Dimitrova2. Stoyan TanevVasil Vasilev.. Nikolay Ivanove Ves#: Apps9 ToastS Jira Cloud& platform-inner-team• MessagesChannel OverviewRehnementsr Files& Pins• BookmarksRetro Action Itemsможе ли едно ревю на PR-a за репортите?https://github.com/jiminny/app/pull/11894Galya Dimitrova 4:38 PMсторита за рефаинмьнт за понеделникFriday. April 17thv• https://minnv.atlassian.net/browse/Jy-200/o• https://minnv.atlassian.net/browse/Jy-20015• https://uminnv.atlassian.net/browse/Jy-20013• https://jiminny.atlassian.net/browse/JY-20500• nttos:///uminnv.atlassian.net/orowse/y-20501JY-20676 Notify the user if a Panorama prompts is deleted but is used in AJ ...Status. RackloeType: Story® Assignee: Unassigned1 Priority: MediumSync thread*AISummariseAdded by Jira CloudJY-20615 Notify the user if a SS is deleted but is used in AJ ReportStatus: BacklogI Type: StoryAssignee: UnassignedT Priority: MediumSync thread+ Al SummariseAdded by lira CloudhJY-0613 Allow owner's role to be selected when setung uo a triallStatus: BacklogType: StoryAssignee: UnassignedT Priority: MediumSync thread*+ Al SummariseAdded bv lira CloudJY-20500 Batch initial svnc for Salesforce.Status: BacklogType: Story®)Assignee: Unassisned1 Prioritv: MediumSync thread*+ Al SummariseAdded hy lim Cloud1Y.20501 Ratch initial svnc for HubsnotStatus: BacklogType: StoryAssignee: Unassigned1 Prioritv: MediumSunc thread*Al SummariseAdded by Jira CloudLukas Kovallk 10-08 AMнещо ми изключи audio изцяло, края не съм чувал нищо, какво да погледна аз от тезиMessage & platform-inner-team+ Аa& 10o Deleted hlevith Ask JiminnyIn QAild docker.desktop PERSONALAsk Gordon BETAContainers Give feedbackContainersImagesContainer CPU usageVolumesNo containers are runnina.KubernetesSearchBuildsContainer IDModelsMCP Toolkit BETA1229ffe7ed37Docker HubkibanaД А 7с3ес7911304Docker ScoutngrokAAN 00a86edb2-8dEytencionehlacktire.121A65207054Manaaejiminny_ext-1 587546c8d3e0elasticsearch e802ad473a4f1- Resource usagedaladod-l0727543fa332mariado-b7e64079c3c3RAM 0.61 GB CPU 4.35%Disk: -.- GB used (limit -.- GB)delete it - deleting it will delete only this specific pdfFix versionsontion to the Revort tvoe tilter so users can tilter the list torSprintow who shared it in the 'Shared' column - & httos://www.fiamDroiect_Dhoenix?node-id-14360-400782t-WeZZfvQ>T||fHuXVIDaysNeed QA+ Create* Improve StoryQ SearchContainer memory usageNo containers are runnina.Only show running containersImagePort(s)reals.o6379:6379kibana/kibana:7.10.2 5601:5601werniaht/narok4040:4040hlacktire/blacktire'1: 8707-8707noceo. lrealpheeaskcseacmeasile0200•0200|Show all ports )oaladoglagento.lCPU(% ActionsN/A DN/ADN/AN/AN/A DN/A DN/A Dmariado:11.4.5You are signed outSian in to share images andcollaborate with your teamNonePlatform Sprint 2 Q2 +1Yes100% LzMon ZUAO 10.19.103 Ask Rovoloyments# Archived work items© PagesMore 3Show chartsSign inGroup: QueriesEPLOY 5revare tallback with email torSO for 'persistentame id tormatJY-206321œee—J Panorama> Don't show internalASK ANYTHING ON ANYTHINGveplovedProphet]B JY-202781@ ••=TIpgrade Python and libraries - Apreployed2 JY-199671@eo —CLONE - [Team insights) Filterets reset automaticallyUPPORT TICKETSleolovedP.IV-2068105"000›_ O Update availableTech Navl Imnrove NenendahotActionsClosed#JY-20696...
|
NULL
|
|
52519
|
1137
|
12
|
2026-04-20T07:19:18.303384+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669558303_m2.jpg...
|
iTerm2
|
NULL
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
Docker DesktopView100% LzMon ZU AOr 10.19.10ii doc Docker DesktopView100% LzMon ZU AOr 10.19.10ii dockerdesktoo PERSONAL• SearchAsk GordonBETAContainersGive feedbackVolumesKuberneresbulldsNo containers are runninaSearchNo containers are runninaOnlv show runnina containersNameContainer DImagePort(s)CPU(% Actions1220ffe7ed276270-6270easeNoskor LnbDocker ScoutFxtensionsManage1 Resource usagehlacktire.fiminny_ext-1 587546c8dBe09200:9200show alllpors AnatrtlainethZokAnzOA2A?calacog lageno.lamariadb:11.4.5collaborate with your team...
|
NULL
|
8110645486230639204
|
NULL
|
visual_change
|
ocr
|
NULL
|
Docker DesktopView100% LzMon ZU AOr 10.19.10ii doc Docker DesktopView100% LzMon ZU AOr 10.19.10ii dockerdesktoo PERSONAL• SearchAsk GordonBETAContainersGive feedbackVolumesKuberneresbulldsNo containers are runninaSearchNo containers are runninaOnlv show runnina containersNameContainer DImagePort(s)CPU(% Actions1220ffe7ed276270-6270easeNoskor LnbDocker ScoutFxtensionsManage1 Resource usagehlacktire.fiminny_ext-1 587546c8dBe09200:9200show alllpors AnatrtlainethZokAnzOA2A?calacog lageno.lamariadb:11.4.5collaborate with your team...
|
52518
|
|
52533
|
1136
|
17
|
2026-04-20T07:21:21.605026+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669681605_m1.jpg...
|
iTerm2
|
DOCKER (-zsh)
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:15:14 on console
Poetry Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $
Menu
⌥1 DOCKER (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥2 PROD (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥3 EU (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥4 STAGE (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥5 QA (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥6 FE (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥7 EXT (-zsh)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DOCKER (-zsh)...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $","depth":4,"bounds":{"left":0.0,"top":0.112222224,"width":0.50069445,"height":0.8877778},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $","is_focused":true},{"role":"AXButton","text":"Menu","depth":3,"bounds":{"left":0.4861111,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥1 DOCKER (-zsh)","depth":3,"bounds":{"left":0.01875,"top":0.09,"width":0.46388888,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.112222224,"width":0.4986111,"height":0.12222222},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.08944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥2 PROD (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.09,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.26222223,"width":0.4986111,"height":0.14222223},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.23944445,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥3 EU (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.24,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.43222222,"width":0.4986111,"height":0.12222222},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.40944445,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥4 STAGE (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.41,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.5822222,"width":0.4986111,"height":0.12222222},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.5594444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥5 QA (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.56,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.7322222,"width":0.4986111,"height":0.12222222},"value":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.70944446,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥6 FE (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.71,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5013889,"top":0.8622222,"width":0.4986111,"height":0.13777778},"value":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.98819447,"top":0.85944444,"width":0.010416667,"height":0.016666668},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥7 EXT (-zsh)","depth":4,"bounds":{"left":0.52013886,"top":0.86,"width":0.46458334,"height":0.015555556},"role_description":"text"},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.19722222,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.2013889,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.39444444,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.3986111,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.59166664,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.59583336,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.7888889,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.79305553,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.95763886,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"DOCKER (-zsh)","depth":1,"bounds":{"left":0.4652778,"top":0.033333335,"width":0.07152778,"height":0.017777778},"role_description":"text"}]...
|
1399546062189663050
|
8197420140589008921
|
click
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:15:14 on console
Poetry Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $
Menu
⌥1 DOCKER (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥2 PROD (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥3 EU (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥4 STAGE (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥5 QA (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥6 FE (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥7 EXT (-zsh)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DOCKER (-zsh)...
|
52529
|
|
52534
|
1137
|
21
|
2026-04-20T07:21:21.436780+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669681436_m2.jpg...
|
iTerm2
|
DOCKER (-zsh)
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:15:14 on console
Poetry Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $
Menu
⌥1 DOCKER (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥2 PROD (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥3 EU (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥4 STAGE (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥5 QA (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥6 FE (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥7 EXT (-zsh)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DOCKER (-zsh)...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $","depth":4,"bounds":{"left":0.27027926,"top":1.0,"width":0.23969415,"height":-0.08060658},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $","is_focused":true},{"role":"AXButton","text":"Menu","depth":3,"bounds":{"left":0.50299203,"top":1.0,"width":0.004986702,"height":-0.06424582},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥1 DOCKER (-zsh)","depth":3,"bounds":{"left":0.27925533,"top":1.0,"width":0.22207446,"height":-0.06464481},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"bounds":{"left":0.5103058,"top":1.0,"width":0.2400266,"height":-0.08060658},"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"bounds":{"left":0.74335104,"top":1.0,"width":0.004986702,"height":-0.06424582},"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥2 PROD (-zsh)","depth":4,"bounds":{"left":0.5192819,"top":1.0,"width":0.22240691,"height":-0.06464481},"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥3 EU (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥4 STAGE (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Mon Apr 20 10:15:14 on console\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥5 QA (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥6 FE (-zsh)","depth":4,"role_description":"text"},{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","depth":5,"value":"Last login: Mon Apr 20 10:16:41 on ttys004\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $","is_focused":true},{"role":"AXButton","text":"Menu","depth":4,"role_description":"button","is_enabled":true,"is_focused":false},{"role":"AXStaticText","text":"⌥7 EXT (-zsh)","depth":4,"role_description":"text"},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.27027926,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.27227393,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.36469415,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.36668882,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.45910904,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.46110374,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.55352396,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.5555186,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.64793885,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.6499335,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.7287234,"top":1.0,"width":0.01861702,"height":-0.023144484},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"DOCKER (-zsh)","depth":1,"bounds":{"left":0.49301863,"top":1.0,"width":0.034242023,"height":-0.02394259},"role_description":"text"}]...
|
1399546062189663050
|
8197420140589008921
|
click
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:15:14 on console
Poetry Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/infrastructure/dev/docker or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $
Menu
⌥1 DOCKER (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥2 PROD (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥3 EU (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥4 STAGE (-zsh)
Last login: Mon Apr 20 10:15:14 on console
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥5 QA (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥6 FE (-zsh)
Last login: Mon Apr 20 10:16:41 on ttys004
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
Poetry could not find a pyproject.toml file in /Users/lukas or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~ $
Menu
⌥7 EXT (-zsh)
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DOCKER (-zsh)...
|
52526
|
|
52629
|
1138
|
17
|
2026-04-20T07:25:42.104382+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669942104_m1.jpg...
|
iTerm2
|
APP (-zsh)
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys007
Poetry Last login: Mon Apr 20 10:16:41 on ttys007
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
APP (-zsh)...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys007\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","depth":4,"bounds":{"left":0.0,"top":0.08777778,"width":1.0,"height":0.9122222},"value":"Last login: Mon Apr 20 10:16:41 on ttys007\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.19722222,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.2013889,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.39444444,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.3986111,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.59166664,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.59583336,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.7888889,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.79305553,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.95763886,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"APP (-zsh)","depth":1,"bounds":{"left":0.47569445,"top":0.033333335,"width":0.05138889,"height":0.017777778},"role_description":"text"}]...
|
-2484984247063990180
|
6234070588200504339
|
click
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys007
Poetry Last login: Mon Apr 20 10:16:41 on ttys007
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
APP (-zsh)...
|
NULL
|
|
52630
|
1139
|
25
|
2026-04-20T07:25:42.200664+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669942200_m2.jpg...
|
iTerm2
|
APP (-zsh)
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys007
Poetry Last login: Mon Apr 20 10:16:41 on ttys007
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
APP (-zsh)...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys007\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","depth":4,"bounds":{"left":0.27027926,"top":1.0,"width":0.4800532,"height":-0.06304872},"value":"Last login: Mon Apr 20 10:16:41 on ttys007\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.27027926,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.27227393,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.36469415,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.36668882,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.45910904,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.46110374,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.55352396,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.5555186,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.64793885,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.6499335,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.7287234,"top":1.0,"width":0.01861702,"height":-0.023144484},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"APP (-zsh)","depth":1,"bounds":{"left":0.49800533,"top":1.0,"width":0.024601065,"height":-0.02394259},"role_description":"text"}]...
|
-2484984247063990180
|
6234070588200504339
|
click
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys007
Poetry Last login: Mon Apr 20 10:16:41 on ttys007
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
APP (-zsh)...
|
NULL
|
|
52631
|
1138
|
18
|
2026-04-20T07:25:42.874393+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669942874_m1.jpg...
|
iTerm2
|
APP (-zsh)
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys007
Poetry Last login: Mon Apr 20 10:16:41 on ttys007
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
APP (-zsh)...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys007\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","depth":4,"bounds":{"left":0.0,"top":0.08777778,"width":1.0,"height":0.9122222},"value":"Last login: Mon Apr 20 10:16:41 on ttys007\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.19722222,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.2013889,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.39444444,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.3986111,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.59166664,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.59583336,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.7888889,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.79305553,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.95763886,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"APP (-zsh)","depth":1,"bounds":{"left":0.47569445,"top":0.033333335,"width":0.05138889,"height":0.017777778},"role_description":"text"}]...
|
-2484984247063990180
|
6234070588200504339
|
click
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys007
Poetry Last login: Mon Apr 20 10:16:41 on ttys007
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
APP (-zsh)...
|
NULL
|
|
52632
|
1139
|
26
|
2026-04-20T07:25:42.874411+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669942874_m2.jpg...
|
iTerm2
|
APP (-zsh)
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys007
Poetry Last login: Mon Apr 20 10:16:41 on ttys007
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
APP (-zsh)...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys007\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","depth":4,"bounds":{"left":0.27027926,"top":1.0,"width":0.4800532,"height":-0.06304872},"value":"Last login: Mon Apr 20 10:16:41 on ttys007\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.27027926,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.27227393,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.36469415,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.36668882,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.45910904,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.46110374,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.55352396,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.5555186,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.64793885,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.6499335,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.7287234,"top":1.0,"width":0.01861702,"height":-0.023144484},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"APP (-zsh)","depth":1,"bounds":{"left":0.49800533,"top":1.0,"width":0.024601065,"height":-0.02394259},"role_description":"text"}]...
|
-2484984247063990180
|
6234070588200504339
|
click
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys007
Poetry Last login: Mon Apr 20 10:16:41 on ttys007
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
APP (-zsh)...
|
NULL
|
|
52633
|
1138
|
19
|
2026-04-20T07:25:50.929349+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669950929_m1.jpg...
|
iTerm2
|
DEV (-zsh)
|
1
|
NULL
|
monitor_1
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys006
Poetry Last login: Mon Apr 20 10:16:41 on ttys006
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DEV (-zsh)...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys006\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","depth":4,"bounds":{"left":0.0,"top":0.08777778,"width":1.0,"height":0.9122222},"value":"Last login: Mon Apr 20 10:16:41 on ttys006\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.0,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.004166667,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.19722222,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.2013889,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.39444444,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.3986111,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.59166664,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.59583336,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.7888889,"top":0.05888889,"width":0.19722222,"height":0.026666667},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.79305553,"top":0.06333333,"width":0.011111111,"height":0.017777778},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.95763886,"top":0.032222223,"width":0.03888889,"height":0.018888889},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"DEV (-zsh)","depth":1,"bounds":{"left":0.47569445,"top":0.033333335,"width":0.052083332,"height":0.017777778},"role_description":"text"}]...
|
5817734767217806
|
-3851742976050575343
|
click
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys006
Poetry Last login: Mon Apr 20 10:16:41 on ttys006
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DEV (-zsh)...
|
52631
|
|
52634
|
1139
|
27
|
2026-04-20T07:25:50.929237+00:00
|
/Users/lukas/.screenpipe/data/data/2026-04-20/1776 /Users/lukas/.screenpipe/data/data/2026-04-20/1776669950929_m2.jpg...
|
iTerm2
|
DEV (-zsh)
|
1
|
NULL
|
monitor_2
|
NULL
|
NULL
|
NULL
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys006
Poetry Last login: Mon Apr 20 10:16:41 on ttys006
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DEV (-zsh)...
|
[{"role":"AXTextArea","text [{"role":"AXTextArea","text":"Last login: Mon Apr 20 10:16:41 on ttys006\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","depth":4,"bounds":{"left":0.27027926,"top":1.0,"width":0.4800532,"height":-0.06304872},"value":"Last login: Mon Apr 20 10:16:41 on ttys006\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\n\nPoetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents\nlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $","is_focused":true},{"role":"AXRadioButton","text":"DOCKER","depth":2,"bounds":{"left":0.27027926,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.27227393,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"DEV (-zsh)","depth":2,"bounds":{"left":0.36469415,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.36668882,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"APP (-zsh)","depth":2,"bounds":{"left":0.45910904,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.46110374,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"-zsh","depth":2,"bounds":{"left":0.55352396,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.5555186,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXRadioButton","text":"screenpipe\"","depth":2,"bounds":{"left":0.64793885,"top":1.0,"width":0.0944149,"height":-0.042298436},"role_description":"radio button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXButton","text":"Close Tab","depth":3,"bounds":{"left":0.6499335,"top":1.0,"width":0.005319149,"height":-0.04549086},"role_description":"button","is_enabled":false,"is_focused":false,"is_selected":false,"is_expanded":false},{"role":"AXStaticText","text":"⌥⌘1","depth":1,"bounds":{"left":0.7287234,"top":1.0,"width":0.01861702,"height":-0.023144484},"automation_id":"_NS:8","role_description":"text"},{"role":"AXStaticText","text":"DEV (-zsh)","depth":1,"bounds":{"left":0.49800533,"top":1.0,"width":0.024933511,"height":-0.02394259},"role_description":"text"}]...
|
5817734767217806
|
-3851742976050575343
|
click
|
accessibility
|
NULL
|
Last login: Mon Apr 20 10:16:41 on ttys006
Poetry Last login: Mon Apr 20 10:16:41 on ttys006
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
Poetry could not find a pyproject.toml file in /Users/lukas/jiminny/app or its parents
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app (JY-18909-automated-reports-ask-jiminny) $
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
-zsh
Close Tab
screenpipe"
Close Tab
⌥⌘1
DEV (-zsh)...
|
52632
|