|
13320
|
kibana | {"type":"log", kibana | {"type":"log","@timestamp":"2026-04-14T12:16:13Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:17Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:17Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:17Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:18Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:19Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:19Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:20Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:20Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:21Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:16:10 Running ['artisan' meeting-bot:schedule-bot]
docker_lamp_1 | Elastica\Exception\Connection\HttpException
docker_lamp_1 |
docker_lamp_1 | Couldn't resolve host
docker_lamp_1 |
docker_lamp_1 | at vendor/ruflin/elastica/src/Transport/Http.php:186
docker_lamp_1 | 182▕ throw new PartialShardFailureException($request, $response);
docker_lamp_1 | 183▕ }
docker_lamp_1 | 184▕
docker_lamp_1 | 185▕ if ($errorNumber > 0) {
docker_lamp_1 | ➜ 186▕ throw new HttpException($errorNumber, $request, $response);
docker_lamp_1 | 187▕ }
docker_lamp_1 | 188▕
docker_lamp_1 | 189▕ return $response;
docker_lamp_1 | 190▕ }
docker_lamp_1 |
docker_lamp_1 | 1 vendor/ruflin/elastica/src/Request.php:183
docker_lamp_1 | Elastica\Transport\Http::exec(Object(Elastica\Request), [])
docker_lamp_1 |
docker_lamp_1 | 2 vendor/ruflin/elastica/src/Client.php:545
docker_lamp_1 | Elastica\Request::send()
docker_lamp_1 |
docker_lamp_1 | 3 vendor/ruflin/elastica/src/Client.php:580
docker_lamp_1 | Elastica\Client::request("activities/_doc/408704", "POST", [])
docker_lamp_1 |
docker_lamp_1 | 4 vendor/ruflin/elastica/src/Index.php:774
docker_lamp_1 | Elastica\Client::requestEndpoint(Object(Elasticsearch\Endpoints\Index))
docker_lamp_1 |
docker_lamp_1 | 5 vendor/ruflin/elastica/src/Index.php:241
docker_lamp_1 | Elastica\Index::requestEndpoint(Object(Elasticsearch\Endpoints\Index))
docker_lamp_1 |
docker_lamp_1 | 6 app/Component/ES/Actions/InstantUpsertAction.php:52
docker_lamp_1 | Elastica\Index::addDocument(Object(Elastica\Document))
docker_lamp_1 |
docker_lamp_1 | 7 app/Component/Datadog/Client.php:20
docker_lamp_1 | Jiminny\Component\ES\Actions\InstantUpsertAction::Jiminny\Component\ES\Actions\{closure}()
docker_lamp_1 |
docker_lamp_1 | 8 app/Component/ES/Actions/InstantUpsertAction.php:51
docker_lamp_1 | Jiminny\Component\Datadog\Client::withTiming(Object(Closure), "jiminny.elasticsearch.refresh.waiting_time")
docker_lamp_1 |
docker_lamp_1 | 9 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:36
docker_lamp_1 | Jiminny\Component\ES\Actions\InstantUpsertAction::handle(Object(Illuminate\Log\LogManager), Object(Jiminny\Component\Datadog\Client), Object(Jiminny\Component\ElasticSearch\Client))
docker_lamp_1 |
docker_lamp_1 | 10 vendor/laravel/framework/src/Illuminate/Container/Util.php:43
docker_lamp_1 | Illuminate\Container\BoundMethod::Illuminate\Container\{closure}()
docker_lamp_1 |
docker_lamp_1 | 11 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:96
docker_lamp_1 | Illuminate\Container\Util::unwrapIfClosure(Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 12 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:35
docker_lamp_1 | Illuminate\Container\BoundMethod::callBoundMethod(Object(Illuminate\Foundation\Application), Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 13 vendor/laravel/framework/src/Illuminate/Container/Container.php:799
docker_lamp_1 | Illuminate\Container\BoundMethod::call(Object(Illuminate\Foundation\Application), [])
docker_lamp_1 |
docker_lamp_1 | 14 vendor/laravel/framework/src/Illuminate/Bus/Dispatcher.php:129
docker_lamp_1 | Illuminate\Container\Container::call()
docker_lamp_1 |
docker_lamp_1 | 15 vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php:180
docker_lamp_1 | Illuminate\Bus\Dispatcher::Illuminate\Bus\{closure}(Object(Jiminny\Component\ES\Actions\InstantUpsertAction))
docker_lamp_1 |
docker_lamp_1 | 16 vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php:137
docker_lamp_1 | Illuminate\Pipeline\Pipeline::Illuminate\Pipeline\{closure}(Object(Jiminny\Component\ES\Actions\InstantUpsertAction))
docker_lamp_1 |
docker_lamp_1 | 17 vendor/laravel/framework/src/Illuminate/Bus/Dispatcher.php:133
docker_lamp_1 | Illuminate\Pipeline\Pipeline::then(Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 18 vendor/laravel/framework/src/Illuminate/Bus/Dispatcher.php:81
docker_lamp_1 | Illuminate\Bus\Dispatcher::dispatchNow(Object(Jiminny\Component\ES\Actions\InstantUpsertAction))
docker_lamp_1 |
docker_lamp_1 | 19 app/Component/ES/Listeners/UpdateSingleTargetListener.php:52
docker_lamp_1 | Illuminate\Bus\Dispatcher::dispatch(Object(Jiminny\Component\ES\Actions\InstantUpsertAction))
docker_lamp_1 |
docker_lamp_1 | 20 app/Component/ES/Listeners/UpdateSingleTargetListener.php:34
docker_lamp_1 | Jiminny\Component\ES\Listeners\UpdateSingleTargetListener::handleSyncEvent(Object(Jiminny\Contracts\ES\Events\UpdateSingleEntity))
docker_lamp_1 |
docker_lamp_1 | 21 vendor/laravel/framework/src/Illuminate/Events/Dispatcher.php:512
docker_lamp_1 | Jiminny\Component\ES\Listeners\UpdateSingleTargetListener::handle(Object(Jiminny\Contracts\ES\Events\UpdateSingleEntity))
docker_lamp_1 |
docker_lamp_1 | 22 vendor/laravel/framework/src/Illuminate/Events/Dispatcher.php:319
docker_lamp_1 | Illuminate\Events\Dispatcher::Illuminate\Events\{closure}("Jiminny\Contracts\ES\Events\UpdateSingleEntity")
docker_lamp_1 |
docker_lamp_1 | 23 vendor/laravel/framework/src/Illuminate/Events/Dispatcher.php:299
docker_lamp_1 | Illuminate\Events\Dispatcher::invokeListeners("Jiminny\Contracts\ES\Events\UpdateSingleEntity")
docker_lamp_1 |
docker_lamp_1 | 24 vendor/laravel/framework/src/Illuminate/Foundation/helpers.php:505
docker_lamp_1 | Illuminate\Events\Dispatcher::dispatch(Object(Jiminny\Contracts\ES\Events\UpdateSingleEntity))
docker_lamp_1 | 25 app/Component/ElasticSearch/Model/Observer.php:71
docker_lamp_1 |
docker_lamp_1 | 26 vendor/laravel/framework/src/Illuminate/Events/Dispatcher.php:512
docker_lamp_1 | Jiminny\Component\ElasticSearch\Model\Observer::saved(Object(Jiminny\Models\Activity))
docker_lamp_1 |
docker_lamp_1 | 27 vendor/laravel/framework/src/Illuminate/Events/Dispatcher.php:319
docker_lamp_1 | Illuminate\Events\Dispatcher::Illuminate\Events\{closure}("eloquent.saved: Jiminny\Models\Activity")
docker_lamp_1 |
docker_lamp_1 | 28 vendor/laravel/framework/src/Illuminate/Events/Dispatcher.php:299
docker_lamp_1 | Illuminate\Events\Dispatcher::invokeListeners("eloquent.saved: Jiminny\Models\Activity")
docker_lamp_1 |
docker_lamp_1 | 29 vendor/laravel/framework/src/Illuminate/Database/Eloquent/Concerns/HasEvents.php:224
docker_lamp_1 | Illuminate\Events\Dispatcher::dispatch("eloquent.saved: Jiminny\Models\Activity", Object(Jiminny\Models\Activity))
docker_lamp_1 |
docker_lamp_1 | 30 vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php:1279
docker_lamp_1 | Illuminate\Database\Eloquent\Model::fireModelEvent("saved")
docker_lamp_1 |
docker_lamp_1 | 31 vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php:1252
docker_lamp_1 | Illuminate\Database\Eloquent\Model::finishSave([])
docker_lamp_1 |
docker_lamp_1 | 32 vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php:1090
docker_lamp_1 | Illuminate\Database\Eloquent\Model::save([])
docker_lamp_1 |
docker_lamp_1 | 33 app/Services/RecallAI/Commands/ScheduleBotCommand.php:94
docker_lamp_1 | Illuminate\Database\Eloquent\Model::update(["failed"])
docker_lamp_1 |
docker_lamp_1 | 34 app/Services/RecallAI/Commands/ScheduleBotCommand.php:39
docker_lamp_1 | Jiminny\Services\RecallAI\Commands\ScheduleBotCommand::scheduleBotsForUpcomingConferences()
docker_lamp_1 |
docker_lamp_1 | 35 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:36
docker_lamp_1 | Jiminny\Services\RecallAI\Commands\ScheduleBotCommand::handle()
docker_lamp_1 |
docker_lamp_1 | 36 vendor/laravel/framework/src/Illuminate/Container/Util.php:43
docker_lamp_1 | Illuminate\Container\BoundMethod::Illuminate\Container\{closure}()
docker_lamp_1 |
docker_lamp_1 | 37 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:96
docker_lamp_1 | Illuminate\Container\Util::unwrapIfClosure(Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 38 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:35
docker_lamp_1 | Illuminate\Container\BoundMethod::callBoundMethod(Object(Illuminate\Foundation\Application), Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 39 vendor/laravel/framework/src/Illuminate/Container/Container.php:799
docker_lamp_1 | Illuminate\Container\BoundMethod::call(Object(Illuminate\Foundation\Application), [])
docker_lamp_1 |
docker_lamp_1 | 40 vendor/laravel/framework/src/Illuminate/Console/Command.php:211
docker_lamp_1 | Illuminate\Container\Container::call()
docker_lamp_1 |
docker_lamp_1 | 41 vendor/symfony/console/Command/Command.php:341
docker_lamp_1 | Illuminate\Console\Command::execute(Object(Symfony\Component\Console\Input\ArgvInput), Object(Illuminate\Console\OutputStyle))
docker_lamp_1 |
docker_lamp_1 | 42 vendor/laravel/framework/src/Illuminate/Console/Command.php:180
docker_lamp_1 | Symfony\Component\Console\Command\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Illuminate\Console\OutputStyle))
docker_lamp_1 |
docker_lamp_1 | 43 app/Console/Commands/Command.php:41
docker_lamp_1 | Illuminate\Console\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 44 vendor/symfony/console/Application.php:1117
docker_lamp_1 | Jiminny\Console\Commands\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 45 vendor/symfony/console/Application.php:356
docker_lamp_1 | Symfony\Component\Console\Application::doRunCommand(Object(Jiminny\Services\RecallAI\Commands\ScheduleBotCommand), Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 46 vendor/symfony/console/Application.php:195
docker_lamp_1 | Symfony\Component\Console\Application::doRun(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 47 vendor/laravel/framework/src/Illuminate/Foundation/Console/Kernel.php:198
docker_lamp_1 | Symfony\Component\Console\Application::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 48 vendor/laravel/framework/src/Illuminate/Foundation/Application.php:1235
docker_lamp_1 | Illuminate\Foundation\Console\Kernel::handle(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 49 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
docker_lamp_1 | 2026-04-14 12:16:22 Jiminny\Events\Activities\StatusUpdated [PASSWORD_DOTS] RUNNING
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:22Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:22Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | .. 12s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-14 12:16:22 Running ['artisan' dialers:monitor-activities] 2026-04-14 12:16:23 Jiminny\Events\Activities\StatusUpdated [PASSWORD_DOTS] 1s DONE
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:23Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:25Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:25Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:26Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:28Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:28Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:28Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:28Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | . 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:31Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:31Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:31Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:33Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:34Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:34Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:36Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:16:30 Running ['artisan' jiminny:monitor-social-accounts] 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:37Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:37Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:38Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:40Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:40Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:41Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:16:37 Running ['artisan' mailbox:skip-lists:refresh] . 5s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:43Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:43Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:43Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:16:42 Running ['artisan' mailbox:batch:process --max-batches=15] 3s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:46Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:46Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:46Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:16:46 Running ['artisan' conference:monitor:count] ... 2s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:49Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:49Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:50Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:51Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:16:48 Running ['artisan' activity:notify-not-logged] . 3s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:notify-not-logged > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:52Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:52Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:55Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:55Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:56Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:58Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:58Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:16:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:16:52 Running ['artisan' activity:status-count] {"canceled":35,"failed":1}...... 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' activity:status-count > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:01Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:01Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:01Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:04Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:04Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:04Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:06Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:17:00 Running ['artisan' mailbox:sync] Queueing 2 inbox(es) for sync.
docker_lamp_1 | [PASSWORD_DOTS] 6s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:sync > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:07Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:07Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-14 12:17:08 Jiminny\Jobs\Mailbox\SyncInbox [PASSWORD_DOTS] RUNNING
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:17:08 Jiminny\Jobs\Mailbox\SyncInbox [PASSWORD_DOTS] 190.21ms DONE
docker_lamp_1 | 2026-04-14 12:17:09 Jiminny\Jobs\Mailbox\SyncInbox [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:17:09 Jiminny\Jobs\Mailbox\SyncInbox [PASSWORD_DOTS] 278.16ms DONE
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:10Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:10Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:11Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:13Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:13Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:13Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:16Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:16Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:16Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:18Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:19Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:19Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:20Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:20Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:21Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:23Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:23Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:23Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:25Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:25Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:26Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:28Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:28Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:28Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:29Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:31Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:31Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:31Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:34Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:34Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:34Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:36Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:37Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:37Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:39Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:40Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:40Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:41Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:43Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:43Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:44Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:46Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:46Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:46Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:49Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:49Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:49Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:50Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:51Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:52Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:54Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:56Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:59Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:17:59Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elastic...
|
iTerm2
|
PROD (ssh)
|
NULL
|
13320
|
|
13255
|
kibana | {"type":"log", kibana | {"type":"log","@timestamp":"2026-04-14T12:13:43Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:46Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:46Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:49Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:49Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:50Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:52Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:52Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:55Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:55Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:13:30 Running ['artisan' calendar:sync --dateMode=daily] 2026-04-14 12:13:55 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... RUNNING
docker_lamp_1 | 2026-04-14 12:13:56 Jiminny\Jobs\Calendar\SyncCalendarEvents . 806.45ms DONE
docker_lamp_1 | 26s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' calendar:sync --dateMode=daily > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:58Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:58Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:13:58 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... RUNNING
docker_lamp_1 | 2026-04-14 12:13:59 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... 1s DONE
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:01Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:01Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:04Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:04Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:07Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:07Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:10Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:10Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:13Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:13Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:13Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:15Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:16Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:16Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:18Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:19Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:19Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-14 12:14:10 Running ['artisan' meeting-bot:schedule-bot] ... 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:20Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:20Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:20Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:22Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:22Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:23Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:25Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:25Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:25Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:14:20 Running ['artisan' dialers:monitor-activities] . 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:28Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:28Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:28Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:28Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:30Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:31Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:31Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:33Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:34Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:34Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:35Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:14:27 Running ['artisan' jiminny:monitor-social-accounts] 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:37Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:37Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:38Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:40Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:40Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:40Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:43Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:43Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:43Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:14:36 Running ['artisan' mailbox:skip-lists:refresh] . 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:46Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:46Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:49Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:49Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:50Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:52Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:52Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:14:44 Running ['artisan' mailbox:batch:process --max-batches=15] 11s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:55Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:55Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:58Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:58Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:01Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:01Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:01Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:04Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:04Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-14 12:14:55 Running ['artisan' conference:monitor:count] ... 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:06Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:07Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:07Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:10Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:10Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:11Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:13Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:13Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:13Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:16Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:16Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:16Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-14 12:15:05 Running ['artisan' crm:sync-objects] Syncing objects for Salesforce Team since 2026-03-19 20:14:10
docker_lamp_1 | Syncing objects for Hubspot since 2026-04-14 11:44:15
docker_lamp_1 | Syncing objects for Edge Communications since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Burlington Textiles Corp of America since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Uber since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Vonage since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Nexmo since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for SF since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for NewAccount since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for TestV since 2026-02-17 15:05:47
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for KioskAccount since 2026-02-17 15:06:27
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Pipedrive, Inc. since 2026-04-08 18:14:13
docker_lamp_1 | Syncing objects for Copper since 2026-04-14 11:44:16
docker_lamp_1 | Syncing objects for Pipedrive External Test since 2026-04-08 18:14:16
docker_lamp_1 | Syncing objects for InsightSquared since 2026-02-17 15:09:59
docker_lamp_1 | Syncing objects for Close since 2026-04-14 11:44:17
docker_lamp_1 | Syncing objects for DeewanyAccount since 2026-02-19 13:27:51
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Aircall Demo since 2026-02-17 15:06:51
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05
docker_lamp_1 | Syncing objects for GoStudent UAT since 2026-02-17 15:09:59
docker_lamp_1 | Syncing objects for Horen test since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed Hamadi since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed Testing since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed jiminny since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for JustCall since 2026-02-17 15:07:41
docker_lamp_1 | Syncing objects for Testers Inc since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | 2026-04-14 12:15:17 Jiminny\Jobs\Crm\SyncObjectsSyncing objects for Copper Team since 2026-02-19 13:26:04
docker_lamp_1 | [PASSWORD_DOTS]Team is not yet assigned an owner. skipping...
docker_lamp_1 | RUNNING
docker_lamp_1 | Syncing objects for TheBestPlace Ever since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Horen's apartments since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for New Org Test since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Pipedrive test org since 2026-02-19 13:27:42
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for BigChairs Inc. since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for GL 500 since 2026-02-19 13:26:04
docker_lamp_1 | Syncing objects for Loren X since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Laravel Company 2024 since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Barbara Fuchs since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Twilio Video since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for My Test Account 3000 since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for test since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Test since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Test salesforce auto sync since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for test ogg auto sync since 2026-02-17 15:07:16
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Globo Tech since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Laravel 11 Company 2024 since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | 2026-04-14 12:15:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 386.96ms DONE
docker_lamp_1 | 2026-04-14 12:15:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 |
docker_lamp_1 | Error
docker_lamp_1 |
docker_lamp_1 | Call to a member function getAttribute() on null
docker_lamp_1 |
docker_lamp_1 | at app/Console/Commands/Crm/SyncObjects.php:42
docker_lamp_1 | 38▕
docker_lamp_1 | 39▕ foreach ($teams as $team) {
docker_lamp_1 | 40▕ $config = $team->getCrmConfiguration();
docker_lamp_1 | 41▕
docker_lamp_1 | ➜ 42▕ if ($config->getAttribute('sync_objects') === false) {
docker_lamp_1 | 43▕ continue;
docker_lamp_1 | 44▕ }
docker_lamp_1 | 45▕
docker_lamp_1 | 46▕ $this->info(
docker_lamp_1 |
docker_lamp_1 | 1 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:36
docker_lamp_1 | Jiminny\Console\Commands\Crm\SyncObjects::handle()
docker_lamp_1 |
docker_lamp_1 | 2 vendor/laravel/framework/src/Illuminate/Container/Util.php:43
docker_lamp_1 | Illuminate\Container\BoundMethod::Illuminate\Container\{closure}()
docker_lamp_1 |
docker_lamp_1 | 3 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:96
docker_lamp_1 | Illuminate\Container\Util::unwrapIfClosure(Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 4 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:35
docker_lamp_1 | Illuminate\Container\BoundMethod::callBoundMethod(Object(Illuminate\Foundation\Application), Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 5 vendor/laravel/framework/src/Illuminate/Container/Container.php:799
docker_lamp_1 | Illuminate\Container\BoundMethod::call(Object(Illuminate\Foundation\Application), [])
docker_lamp_1 |
docker_lamp_1 | 6 vendor/laravel/framework/src/Illuminate/Console/Command.php:211
docker_lamp_1 | Illuminate\Container\Container::call()
docker_lamp_1 |
docker_lamp_1 | 7 vendor/symfony/console/Command/Command.php:341
docker_lamp_1 | Illuminate\Console\Command::execute(Object(Symfony\Component\Console\Input\ArgvInput), Object(Illuminate\Console\OutputStyle))
docker_lamp_1 |
docker_lamp_1 | 8 vendor/laravel/framework/src/Illuminate/Console/Command.php:180
docker_lamp_1 | Symfony\Component\Console\Command\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Illuminate\Console\OutputStyle))
docker_lamp_1 |
docker_lamp_1 | 9 app/Console/Commands/Command.php:41
docker_lamp_1 | Illuminate\Console\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 10 vendor/symfony/console/Application.php:1117
docker_lamp_1 | Jiminny\Console\Commands\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 11 vendor/symfony/console/Application.php:356
docker_lamp_1 | Symfony\Component\Console\Application::doRunCommand(Object(Jiminny\Console\Commands\Crm\SyncObjects), Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 12 vendor/symfony/console/Application.php:195
docker_lamp_1 | Symfony\Component\Console\Application::doRun(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 13 vendor/laravel/framework/src/Illuminate/Foundation/Console/Kernel.php:198
docker_lamp_1 | Symfony\Component\Console\Application::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 14 vendor/laravel/framework/src/Illuminate/Foundation/Application.php:1235
docker_lamp_1 | Illuminate\Foundation\Console\Kernel::handle(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 15 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:18Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | [PASSWORD_DOTS] 13s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:sync-objects > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | 2026-04-14 12:15:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 896.59ms DONE
docker_lamp_1 | 2026-04-14 12:15:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 115.49ms DONE
docker_lamp_1 | 2026-04-14 12:15:19 Jiminny\Jobs\Crm\SyncObjectsrun_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 | [PASSWORD_DOTS] RUNNING
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:19Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:19Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:20Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:20Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
docker_lamp_1 | 2026-04-14 12:15:20 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-14 12:15:20 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:20Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:15:21 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 75.20ms DONE
docker_lamp_1 | 2026-04-14 12:15:21 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:21 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 51.77ms DONE
docker_lamp_1 | 2026-04-14 12:15:21 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:22Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:22Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 75.17ms DONE
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 42.25ms DONE
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 52.64ms DONE
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:23 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 38.54ms DONE
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:23Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:25Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:25Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:25Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:28Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:28Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"20...
|
iTerm2
|
PROD (-zsh)
|
NULL
|
13255
|
|
13257
|
kibana | {"type":"log", kibana | {"type":"log","@timestamp":"2026-04-14T12:13:43Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:46Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:46Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:49Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:49Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:50Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:52Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:52Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:55Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:55Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:13:30 Running ['artisan' calendar:sync --dateMode=daily] 2026-04-14 12:13:55 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... RUNNING
docker_lamp_1 | 2026-04-14 12:13:56 Jiminny\Jobs\Calendar\SyncCalendarEvents . 806.45ms DONE
docker_lamp_1 | 26s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' calendar:sync --dateMode=daily > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | run_artisan_schedule: Done waiting for schedule:run
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:58Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:58Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:13:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:13:58 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... RUNNING
docker_lamp_1 | 2026-04-14 12:13:59 Jiminny\Jobs\Calendar\SyncCalendarEvents ....... 1s DONE
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:01Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:01Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:04Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:04Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:07Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:07Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:10Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:10Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:13Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:13Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:13Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:15Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:16Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:16Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:18Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:19Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:19Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-14 12:14:10 Running ['artisan' meeting-bot:schedule-bot] ... 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:20Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:20Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:20Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:22Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:22Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:23Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:25Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:25Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:25Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:14:20 Running ['artisan' dialers:monitor-activities] . 7s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:28Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:28Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:28Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:28Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:30Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:31Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:31Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:33Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:34Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:34Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:35Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:14:27 Running ['artisan' jiminny:monitor-social-accounts] 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:37Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:37Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:38Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:40Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:40Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:40Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:43Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:43Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:43Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:14:36 Running ['artisan' mailbox:skip-lists:refresh] . 8s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:46Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:46Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:49Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:49Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:49Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:50Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:50Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:52Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:52Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:14:44 Running ['artisan' mailbox:batch:process --max-batches=15] 11s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:55Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:55Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:58Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:58Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:14:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:01Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:01Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:01Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:04Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:04Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-14 12:14:55 Running ['artisan' conference:monitor:count] ... 9s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' conference:monitor:count > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:06Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:07Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:07Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:10Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:10Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:11Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:13Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:13Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:13Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:16Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:16Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:16Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-14 12:15:05 Running ['artisan' crm:sync-objects] Syncing objects for Salesforce Team since 2026-03-19 20:14:10
docker_lamp_1 | Syncing objects for Hubspot since 2026-04-14 11:44:15
docker_lamp_1 | Syncing objects for Edge Communications since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Burlington Textiles Corp of America since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Uber since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Vonage since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Nexmo since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for SF since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for NewAccount since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for TestV since 2026-02-17 15:05:47
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for KioskAccount since 2026-02-17 15:06:27
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Pipedrive, Inc. since 2026-04-08 18:14:13
docker_lamp_1 | Syncing objects for Copper since 2026-04-14 11:44:16
docker_lamp_1 | Syncing objects for Pipedrive External Test since 2026-04-08 18:14:16
docker_lamp_1 | Syncing objects for InsightSquared since 2026-02-17 15:09:59
docker_lamp_1 | Syncing objects for Close since 2026-04-14 11:44:17
docker_lamp_1 | Syncing objects for DeewanyAccount since 2026-02-19 13:27:51
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Aircall Demo since 2026-02-17 15:06:51
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05
docker_lamp_1 | Syncing objects for GoStudent UAT since 2026-02-17 15:09:59
docker_lamp_1 | Syncing objects for Horen test since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed Hamadi since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed Testing since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Ahmed jiminny since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for JustCall since 2026-02-17 15:07:41
docker_lamp_1 | Syncing objects for Testers Inc since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | 2026-04-14 12:15:17 Jiminny\Jobs\Crm\SyncObjectsSyncing objects for Copper Team since 2026-02-19 13:26:04
docker_lamp_1 | [PASSWORD_DOTS]Team is not yet assigned an owner. skipping...
docker_lamp_1 | RUNNING
docker_lamp_1 | Syncing objects for TheBestPlace Ever since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Horen's apartments since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for New Org Test since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Pipedrive test org since 2026-02-19 13:27:42
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for BigChairs Inc. since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for GL 500 since 2026-02-19 13:26:04
docker_lamp_1 | Syncing objects for Loren X since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Laravel Company 2024 since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Barbara Fuchs since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Twilio Video since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for My Test Account 3000 since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for test since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Test since 2026-02-17 15:09:59
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Test salesforce auto sync since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Bullhorn since 2026-02-19 13:28:05
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for test ogg auto sync since 2026-02-17 15:07:16
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Globo Tech since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | Syncing objects for Laravel 11 Company 2024 since 2026-02-19 13:26:04
docker_lamp_1 | Team is not yet assigned an owner. skipping...
docker_lamp_1 | 2026-04-14 12:15:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 386.96ms DONE
docker_lamp_1 | 2026-04-14 12:15:18 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 |
docker_lamp_1 | Error
docker_lamp_1 |
docker_lamp_1 | Call to a member function getAttribute() on null
docker_lamp_1 |
docker_lamp_1 | at app/Console/Commands/Crm/SyncObjects.php:42
docker_lamp_1 | 38▕
docker_lamp_1 | 39▕ foreach ($teams as $team) {
docker_lamp_1 | 40▕ $config = $team->getCrmConfiguration();
docker_lamp_1 | 41▕
docker_lamp_1 | ➜ 42▕ if ($config->getAttribute('sync_objects') === false) {
docker_lamp_1 | 43▕ continue;
docker_lamp_1 | 44▕ }
docker_lamp_1 | 45▕
docker_lamp_1 | 46▕ $this->info(
docker_lamp_1 |
docker_lamp_1 | 1 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:36
docker_lamp_1 | Jiminny\Console\Commands\Crm\SyncObjects::handle()
docker_lamp_1 |
docker_lamp_1 | 2 vendor/laravel/framework/src/Illuminate/Container/Util.php:43
docker_lamp_1 | Illuminate\Container\BoundMethod::Illuminate\Container\{closure}()
docker_lamp_1 |
docker_lamp_1 | 3 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:96
docker_lamp_1 | Illuminate\Container\Util::unwrapIfClosure(Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 4 vendor/laravel/framework/src/Illuminate/Container/BoundMethod.php:35
docker_lamp_1 | Illuminate\Container\BoundMethod::callBoundMethod(Object(Illuminate\Foundation\Application), Object(Closure))
docker_lamp_1 |
docker_lamp_1 | 5 vendor/laravel/framework/src/Illuminate/Container/Container.php:799
docker_lamp_1 | Illuminate\Container\BoundMethod::call(Object(Illuminate\Foundation\Application), [])
docker_lamp_1 |
docker_lamp_1 | 6 vendor/laravel/framework/src/Illuminate/Console/Command.php:211
docker_lamp_1 | Illuminate\Container\Container::call()
docker_lamp_1 |
docker_lamp_1 | 7 vendor/symfony/console/Command/Command.php:341
docker_lamp_1 | Illuminate\Console\Command::execute(Object(Symfony\Component\Console\Input\ArgvInput), Object(Illuminate\Console\OutputStyle))
docker_lamp_1 |
docker_lamp_1 | 8 vendor/laravel/framework/src/Illuminate/Console/Command.php:180
docker_lamp_1 | Symfony\Component\Console\Command\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Illuminate\Console\OutputStyle))
docker_lamp_1 |
docker_lamp_1 | 9 app/Console/Commands/Command.php:41
docker_lamp_1 | Illuminate\Console\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 10 vendor/symfony/console/Application.php:1117
docker_lamp_1 | Jiminny\Console\Commands\Command::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 11 vendor/symfony/console/Application.php:356
docker_lamp_1 | Symfony\Component\Console\Application::doRunCommand(Object(Jiminny\Console\Commands\Crm\SyncObjects), Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 12 vendor/symfony/console/Application.php:195
docker_lamp_1 | Symfony\Component\Console\Application::doRun(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 13 vendor/laravel/framework/src/Illuminate/Foundation/Console/Kernel.php:198
docker_lamp_1 | Symfony\Component\Console\Application::run(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 14 vendor/laravel/framework/src/Illuminate/Foundation/Application.php:1235
docker_lamp_1 | Illuminate\Foundation\Console\Kernel::handle(Object(Symfony\Component\Console\Input\ArgvInput), Object(Symfony\Component\Console\Output\ConsoleOutput))
docker_lamp_1 |
docker_lamp_1 | 15 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:18Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | [PASSWORD_DOTS] 13s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' crm:sync-objects > '/proc/1/fd/1' 2>&1
docker_lamp_1 |
docker_lamp_1 | 2026-04-14 12:15:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 896.59ms DONE
docker_lamp_1 | 2026-04-14 12:15:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:19 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 115.49ms DONE
docker_lamp_1 | 2026-04-14 12:15:19 Jiminny\Jobs\Crm\SyncObjectsrun_artisan_schedule: Done waiting for schedule:run
docker_lamp_1 | [PASSWORD_DOTS] RUNNING
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:19Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mnyadn6600070bdf86d69yx1 - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:19Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:19Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:20Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:20Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:20Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
docker_lamp_1 | 2026-04-14 12:15:20 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-14 12:15:20 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:20Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-14 12:15:21 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 75.20ms DONE
docker_lamp_1 | 2026-04-14 12:15:21 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:21 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 51.77ms DONE
docker_lamp_1 | 2026-04-14 12:15:21 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:22Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:22Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 1s DONE
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 75.17ms DONE
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 42.25ms DONE
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 52.64ms DONE
docker_lamp_1 | 2026-04-14 12:15:22 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] RUNNING
docker_lamp_1 | 2026-04-14 12:15:23 Jiminny\Jobs\Crm\SyncObjects [PASSWORD_DOTS] 38.54ms DONE
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:23Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:25Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:25Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:25Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:28Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-14T12:15:28Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"20...
|
iTerm2
|
PROD (-zsh)
|
NULL
|
13257
|
|
19759
|
key
SlackFileEditViewGoHistoryWindowHelp(ih|[Platf key
SlackFileEditViewGoHistoryWindowHelp(ih|[Platform) Planning... 2 m leftmeet.google.com/tgb-pyuf-dri?authuser=lukas.kovalik%40jiminny.com100% C8• Wed 15 Apr 10:58:429Galya Dimitrova (Presenting, annotating)SafariFileEditViewHistoryBookmarksDevelopkeyA Sentry3 HubSpot|# Platform Tea.88For youRecent# Starred89 Apps% PlansQ SpacesStarredJiminny (New)| I Platform TeamIID Processing TeamIID SE KanbanIID Capture TeamШD Enterprise Stability I...DiscoveryProductRecentService-Desk= More spaces= FiltersIB DashboardsOperations@ ОрeпA!|#i Seit ServiaS GTMSearchSpaces / Jiminny (New)Platform Team88® Summary& TimelinSearch backlog• v Platform Sprint 2A JY-18909Pal@ JY-19798EvalA JY-20553DelA JY-20632PreA JY-20489D JY-9712Char© JY-20564N JY-20372 AIFA JY-20157 SenA JY-20508 NotC JY-20278AJ Panorama> Don't show internal errors to customers& Customersmeet.google.com is sharing your screen.Keychain Access.app/Applications/Utilities/Keychain Access.appKeyboard Maestro.app/Applications/Keyboard Maestro.appKeyboardSystem Settings → KeyboardKeychron_Screen~/Keychron_Screentest_stream_keepalive.py-/jiminny/prophet/tests/component/streaming/test_stream_keepalive.pystream_keepalive.py~/jiminny/prophet/src/component/streaming/stream_keepalive.pykey_points_request.py~/jiminny/prophet/src/component/call_processing/schemes/key_points_request.pykey_points.txt-/jiminny/prophet/src/component/call_processing/prompts/key_points/en/key_points.txtkey_points_prompt_builder.py-/jiminny/prophet/src/component/call__processing/[API_KEY].pyBACKLOGweeouyCancel|StartBACKLOGASK ANYTHING ON A.BACKLOG282883₴84₴5286$87488489Stetka StoyanovaGalya Dimitrova4 othersNikolay Nikolov21Stop sharing10:58 AM | [Platform] Planning | Session i7Lukas Kovalik58:29...
|
Alfred
|
Alfred
|
NULL
|
19759
|
|
19758
|
key
FileDMsAchivityFilesMoreJiminny ...= Unreads@ key
FileDMsAchivityFilesMoreJiminny ...= Unreads@ Threads6 HuddlesDrafts & sent8 DirectoriesEh External connections# Starred& platform-inner-team# Channels# ai-chapter# alerts# backend#: confusion-clinic# curiosity_labiengineering# frontend# general# infra-changes# jiminny-bg# platform-tickets# product_launches# randomit releases# sofia-office# support# thank-yous# the people of jimi...Direct messazesAneliya Angelova,€. Vasil VasilevR. Steliyan Georgiev E3 Adelina Petrova, Ili...Adelina PetrovaP. Galya Dimitrova iD. Nikolay Nikolov2 Galya Dimitrova, Ni...2 Galya Dimitrova, Ni...#: AppsE Jira CloudToastGoogle Cale….HistoryWindowHelpQ Search Jiminny IncJira CloudHomeMessagesAboutJuis LeccventurerdeorveiWednesday, April 1st~commentJira Cloud APP 3:00 PM4 ®Nikolay Nikolov edited a comment on a Bug you reportedJY-20530 Pipedrive opportunity sync issueWe have re-synced missed updated deals from 01.01.2026 - now:-- LUceaneo inameoldnew2ZZ | Caribou44344lne Electric car schemeliSU11450/377 ISEP22458ooy lAnaco1511805g-- US-- teamID Inameoldnew145levermedoou cenptans648 lAdvanto786| ColormaticsXXodevonkev9401 Shellworks1oolldrowchrad1086|Pipedrive1113|LeadVenturePipedrivel823975611X/28429• 1878113426104316878CommentMore actions...Monday. April 6th vJira Cloud APP 11:31 AM• @Galya Dimitrova updated the description of a Story you are assigned toJY-20543 AJ Reports > TrackingStatus: Backlog• Type: StoryLK) Assignee: Lukas KovalikT Priority: MediumCommentTransitionToday "Jira Cloud APP 10:10 AM4 @Stefka Stoyanova transitioned a Story you are assigned to from |Code Review → Ready for DeploymentJY-20543 AJ Reports > TrackingStatus: Ready for Deploymentlype. storAssignee: Lukas Kovalik Priority: MediumCommentMore actions....@Stefka Stoyanova transitioned a Story you are assigned to from Ready for Deployment → ClosedJY-20543 AJ Reports > IrackingStatus: ClosedType: StoryLK Assignee: Lukas KovalikT Priority: MediumcommentMore actions.Message Jira CloudAaInu 1oFri 17$0.l [Platform] Planning... 2m left(A100% 778• Wed 15 Apr 10:58:41Week vTodaysatloSun 19Daily - Platform 09:45Backend ChapterJimid -Y Tech Day# Support Daily 15:00Tech Day Review...
|
Alfred
|
Alfred
|
NULL
|
19758
|
|
72350
|
kenazk/gong-mcp: Simple Gong MCP server
github.com kenazk/gong-mcp: Simple Gong MCP server
github.com
Platform Sprint 2 Q2 - Platform Team - Scrum Board - Jira
SevenShores\Hubspot\Exceptions\BadRequest: Client error: `POST https://api.hubapi.com/crm/v3/objects/contact/search` resulted in a `429 Too Many Requests` response: {"status":"error","message":"You have reached your secondly limit.","errorType":"RATE_LIMIT
SevenShores\Hubspot\Exceptions\BadRequest: Client error: `POST https://api.hubapi.com/crm/v3/objects/contact/search` resulted in a `429 Too Many Requests` response: {"status":"error","message":"You have reached your secondly limit.","errorType":"RATE_LIMIT
[JY-20372] AI Reports > Empty page design and promotion - Jira
[JY-20372] AI Reports > Empty page design and promotion - Jira
Jiminny MCP Connector - Product - Confluence
Jiminny MCP Connector - Product - Confluence
Jiminny
Jiminny
Workers | Datadog
Workers | Datadog
Pull requests · jiminny/app
Pull requests · jiminny/app
[JY-20728] [HubSpot] Find the root cause of 429 hit and tweak API client rate limiter - Jira
[JY-20728] [HubSpot] Find the root cause of 429 hit and tweak API client rate limiter - Jira
JY-9712 | Nuges to expire after one year by nikolaybiaivanov · Pull Request #11981 · jiminny/app
JY-9712 | Nuges to expire after one year by nikolaybiaivanov · Pull Request #11981 · jiminny/app
Jiminny
Jiminny
CloudWatch | us-east-2
CloudWatch | us-east-2
Your 'Not enpough activities' report wasn't generated - [EMAIL] - Jiminny Mail
Your 'Not enpough activities' report wasn't generated - [EMAIL] - Jiminny Mail
Jiminny MCP Connector - Product - Confluence
Jiminny MCP Connector - Product - Confluence
kenazk/gong-mcp: Simple Gong MCP server
kenazk/gong-mcp: Simple Gong MCP server
Close tab
JY-9712 | change nudges schema by nikolaybiaivanov · Pull Request #11983 · jiminny/app
JY-9712 | change nudges schema by nikolaybiaivanov · Pull Request #11983 · jiminny/app
JY-9712 | Nuges to expire after one year by nikolaybiaivanov · Pull Request #11981 · jiminny/app
JY-9712 | Nuges to expire after one year by nikolaybiaivanov · Pull Request #11981 · jiminny/app
Architecture overview - Model Context Protocol
Architecture overview - Model Context Protocol
Laravel MCP - AI tooling for Laravel, by the Laravel team | Laravel - The clean stack for Artisans and agents
Laravel MCP - AI tooling for Laravel, by the Laravel team | Laravel - The clean stack for Artisans and agents
Laravel MCP | Laravel 13.x - The clean stack for Artisans and agents
Laravel MCP | Laravel 13.x - The clean stack for Artisans and agents
New Tab
Customize sidebar
Open Google Gemini (⌃X)
Tabs from other devices
Open history (⇧⌘H)
Open bookmarks (⌘B)
Skip to content
Skip to content
Open menu
Homepage (g then d)
kenazk
kenazk
gong-mcp
gong-mcp
Search or jump to…
Type
/
to search
Chat with Copilot
Open Copilot…
Create new...
Issues
Pull requests
Repositories
You have unread notifications(g then n)
Open user navigation menu
Repository navigation
Repository navigation
Code
Code
Issues (1)
Issues
(
1
)
Pull requests (10)
Pull requests
(
10
)
Agents
Agents
Actions
Actions
Projects
Projects
Security and quality
Security and quality
Insights
Insights
Important update
Important update
On April 24 we'll start using GitHub Copilot interaction data for AI model training unless you opt out.
Review this update
Review this update
and manage your preferences in your
GitHub account settings
GitHub account settings
.
Dismiss banner
gong-mcp
gong-mcp
Public
Watch: Participating in kenazk/gong-mcp
Watch
(
2
)
Fork 21
Fork
21
See your forks of this repository
Star this repository (31)
Star
31
Add this repository to a list
kenazk/gong-mcp
kenazk/gong-mcp
master branch
master
4 Branches
4
Branches
0 Tags
0
Tags
Go to file
Add file
Add file
Add file
Add file
Code
Code
Folders and files
Folders and files
Name
Last commit message
Last commit date
Latest commit
Latest commit
kenazk
commits by kenazk
kenazk
Update README.md
Update README.md
Commit aa952b1
aa952b1
·
last year
History
History
2 Commits
2 Commits
src, (Directory)
src
Initial commit
Initial commit
last year
.env.example, (File)
.env.example
Initial commit
Initial commit
last year
.gitignore, (File)
.gitignore
Initial commit
Initial commit
last year
Dockerfile, (File)
Dockerfile
Initial commit
Initial commit
last year
LICENSE, (File)
LICENSE
Initial commit
Initial commit
last year
README.md, (File)
README.md
Update README.md
Update README.md
last year
package.json, (File)
package.json
Initial commit
Initial commit
last year
tsconfig.json, (File)
tsconfig.json
Initial commit
Initial commit
last year
Name
Latest commit
Latest commit
kenazk
commits by kenazk
kenazk
Update README.md
Update README.md
Commit aa952b1
aa952b1
·
last year
History
History
2 Commits
2 Commits
src, (Directory)
src
.env.example, (File)
.env.example
.gitignore, (File)
.gitignore
Dockerfile, (File)
Dockerfile
LICENSE, (File)
LICENSE
README.md, (File)
README.md
package.json, (File)
package.json
tsconfig.json, (File)
tsconfig.json
Last commit message
Latest commit
Latest commit
kenazk
commits by kenazk
kenazk
Update README.md
Update README.md
Commit aa952b1
aa952b1
·
last year
History
History
2 Commits
2 Commits
Initial commit
Initial commit
Initial commit
Initial commit
Initial commit
Initial commit
Initial commit
Initial commit
Initial commit
Initial commit
Update README.md
Update README.md
Initial commit
Initial commit
Initial commit
Initial commit
Last commit date
Latest commit
Latest commit
kenazk
commits by kenazk
kenazk
Update README.md
Update README.md
Commit aa952b1
aa952b1
·
last year
History
History
2 Commits
2 Commits
last year
last year
last year
last year
last year
last year
last year
last year
Repository files navigation
Repository files navigation
README
README
MIT license
MIT license
Edit file
Outline
Gong MCP Server
Gong MCP Server
Permalink: Gong MCP Server...
|
Firefox
|
kenazk/gong-mcp: Simple Gong MCP server — Work
|
github.com/kenazk/gong-mcp
|
72350
|
|
72351
|
kenazk/gong-mcp: Simple Gong MCP server
github.com kenazk/gong-mcp: Simple Gong MCP server
github.com
Platform Sprint 2 Q2 - Platform Team - Scrum Board - Jira
SevenShores\Hubspot\Exceptions\BadRequest: Client error: `POST [URL_WITH_CREDENTIALS]
kenazk Kenaz Kwa
kenazk
Kenaz Kwa
Languages
Languages
JavaScript 96.8%
JavaScript
96.8%
Dockerfile 3.2%
Dockerfile
3.2%
Footer
Footer
GitHub Homepage
© 2026 GitHub, Inc.
Footer navigation
Footer navigation
Terms
Terms
Privacy
Privacy
Security
Security
Status
Status
Community
Community
Docs
Docs
Contact
Contact
Manage cookies
Do not share my personal information...
|
Firefox
|
kenazk/gong-mcp: Simple Gong MCP server — Work
|
github.com/kenazk/gong-mcp
|
72351
|
|
19757
|
k
SlackFileEditViewGoHistoryWindowHelpllol| [Platf k
SlackFileEditViewGoHistoryWindowHelpllol| [Platform] Planning... 2 m leftmeet.google.com/tgb-pyuf-dri?authuser=lukas.kovalik%40jiminny.comGalya Dimitrova (Presenting, annotating)SafariFileEditViewHistoryBookmarksDevelopA Sentry3 HubSpot|@ ОрeпA!|#i Seit Service# Platform Tes.88For youRecent# Starred89 Apps% Plans0 SpacesStarredJiminny (New)| II Platform TeamIID Processing TeamIID SE KanbanIID Capture TeamШD Enterprise Stability I...DiscoveryProductRecent4 Service-Desk= More spaces= FiltersSearchAWS USAWS EU7 A Journey tr.Start Sprint83 Features Ad.E Ask Sminey -Spaces / Jiminny (New)Platform Team88® Summary& TimelineWork items JY-20586, JY-20550, JY-20567, JY-20568 and 5 others do nothave a value for the 'Estimate' field. Values entered after the start of thesprint will be treated as scope change.Search backlog15 work items will be included in this sprint.Required fields are marked with an asterisk •Sprint name *Platform Sprint 2 Q2• v Platform Sprint 2 €2 weeksA JY-18909[PaStart date *© JY-19798 EvaA JY-20553 Del15/04/202610:56Planned start date: 15/04/2026, 01:00A sprint's start date impacts velocity and scope in reports. Learn more, (?End date *A JY-20632Pre01:00 |A JY-20489RevD JY-9712Chan© JY-20564 InwSprint goalLaunch AJ Panorama Reports for Jiminny usersOptimise nudges - lower the resources they need for executionN JY-20372 AIFA JY-20157 Sen1 LLM Evaluat• UpgradeAsk Rovo•8B List& Forms• ComponentsMore 9++29.59Start sprintIN DEV VCODE REVIEW1IN DEV V2BUG FIXING VECTIVE AN.BACKLOGECTIVE AN...BACKLOGREADY FOR DEVBACKLOG=Cancel|Stak2IB DashboardsA JY-20508 NotBACKLOG1OperationsC JY-20278AJ Panorama> Don't show internal errors to customersASK ANYTHING ON A..& CustomersBACKLOGmeet.google.com is sharing your screen.Stop sharing10:58 AM | [Platform] Planning | Session zSộ3•)).100% C8• Wed 15 Apr 10:58:399Stefka StoyanovaGalya Dimitrova4 othersNikolay NikoloyLukas Kovalik58:26...
|
Alfred
|
Alfred
|
NULL
|
19757
|
|
52314
|
jiminny.atlassian.net/jira/software/c/projects/JY/ jiminny.atlassian.net/jira/software/c/projects/JY/boards/37
[SRD-6793] Les Mills activity types not pulling in - Jira
[SRD-6793] Les Mills activity types not pulling in - Jira
JY-20698 handle failed field sync on playbook import activity types by LakyLak · Pull Request #11988 · jiminny/app
JY-20698 handle failed field sync on playbook import activity types by LakyLak · Pull Request #11988 · jiminny/app
JY-20692 change confirmation parameter by LakyLak · Pull Request #11986 · jiminny/app
JY-20692 change confirmation parameter by LakyLak · Pull Request #11986 · jiminny/app
[JY-20543] AJ Reports > Tracking - Jira
[JY-20543] AJ Reports > Tracking - Jira
[JY-18909] [Part2] Automated reports with Ask Jiminny - Jira
[JY-18909] [Part2] Automated reports with Ask Jiminny - Jira
Ask Jiminny Reports by nikolay-yankov · Pull Request #11894 · jiminny/app
Ask Jiminny Reports by nikolay-yankov · Pull Request #11894 · jiminny/app
New Tab
New Tab
Product Growth Platform | Userpilot
Product Growth Platform | Userpilot
Userpilot | Events
Userpilot | Events...
|
Firefox
|
Work — Mozilla Firefox
|
jiminny.atlassian.net/jira/software/c/projects/JY/ jiminny.atlassian.net/jira/software/c/projects/JY/boards/37...
|
52314
|
|
79311
|
ith Redis. To fix this issue run the command ' ith Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.
redis | 1:M 24 Apr 2026 15:57:13.349 * Reading RDB preamble from AOF file...
redis | 1:M 24 Apr 2026 15:57:13.351 * Reading the remaining AOF tail...
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="no configuration paths supplied"
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="using configuration at default config path" path=/home/ngrok/.ngrok2/ngrok.yml
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="open config file" path=/home/ngrok/.ngrok2/ngrok.yml err=nil
ngrok | t=2026-04-24T15:57:14+0000 lvl=info msg="starting web service" obj=web addr=[IP_ADDRESS]:4040
blackfire-1 exited with code 1
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="tunnel session started" obj=tunnels.session
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="client session established" obj=csess id=56b3c75875b2
ngrok | t=2026-04-24T15:57:16+0000 lvl=info msg="started tunnel" obj=tunnels name="command_line (http)" addr=[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:24,909Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:25,711Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:28,712Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,987Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,998Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:33,530Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
docker_lamp_1 |
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:37,166Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,048Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 232, version: 8596, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,723Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 232, version: 8596, reason: Publication{term=232, version=8596}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:41,707Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][10] overhead, spent [429ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,374Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,409Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:43,780Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][12] overhead, spent [388ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:52,923Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][21] overhead, spent [460ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:55,229Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][23] overhead, spent [428ms] collecting in the last [1.2s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", ...
|
iTerm2
|
DOCKER (docker-compose)
|
NULL
|
79311
|
|
79312
|
ith Redis. To fix this issue run the command ' ith Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.
redis | 1:M 24 Apr 2026 15:57:13.349 * Reading RDB preamble from AOF file...
redis | 1:M 24 Apr 2026 15:57:13.351 * Reading the remaining AOF tail...
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="no configuration paths supplied"
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="using configuration at default config path" path=/home/ngrok/.ngrok2/ngrok.yml
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="open config file" path=/home/ngrok/.ngrok2/ngrok.yml err=nil
ngrok | t=2026-04-24T15:57:14+0000 lvl=info msg="starting web service" obj=web addr=[IP_ADDRESS]:4040
blackfire-1 exited with code 1
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="tunnel session started" obj=tunnels.session
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="client session established" obj=csess id=56b3c75875b2
ngrok | t=2026-04-24T15:57:16+0000 lvl=info msg="started tunnel" obj=tunnels name="command_line (http)" addr=[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:24,909Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:25,711Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:28,712Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,987Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,998Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:33,530Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
docker_lamp_1 |
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:37,166Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,048Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 232, version: 8596, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,723Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 232, version: 8596, reason: Publication{term=232, version=8596}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:41,707Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][10] overhead, spent [429ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,374Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,409Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:43,780Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][12] overhead, spent [388ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:52,923Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][21] overhead, spent [460ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:55,229Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][23] overhead, spent [428ms] collecting in the last [1.2s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", ...
|
iTerm2
|
DOCKER (docker-compose)
|
NULL
|
79312
|
|
79313
|
ith Redis. To fix this issue run the command ' ith Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.
redis | 1:M 24 Apr 2026 15:57:13.349 * Reading RDB preamble from AOF file...
redis | 1:M 24 Apr 2026 15:57:13.351 * Reading the remaining AOF tail...
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="no configuration paths supplied"
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="using configuration at default config path" path=/home/ngrok/.ngrok2/ngrok.yml
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="open config file" path=/home/ngrok/.ngrok2/ngrok.yml err=nil
ngrok | t=2026-04-24T15:57:14+0000 lvl=info msg="starting web service" obj=web addr=[IP_ADDRESS]:4040
blackfire-1 exited with code 1
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="tunnel session started" obj=tunnels.session
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="client session established" obj=csess id=56b3c75875b2
ngrok | t=2026-04-24T15:57:16+0000 lvl=info msg="started tunnel" obj=tunnels name="command_line (http)" addr=[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:24,909Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:25,711Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:28,712Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,987Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,998Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:33,530Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
docker_lamp_1 |
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:37,166Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,048Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 232, version: 8596, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,723Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 232, version: 8596, reason: Publication{term=232, version=8596}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:41,707Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][10] overhead, spent [429ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,374Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,409Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:43,780Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][12] overhead, spent [388ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:52,923Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][21] overhead, spent [460ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:55,229Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][23] overhead, spent [428ms] collecting in the last [1.2s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", ...
|
iTerm2
|
DOCKER (docker-compose)
|
NULL
|
79313
|
|
79314
|
ith Redis. To fix this issue run the command ' ith Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.
redis | 1:M 24 Apr 2026 15:57:13.349 * Reading RDB preamble from AOF file...
redis | 1:M 24 Apr 2026 15:57:13.351 * Reading the remaining AOF tail...
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="no configuration paths supplied"
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="using configuration at default config path" path=/home/ngrok/.ngrok2/ngrok.yml
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="open config file" path=/home/ngrok/.ngrok2/ngrok.yml err=nil
ngrok | t=2026-04-24T15:57:14+0000 lvl=info msg="starting web service" obj=web addr=[IP_ADDRESS]:4040
blackfire-1 exited with code 1
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="tunnel session started" obj=tunnels.session
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="client session established" obj=csess id=56b3c75875b2
ngrok | t=2026-04-24T15:57:16+0000 lvl=info msg="started tunnel" obj=tunnels name="command_line (http)" addr=[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:24,909Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:25,711Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:28,712Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,987Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,998Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:33,530Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
docker_lamp_1 |
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:37,166Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,048Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 232, version: 8596, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,723Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 232, version: 8596, reason: Publication{term=232, version=8596}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:41,707Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][10] overhead, spent [429ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,374Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,409Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:43,780Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][12] overhead, spent [388ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:52,923Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][21] overhead, spent [460ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:55,229Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][23] overhead, spent [428ms] collecting in the last [1.2s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", ...
|
iTerm2
|
DOCKER (docker-compose)
|
NULL
|
79314
|
|
79315
|
ith Redis. To fix this issue run the command ' ith Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.
redis | 1:M 24 Apr 2026 15:57:13.349 * Reading RDB preamble from AOF file...
redis | 1:M 24 Apr 2026 15:57:13.351 * Reading the remaining AOF tail...
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="no configuration paths supplied"
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="using configuration at default config path" path=/home/ngrok/.ngrok2/ngrok.yml
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="open config file" path=/home/ngrok/.ngrok2/ngrok.yml err=nil
ngrok | t=2026-04-24T15:57:14+0000 lvl=info msg="starting web service" obj=web addr=[IP_ADDRESS]:4040
blackfire-1 exited with code 1
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="tunnel session started" obj=tunnels.session
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="client session established" obj=csess id=56b3c75875b2
ngrok | t=2026-04-24T15:57:16+0000 lvl=info msg="started tunnel" obj=tunnels name="command_line (http)" addr=[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:24,909Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:25,711Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:28,712Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,987Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,998Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:33,530Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
docker_lamp_1 |
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:37,166Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,048Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 232, version: 8596, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,723Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 232, version: 8596, reason: Publication{term=232, version=8596}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:41,707Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][10] overhead, spent [429ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,374Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,409Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:43,780Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][12] overhead, spent [388ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:52,923Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][21] overhead, spent [460ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:55,229Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][23] overhead, spent [428ms] collecting in the last [1.2s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", ...
|
iTerm2
|
DOCKER (docker-compose)
|
NULL
|
79315
|
|
79316
|
ith Redis. To fix this issue run the command ' ith Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.
redis | 1:M 24 Apr 2026 15:57:13.349 * Reading RDB preamble from AOF file...
redis | 1:M 24 Apr 2026 15:57:13.351 * Reading the remaining AOF tail...
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="no configuration paths supplied"
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="using configuration at default config path" path=/home/ngrok/.ngrok2/ngrok.yml
ngrok | t=2026-04-24T15:57:13+0000 lvl=info msg="open config file" path=/home/ngrok/.ngrok2/ngrok.yml err=nil
ngrok | t=2026-04-24T15:57:14+0000 lvl=info msg="starting web service" obj=web addr=[IP_ADDRESS]:4040
blackfire-1 exited with code 1
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="tunnel session started" obj=tunnels.session
ngrok | t=2026-04-24T15:57:15+0000 lvl=info msg="client session established" obj=csess id=56b3c75875b2
ngrok | t=2026-04-24T15:57:16+0000 lvl=info msg="started tunnel" obj=tunnels name="command_line (http)" addr=[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:24,909Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:25,711Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:28,712Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,987Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:30,998Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:33,530Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
docker_lamp_1 |
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:37,166Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,048Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 232, version: 8596, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:38,723Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{9lc--KiSQXiV2S7EF2U1cQ}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 232, version: 8596, reason: Publication{term=232, version=8596}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:39,543Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:41,707Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][10] overhead, spent [429ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,374Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:42,409Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:43,780Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][12] overhead, spent [388ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:52,923Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][21] overhead, spent [460ms] collecting in the last [1s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-24T15:59:55,229Z", "level": "INFO", "component": "o.e.m.j.JvmGcMonitorService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "[gc][23] overhead, spent [428ms] collecting in the last [1.2s]", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", ...
|
iTerm2
|
DOCKER (docker-compose)
|
NULL
|
79316
|
|
24441
|
iterm2ShellEditViewSessionScriptsProfilesWindowHel iterm2ShellEditViewSessionScriptsProfilesWindowHelp‹ →0 ll 0§ Sprint Review - in 39 mA100% (<47-zshDOCKERO $1DEV (-zsh)₴82APP (-zsh)83ec2-user@ip-10-... 884-zshlukas@Lukas-Kovaliks-MacBook-Pro-Jiminny~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "SELECT frame_id, app_name, window_name, browser_url, focused,substr(text, 1, 100) as text_previewFROM ocr_textORDER BY frame_id DESCLIMIT 5;*5-zsh86-zshError: in prepare, no such column: browser_urlSELECT frame_id, app_name, window_name, browser_url, focused,substrCerror here ---^lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ clearlukas®Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ ls ~/.screenpipe/data/ | head -20datapending-transcriptionsscreenpipe.dblukas®Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ ls ~/.screenpipe/data/S(ls ~/.screenpipe/data/ | head -1)/ | head -302026-04-092026-04-112026-04-122026-04-132026-04-142026-04-15lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ find ~/.screenpipe/data -name "*.jpg" | wc - 1146lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ find ~/.screenpipe/data -name "*.mp4" | wc -1526lukas®Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ find ~/.screenpipe/data -name "*.jpg" -exec du -sh (} + 2>/dev/null | tail -1256K/Users/lukas/.screenpipe/data/data/2026-04-15/1776255157610_m2. jpglukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ du -sh ~/.screenpipe/data/**/*.mp4 2>/dev/null | tail -12.0M/Users/lukas/.screenpipe/data/data/2026-04-15/compact_monitor_2_1776255640728.mp4lukas®Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $* Unable to a...O 88Wed 15 Apr 15:21:32-zshT81+...
|
NULL
|
NULL
|
NULL
|
24441
|
|
42690
|
iterm2ShellEditViewSessionScriptsProfilesDOCKERZbi iterm2ShellEditViewSessionScriptsProfilesDOCKERZbigniew OlesnickiThemistocleskovaliklukas (you)₴1570756765574DEV (-zsh)₴82WindowHelpS0 lhl* Review screenpipe usage and Boosteroid integrationAPP (-zsh)&3-zsha Backend Chapter • now-zsh®• ₴5100% (47 8• Fri 17 Apr 10:32:21T81* Review screenpipe usage and...*6You finished last in score among 8 players but were still on the winning team (Briton Victory). The streaming setup used soundcore AeroClip for audio output.give me timesod the day when I played form to if multiple sesions when• Basedon theframe data, here's the breakdown:Session 1 - ~10:04 (brief, ~2 min)- Boosteroid opened, showing a post-game rating screen ("Please rate your experience") - likely just launched to check something or wrap up a prior game- No gameplay frames, closed quicklyGap - 10:06 to 15:26 (~5 hours, no Boosteroid activity)Session 2 - 15:26 to 16:36 (~70 min window, ~47 min active)- 15:26 - Launched Boosteroid,browsing launcher- 15:51 - In-game, Dark Age (early game, paused briefly)- 15:55 - Active Dark Age gameplay (Dravidians)- 16:13 - Castle Age reached- 16:35 - Imperial Age- 16:36 - Game over, Briton Victory (match timer: 40:37)So in practice it was one real play session in the afternoon (~15:26-16:36), with a very brief app open in the morning at 10:04. The 49.3min total reflects the ~40 min match+ lobby time before and after.* Cogitated for 1m 4snice• Glad it was useful!* recap: You reviewed today's screenpipe activity focusing on Boosteroid. There was one real play session from 15:26-16:36 where you won an Age of Empires II match as Britons.No next action needed. (disable recaps in /config)> I? for shortcuts...
|
NULL
|
NULL
|
NULL
|
42690
|
|
42687
|
iterm2ShellEditViewSessionScriptsProfilesDOCKERZbi iterm2ShellEditViewSessionScriptsProfilesDOCKERZbigniew OlesnickiThemistocleskovaliklukas (you)₴1570756765574DEV (-zsh)₴82WindowHelp>0 lhl* Review screenpipe usage and Boosteroid integrationAPP (-zsh)&3-zsha Backend Chapter • now-zsh®• ₴5100% (47 8• Fri 17 Apr 10:32:20T*1* Review screenpipe usage and...X6Name: * Review screenpipe usage andBoosteroid integration (claude)Profile: DefaultCommand: loginYou finished last in score among 8 players but were still on the winning team (Briton Victory). The streaming setup used soundcore AeroClip for audio output.give me timesod the day when I played form to if multiple sesions when• Basedon theframe data, here's the breakdown:Session 1 - ~10:04 (brief, ~2 min)- Boosteroid opened, showing a post-game rating screen ("Please rate your experience") - likely just launched to check something or wrap up a prior game- No gameplay frames, closed quicklyGap - 10:06 to 15:26 (~5 hours, no Boosteroid activity)Session 2 - 15:26 to 16:36 (~70 min window, ~47 min active)- 15:26 - Launched Boosteroid,browsing launcher- 15:51 - In-game, Dark Age (early game, paused briefly)- 15:55 - Active Dark Age gameplay (Dravidians)- 16:13 - Castle Age reached- 16:35 - Imperial Age- 16:36 - Game over, Briton Victory (match timer: 40:37)So in practice it was one real play session in the afternoon (~15:26-16:36), with a very brief app open in the morning at 10:04. The 49.3min total reflects the ~40 min match+ lobby time before and after.* Cogitated for 1m 4snice• Glad it was useful!* recap: You reviewed today's screenpipe activity focusing on Boosteroid. There was one real play session from 15:26-16:36 where you won an Age of Empires II match as Britons.No next action needed. (disable recaps in /config)> I? for shortcuts...
|
NULL
|
NULL
|
NULL
|
42687
|
|
8471
|
irefoxFileEoitViewHistoryBookmarksProfilesToolsWin irefoxFileEoitViewHistoryBookmarksProfilesToolsWindow Helpjiminny.atlassian.net/jira/software/c/projects/JY/boards/37?selectedlssue=JY-19798JIMINNYQ SearchPlatform Sprint 1 Q2 - Platformg For you- New lao• Recent# Starred04 Apps0, SpacesRecent(2) Jiminny (New)I Œ Platform Teamspaces / Jiminny (NeW)Platform Team 884 JY-18679 / © JDevelopment JY-19798# SummaryTimelineB BаEvaluationBranches Commits Pull requests Builds Deployments Feature flags Other linksQ Search board |READY FOR DEV1T ***Uu s= KanbanInvestigate and fix why exceedFontawesome package limitsMAINTENANCEReady for Dev• JY-2056400 •000=y DescriptionA few months agoadding a descriptithis feature.CD Capture TeamD Enterprise Stability I...II Processing Team9 Service-Desk• create automa• use data from•make sure we• use a smart m= More spaces• then choose o• make sure the= FiltersB DashboardsC OperationsSubtasksAdd subtask(& Confluence88 Teams2= Customise sidebarv Linked work iteirelates toA JY-20510 Ccv ActivityCommerAdd a comment...Suggest a reply... Status update...Thanks...Pro tip: press M to commentNikyay Tvanovhttps://docs.google.com/spreadsheets/d/1QGD9igXbZDs6ydoT_zPZY8wK0ZBHAGzJHIIW1EzWIBQ/edit?gid=1867196262#gid=1867196262Nikolay Nik 1t 13:48 (edired)comments.• Similar to identify participants evaluation203+ Create& Give feedbacknical storyo days ago(OPEN]Priorityrix versionssorintvaysNeed QACanny LinksMediumNonePiauorm s orintt z+ JY-18679 Auto-detected actiyOpen Canny Links> More fields Original estimate, Time tracking> Automation 4kule executions› featureOs 5Open featureOS> 0| Daily - Platform • now100% CD8 • Tue 14 Apr 9:48:02Ask Rovo• DeploymentsE Archived work itemsMore4Complete sprintGroup: QueriesDEPLOY 8Fix downtime during to featurelide reiease(SUPPORT TICKETSDeployed** JY-20565AI Review - Q1GROWTH - MAINTAIN OUR COMPEIITI..Deployed# JY-19982[PASSWORD_DOTS]0=ino suen column "Ivoe on enuuy'Event'PLATIFORM STTABTLOMDeployed© JY-203840.5. •eee:Missed deletion of stale dealsrisksREDUCE CHURNClosedi#K JY-206191| 10 00ee=[Pipedrive] DI > status notlabeled, date data missingREDUCE CHURNDeployed[ JY-206140.5. 0000=OD/TI > Refresh the resultsWnen une Tielc is clearea wiunthe 'x'FE IMPROVEMENTSDeployedW JY-205/1os eeee=Fix lhubspoy upaatingassociationsREDUCE CHURNClosedS- JY-Z00212ф•0=...
|
NULL
|
NULL
|
NULL
|
8471
|
|
60662
|
install screenpipe - screenpipe docs
docs.screenpi install screenpipe - screenpipe docs
docs.screenpi.pe
DXP4800PLUS-B5F8
Статус на поръчка в stantek.com - [EMAIL] - Gmail
Stantek
Stantek
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX) от 241,72 € (472,76 лв.) Вътрешен хард диск Western Digital - Pazaruvaj.com
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX) от 241,72 € (472,76 лв.) Вътрешен хард диск Western Digital - Pazaruvaj.com
Close tab
| Senetic
| Senetic
Stantek
Stantek
| Senetic
| Senetic
SQLite Web: db.sqlite
SQLite Web: db.sqlite
Screenpipe — Archive
Screenpipe — Archive
[IP_ADDRESS]:8768/
[IP_ADDRESS]:8768/
Today's Deals
Today's Deals
Shameless • HBO Max
Shameless • HBO Max
install screenpipe - screenpipe docs
install screenpipe - screenpipe docs
Close tab
New Tab
New Tab
Screenpipe — Archive
Screenpipe — Archive
SQLite Web: db.sqlite
SQLite Web: db.sqlite
Claude Platform
Claude Platform
New Tab
Customize sidebar
Open Google Gemini (⌃X)
Open history (⇧⌘H)
Open bookmarks (⌘B)
Bitwarden
Pazaruvaj.com - Надежден помощник за покупки, сравнение на цени и продуктови оферти
Какво търсите?
Изтрий
Търсене
Търсене
Намаления
Намаления
Електроника
Електроника
Информатика
Информатика
Здраве и красота
Здраве и красота
Битова техника
Битова техника
Дом и градина
Дом и градина
Още (15)
Още (15)
Pazaruvaj.com
Pazaruvaj.com
Информатика
Информатика
Хардуер
Хардуер
Вътрешен хард диск
Вътрешен хард диск
Вътрешен хард диск Western Digital
Вътрешен хард диск Western Digital
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
previous slide / item
Image 1 of Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
Image 1 of Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
next slide / item
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
100
%
5 мнения
5
мнения
Производител: Western Digital Модел: Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX) Описание: Високопроизводител ни външни hdd дискове, създадени за надеждност и енергийна...
Пълно описание
Варианти на продукта:
Варианти на продукта
:
2000 GB
2000 GB
3000 GB
3000 GB
4000 GB
4000 GB
6000 GB
8000 GB
8000 GB
289,99 €
|
567,17 лв.
*
(2,59 €)
Доставка
5,07 лв.
,
В наличност
Препоръчана оферта
Plasico IT Superstore
към магазина
към магазина
Оферти с цени
Оферти с цени
Оферти с цени
Описание
Описание
Описание
Мнения (5)
Мнения (5)
Мнения
(
5
)
Препоръчани оферти
Препоръчани оферти
Най-евтини
Най-евтини
Всички филтри
Всички филтри
Препоръчани оферти (3)
Препоръчани оферти
(
3
)
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
към магазина Plasico IT Superstore
97%
(
9506
×)
В наличност
В наличност
1 отличие
1
отличие
Red Plus 6TB NAS, 3.5 , 256MB, 5400RPM | WD60EFPX (WD60EFPX)
към магазина Plasico IT Superstore
289,99 €
567,17 лв.
*
Доставка от 2,59 € (5,07 лв.)
Доставка
от
2,59 € (5,07 лв.)
към магазина Plasico IT Superstore
към магазина
към магазина PIC.bg - купи лаптоп изгодно
99%
(
8173
×)
В наличност
В наличност
1 отличие
1
отличие
WD Red Plus 6TB SATA 6Gb (WD60EFPX)
към магазина PIC.bg - купи лаптоп изгодно
299,00 €
584,79 лв.
*
Безплатна доставка
към магазина PIC.bg - купи лаптоп изгодно
към магазина
към магазина МегаМаг България
98%
(
4806
×)
В наличност
В наличност
1 отличие
1
отличие
Твърд диск WD Red Plus 6TB NAS HDD 3.5" SATA 6Gb/s 256MB (WD60EFPX)
към магазина МегаМаг България
292,04 €
571,18 лв.
*
към магазина МегаМаг България
към магазина
Най-евтини оферти (21)
Най-евтини оферти
(
21
)
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
към магазина Стантек
17
мнения
В наличност
В наличност
HDD NAS WD Red Plus (3.5'', 6TB, 256MB, 5400 RPM, SATA 6 Gb/s) - WD60EFPX (WD60EFPX)
към магазина Стантек
249,00 €
487,00 лв.
*
Безплатна доставка
към магазина Стантек
към магазина
към магазина Senetic
73
мнения
Информация в магазина
Информация в магазина
Red Plus 6TB WD60EFPX (WD60EFPX)
към магазина Senetic
250,14 €
489,23 лв.
*
Доставка от 9,82 € (19,21 лв.)
Доставка
от
9,82 € (19,21 лв.)
към магазина Senetic
към магазина
към магазина xtreme.bg
99%
(
5953
×)
В наличност
В наличност
1 отличие
1
отличие
Твърд диск, Western Digital Red 6TB Plus ( 3.5", 256MB, 5400 RPM, SATA 6Gb/s ) (WD60EFPX)
към магазина xtreme.bg
272,82 €
533,59 лв.
*
към магазина xtreme.bg
към магазина
към магазина SIGMATRONIC
3
мнения
В наличност
В наличност
HDD NAS WD Red Plus (3.5 (WD60EFPX)
към магазина SIGMATRONIC
272,82 €
533,59 лв.
*
към магазина SIGMATRONIC
към магазина
към магазина Optimal Computers
67
мнения
В наличност
В наличност
HDD NAS WD Red Plus (3.5'', 6TB, 256MB, 5400 RPM, SATA 6 Gb/ (WD60EFPX)
към магазина Optimal Computers
277,09 €
541,94 лв.
*
Безплатна доставка
към магазина Optimal Computers
към магазина
към магазина eZona.net
100%
(
725
×)
В наличност
В наличност
1 отличие
1
отличие
HDD NAS WD Red Plus (3.5'', 6TB, 256MB, 5400 RPM, SATA 6 Gb/s) (WD60EFPX)
към магазина eZona.net
278,59 €
544,87 лв.
*
към магазина eZona.net
към магазина
към магазина CyberTrade.bg
99%
(
2085
×)
В наличност
В наличност
1 отличие
1
отличие
Хард диск WD Red Plus, 6TB NAS, 3.5 (WD60EFPX)
към магазина CyberTrade.bg
278,88 €
545,44 лв.
*
към магазина CyberTrade.bg
към магазина
към магазина Plesio
83%
(
20 368
×)
Информация в магазина
Информация в магазина
1 отличие
1
отличие
HDD Red Plus 6TB 3.5 (WD60EFPX)
към магазина Plesio
279,98 €
547,59 лв.
*
Безплатна доставка
към магазина Plesio
към магазина
към магазина THX.bg
97%
(
1076
×)
В наличност
В наличност
1 отличие
1
отличие
WD Red Plus NAS 6TB (WD60EFPX)
към магазина THX.bg
284,00 €
555,46 лв.
*
Безплатна доставка
към магазина THX.bg
към магазина
към магазина pro-bg
100%
(
2351
×)
В наличност
В наличност
1 отличие
1
отличие
HDD NAS WD Red Plus (3.5'', 6TB, 256MB, 5400 RPM, SATA 6 Gb/s) (WD60EFPX)
към магазина pro-bg
289,36 €
565,94 лв.
*
Безплатна доставка
към магазина pro-bg
към магазина
към магазина Plasico IT Superstore
97%
(
9506
×)
В наличност
В наличност
1 отличие
1
отличие
Red Plus 6TB NAS, 3.5 , 256MB, 5400RPM | WD60EFPX (WD60EFPX)
към магазина Plasico IT Superstore
289,99 €
567,17 лв.
*
Доставка от 2,59 € (5,07 лв.)
Доставка
от
2,59 € (5,07 лв.)
към магазина Plasico IT Superstore
към магазина
към магазина DataCom Ltd
279
мнения
В наличност
В наличност
Red 6TB Plus ( 3.5 (WD60EFPX)
към магазина DataCom Ltd
292,00 €
571,10 лв.
*
към магазина DataCom Ltd
към магазина
към магазина PCBUILD.BG
602
мнения
Информация в магазина
Информация в магазина
Твърд диск WD Red Plus, 6TB NAS 3.5" 256MB 5400RPM - WD60EFPX (WD60EFPX)
към магазина PCBUILD.BG
292,00 €
571,10 лв.
*
към магазина PCBUILD.BG
към магазина
към магазина МегаМаг България
98%
(
4806
×)
В наличност
В наличност
1 отличие
1
отличие
Твърд диск WD Red Plus 6TB NAS HDD 3.5" SATA 6Gb/s 256MB (WD60EFPX)
към магазина МегаМаг България
292,04 €
571,18 лв.
*
към магазина МегаМаг България
към магазина
към магазина Technomall.BG
98%
(
633
×)
В наличност
В наличност
1 отличие
1
отличие
HDD 6TB WD Red Plus (WD60EFPX), 3.5", SATA 6Gb/s, 5400rpm, 256MB cache, for NAS (WD60EFPX)
към магазина Technomall.BG
293,00 €
573,06 лв.
*
към магазина Technomall.BG
към магазина
към магазина Йон компютърс
1503
мнения
В наличност
В наличност
Твърд диск Western Digital Red Plus, 6TB NAS, 3.5", 256MB, 5400RPM - WD60EFPX (WD60EFPX)
към магазина Йон компютърс
294,00 €
575,01 лв.
*
Безплатна доставка
към магазина Йон компютърс
към магазина
към магазина PIC.bg - купи лаптоп изгодно
99%
(
8173
×)
В наличност
В наличност
1 отличие
1
отличие
WD Red Plus 6TB SATA 6Gb (WD60EFPX)
към магазина PIC.bg - купи лаптоп изгодно
299,00 €
584,79 лв.
*
Безплатна доставка
към магазина PIC.bg - купи лаптоп изгодно
към магазина
към магазина BestPC.BG
99%
(
1266
×)
Информация в магазина
Информация в магазина
1 отличие
1
отличие
Red Plus, 6TB HDD NAS, SATA III, 5640rpm, 3.5 (WD60EFPX)
към магазина BestPC.BG
299,90 €
586,55 лв.
*
към магазина BestPC.BG
към магазина
към магазина Ardes.bg - национална верига за лаптопи и таблети
96%
(
5667
×)
В рамките на 3 дни
В рамките на 3 дни
1 отличие
1
отличие
6TB WD Red Plus, WD60EFPX (WD60EFPX)
към магазина Ardes.bg - национална верига за лаптопи и таблети
303,00 €
592,62 лв.
*
Безплатна доставка
към магазина Ardes.bg - национална верига за лаптопи и таблети
към магазина
към магазина ZettaByte
98%
(
312
×)
В наличност
В наличност
1 отличие
1
отличие
HDD NAS WD Red Plus (3.5'', 6TB, 256MB, 5400 RPM, SATA 6 Gb/s) (WD60EFPX)
към магазина ZettaByte
315,11 €
616,30 лв.
*
Безплатна доставка
към магазина ZettaByte
към магазина
Вижте и други предложения
Подобни продукти
Подобни продукти
TOP
1
Seagate IronWolf 4TB SATA3 (ST4000VN006)
Seagate IronWolf 4TB SATA3 (ST4000VN006)
от 175,94 € 344,11 лв.*
от
175,94 €
344,11 лв.
*
Сравнение
Сравнение
Seagate IronWolf 4TB SATA3 (ST4000VN006)
Seagate IronWolf 4TB SATA3 (ST4000VN006)
TOP
2
Western Digital Red Plus 3.5 4TB 5400rpm 256MB SATA (WD40EFPX)
Western Digital Red Plus 3.5 4TB 5400rpm 256MB SATA (WD40EFPX)
от 198,38 € 388,00 лв.*
от
198,38 €
388,00 лв.
*
Сравнение
Сравнение
Western Digital Red Plus 3.5 4TB 5400rpm 256MB SATA (WD40EFPX)
Western Digital Red Plus 3.5 4TB 5400rpm 256MB SATA (WD40EFPX)
TOP
3
Seagate IronWolf NAS 3.5 8TB 7200rpm 256MB SATA3 (ST8000VN004)
Seagate IronWolf NAS 3.5 8TB 7200rpm 256MB SATA3 (ST8000VN004)
от 324,99 € 635,63 лв.*
от
324,99 €
635,63 лв.
*
Сравнение
Сравнение
Seagate IronWolf NAS 3.5 8TB 7200rpm 256MB SATA3 (ST8000VN004)
Seagate IronWolf NAS 3.5 8TB 7200rpm 256MB SATA3 (ST8000VN004)
TOP
4
Seagate Surveillance 3.5 4TB 5400rpm 256MB SATA (ST4000VX016)
Seagate Surveillance 3.5 4TB 5400rpm 256MB SATA (ST4000VX016)
от 145,92 € 285,39 лв.*
от
145,92 €
285,39 лв.
*
Сравнение
Сравнение
Seagate Surveillance 3.5 4TB 5400rpm 256MB SATA (ST4000VX016)
Seagate Surveillance 3.5 4TB 5400rpm 256MB SATA (ST4000VX016)
TOP
5
Western Digital Purple 3.5 6TB 5400rpm SATA3 256MB (WD64PURZ)
Western Digital Purple 3.5 6TB 5400rpm SATA3 256MB (WD64PURZ)
от 229,00 € 447,89 лв.*
от
229,00 €
447,89 лв.
*
Сравнение
Сравнение
Western Digital Purple 3.5 6TB 5400rpm SATA3 256MB (WD64PURZ)
Western Digital Purple 3.5 6TB 5400rpm SATA3 256MB (WD64PURZ)
Реклама
Описание на продукта
Описание на продукта
Производител: Western Digital Модел: Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX) Описание: Високопроизводител ни външни hdd дискове, създадени за надеждност и енергийна...
Прочетете още
Параметри на продукта
Параметри на продукта
Производител
Western Digital
Капацитет
6000 GB
Скорост на диска
5400 rpm
Cache размер
256 MB
Максимален пренос на данни
6 Gbit/s
Всички параметри
Мнения
Мнения
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
Проверен клиент
Дими
добавен 16 февруари 2026 г.
, закупен в
Plasico IT Superstore
Plasico IT Superstore
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
Препоръчан продукт
100
%
Бърз, стабилен и много важно - доста добре обезшумен.
Супер. Бърз, стабилен и много важно - доста добре обезшумен.
Проверен клиент
anonymous
добавен 20 октомври 2025 г.
, закупен в
Plasico IT Superstore
Plasico IT Superstore
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
Препоръчан продукт
100
%
...
Проверен клиент
anonymous
добавен 15 декември 2024 г.
, закупен в
Plasico IT Superstore
Plasico IT Superstore
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX)
Препоръчан продукт
100
%
Изключително тих
Добър тих диск, на изгодна цена.
Всички мнения (5)
Всички мнения
(
5
)...
|
Firefox
|
Western Digital Red Plus 3.5 6TB 5400rpm 256MB SAT Western Digital Red Plus 3.5 6TB 5400rpm 256MB SATA3 (WD60EFPX) от 241,72 € (472,76 лв.) Вътрешен хард диск Western Digital - Pazaruvaj.com — Personal...
|
www.pazaruvaj.com/p/western-digital-red-plus-3-5-6 www.pazaruvaj.com/p/western-digital-red-plus-3-5-6tb-5400rpm-256mb-sata3-wd60efpx-pP883930389/#overviews/...
|
60662
|
|
52156
|
ins","taskManager","taskManager ins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:42Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:44Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:45Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:47Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:48Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:49Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:50Z","tags":["error","plugins","reporting","esqueue","queue-worker","error"],"pid":7,"message":"mo2l855600070bdf8650akqy - job querying failed: Error: No Living connections\n at sendReqWithConnection (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:266:15)\n at next (/usr/share/kibana/node_modules/elasticsearch/src/lib/connection_pool.js:243:7)\n at process._tickCallback (internal/process/next_tick.js:61:11)"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","elasticsearch","monitoring"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:51Z","tags":["warning","plugins","licensing"],"pid":7,"message":"License information could not be obtained from Elasticsearch due to Error: No Living connections error"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:53Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:55Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:56Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:58Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:26:59Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:00Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:02Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 |
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:03Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:02 Running ['artisan' meeting-bot:schedule-bot] ... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' meeting-bot:schedule-bot > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:05Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:04 Running ['artisan' dialers:monitor-activities] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' dialers:monitor-activities > '/proc/1/fd/1' 2>&1
docker_lamp_1 | 2026-04-18 13:27:05 Running ['artisan' jiminny:monitor-social-accounts] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' jiminny:monitor-social-accounts > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:06 Running ['artisan' mailbox:skip-lists:refresh] . 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:skip-lists:refresh > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:08Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
docker_lamp_1 | 2026-04-18 13:27:08 Running ['artisan' mailbox:batch:process --max-batches=15] 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:process --max-batches=15 > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:10Z","tags":["error","elasticsearch","data"],"pid":7,"message":"[ConnectionError]: getaddrinfo ENOTFOUND elasticsearch elasticsearch:9200"}
docker_lamp_1 | 2026-04-18 13:27:09 Running ['artisan' mailbox:batch:create] ....... 1s DONE
docker_lamp_1 | ⇂ '/usr/local/bin/php' 'artisan' mailbox:batch:create > '/proc/1/fd/1' 2>&1
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["warning","elasticsearch","data"],"pid":7,"message":"Unable to revive connection: [URL_WITH_CREDENTIALS] living connections"}
kibana | {"type":"log","@timestamp":"2026-04-18T13:27:11Z","tags":["error","plugins","taskManager","taskManager"],"pid":7,"message":"Failed to poll for work: Error: No Living connections"}
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/infrastructure/dev/docker (develop) $ work
WARN[0000] /Users/lukas/jiminny/infrastructure/dev/docker/docker-compose.yml: the attribute `version` is obsolete, it will be ignored, please remove it to avoid potential confusion
[+] Running 2/2
✔ Container redis Running 0.0s
✔ Container docker_lamp_1 Running 0.0s
Attaching to blackfire-1, datadog-1, jiminny_ext-1, mariadb-1, docker_lamp_1, elasticsearch, kibana, ngrok, redis
docker_lamp_1 |
docker_lamp_1 | Illuminate\Database\QueryException
docker_lamp_1 |
docker_lamp_1 | SQLSTATE[HY000] [2002] php_network_getaddresses: getaddrinfo for mariadb failed: Name or service not known (Connection: mysql, Host: mariadb, Port: 3306, Database: jiminny, SQL: select exists (select 1 from information_schema.tables where table_schema = schema() and table_name = 'migrations' and table_type in ('BASE TABLE', 'SYSTEM VERSIONED')) as `exists`)
docker_lamp_1 |
docker_lamp_1 | at vendor/laravel/framework/src/Illuminate/Database/Connection.php:838
docker_lamp_1 | 834▕ $exceptionType = $this->isUniqueConstraintError($e)
docker_lamp_1 | 835▕ ? UniqueConstraintViolationException::class
docker_lamp_1 | 836▕ : QueryException::class;
docker_lamp_1 | 837▕
docker_lamp_1 | ➜ 838▕ throw new $exceptionType(
docker_lamp_1 | 839▕ $this->getNameWithReadWriteType(),
docker_lamp_1 | 840▕ $query,
docker_lamp_1 | 841▕ $this->prepareBindings($bindings),
docker_lamp_1 | 842▕ $e,
docker_lamp_1 |
docker_lamp_1 | +42 vendor frames
docker_lamp_1 |
docker_lamp_1 | 43 artisan:13
docker_lamp_1 | Illuminate\Foundation\Application::handleCommand(Object(Symfony\Component\Console\Input\ArgvInput))
docker_lamp_1 |
mariadb-1 | 2026-04-20 06:38:50+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:11.4.5+maria~ubu2404 started.
blackfire-1 | [2026-04-20T06:38:51Z] ERROR: The server ID parameter is not set. Please run 'blackfire-agent -register' to configure it.
blackfire-1 | usage blackfire-agent [options]
blackfire-1 | --collector="[URL_WITH_CREDENTIALS] controller (64 bit): Version 7.10.2 (Build 40a3af639d4698) Copyright (c) 2020 Elasticsearch BV" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,521Z", "level": "INFO", "component": "o.e.t.NettyAllocator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "creating NettyAllocator with the following configs: [name=unpooled, suggested_max_allocation_size=256kb, factors={es.unsafe.use_unpooled_allocator=null, g1gc_enabled=true, g1gc_region_size=1mb, heap_size=700mb}]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:21,695Z", "level": "INFO", "component": "o.e.d.DiscoveryModule", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "using discovery type [single-node] and seed hosts providers [settings]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:22,462Z", "level": "WARN", "component": "o.e.g.DanglingIndicesState", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,302Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "initialized" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,303Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "starting ..." }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:23,565Z", "level": "INFO", "component": "o.e.t.TransportService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9300}, bound_addresses {[::]:9300}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,081Z", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "cluster UUID [8uh2w1CUSGyWYR_OvaKx6g]" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,252Z", "level": "INFO", "component": "o.e.c.s.MasterService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "elected-as-master ([1] nodes joined)[{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 223, version: 8391, delta: master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,401Z", "level": "INFO", "component": "o.e.c.s.ClusterApplierService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "master node changed {previous [], current [{e802ad473a4f}{e2ZKzgw4Q4aCf2w5ljWr1A}{X0gI228rTyiXo10jkhrkng}{[IP_ADDRESS]}{[IP_ADDRESS]:9300}{cdhilmrstw}{ml.machine_memory=4109217792, xpack.installed=true, transform.node=true, ml.max_open_jobs=20}]}, term: 223, version: 8391, reason: Publication{term=223, version=8391}" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.h.AbstractHttpServerTransport", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "publish_address {[IP_ADDRESS]:9200}, bound_addresses {[::]:9200}", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:24,521Z", "level": "INFO", "component": "o.e.n.Node", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "started", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,162Z", "level": "INFO", "component": "o.e.l.LicenseService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "license [85e882e5-5714-4173-a5dd-9baa841494a0] mode [basic] - valid", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:25,169Z", "level": "INFO", "component": "o.e.g.GatewayService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "recovered [15] indices into cluster_state", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
elasticsearch | {"type": "server", "timestamp": "2026-04-20T06:39:29,975Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "docker-cluster", "node.name": "e802ad473a4f", "message": "Cluster health status changed from [RED] to [YELLOW] (reason: [shards started [[activities_testing][0]]]).", "cluster.uuid": "8uh2w1CUSGyWYR_OvaKx6g", "node.id": "e2ZKzgw4Q4aCf2w5ljWr1A" }
docker_lamp_1 | + find /home/jiminny/storage -maxdepth 100 -exec chown jiminny:jiminny '{}' ';'
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:02Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"auditTrail\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-service"],"pid":7,"message":"Plugin \"visTypeXy\" is disabled."}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["warning","config","deprecation"],"pid":7,"message":"Config key [monitoring.cluster_alerts.email_notifications.email_address] will be required for email notifications to work in 8.0.\""}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:03Z","tags":["info","plugins-system"],"pid":7,"message":"Setting up [96] plugins: [taskManager,licensing,globalSearch,globalSearchProviders,code,mapsLegacy,usageCollection,xpackLegacy,telemetryCollectionManager,telemetry,telemetryCollectionXpack,kibanaUsageCollection,securityOss,kibanaLegacy,newsfeed,translations,share,legacyExport,embeddable,uiActionsEnhanced,expressions,data,home,observability,cloud,console,consoleExtensions,apmOss,searchprofiler,painlessLab,grokdebugger,management,indexPatternManagement,advancedSettings,fileUpload,savedObjects,dashboard,visualizations,visTypeVega,visTypeTimelion,timelion,features,upgradeAssistant,security,snapshotRestore,encryptedSavedObjects,ingestManager,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,enterpriseSearch,dashboardMode,beatsManagement,transform,ingestPipelines,maps,licenseManagement,graph,dataEnhanced,visTypeTable,visTypeMarkdown,tileMap,regionMap,inputControlVis,visualize,esUiShared,bfetch,canvas,charts,lens,visTypeTimeseries,rollup,visTypeTagcloud,visTypeVislib,visTypeMetric,watcher,discover,discoverEnhanced,savedObjectsManagement,spaces,reporting,eventLog,actions,case,alerts,stackAlerts,triggersActionsUi,lists,ml,securitySolution,infra,monitoring,logstash,apm,uptime]"}
docker_lamp_1 | + create_dot_env_local_file
docker_lamp_1 | + cp -f /home/jiminny/.env.local /home/jiminny/.env.local.bak
docker_lamp_1 | + create_dot_env
docker_lamp_1 | + [[ -f /home/jiminny/.env ]]
docker_lamp_1 | + return
docker_lamp_1 | + declare DB_ADMIN_PASSWORD
docker_lamp_1 | + declare DB_ADMIN_USERNAME
docker_lamp_1 | + declare DB_DEV_PASSWORD
docker_lamp_1 | + declare DB_DEV_USERNAME
docker_lamp_1 | + declare DB_ROOT_PASSWORD
docker_lamp_1 | + declare DB_ROOT_USERNAME
docker_lamp_1 | + declare DB_WEB_PASSWORD
docker_lamp_1 | + declare DB_WEB_USERNAME
docker_lamp_1 | ++ jq -r .DB_ADMIN_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ADMIN_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | ++ jq -r .DB_DEV_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_DEV_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_DEV_USERNAME=jmnydev
docker_lamp_1 | ++ jq -r .DB_ROOT_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_ROOT_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_ROOT_USERNAME=root
docker_lamp_1 | ++ jq -r .DB_WEB_PASSWORD /home/jiminny/dev.json
docker_lamp_1 | + [ENV_SECRET]
docker_lamp_1 | ++ jq -r .DB_WEB_USERNAME /home/jiminny/dev.json
docker_lamp_1 | + DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + readonly DB_ADMIN_PASSWORD
docker_lamp_1 | + readonly DB_ADMIN_USERNAME
docker_lamp_1 | + readonly DB_DEV_PASSWORD
docker_lamp_1 | + readonly DB_DEV_USERNAME
docker_lamp_1 | + readonly DB_ROOT_PASSWORD
docker_lamp_1 | + readonly DB_ROOT_USERNAME
docker_lamp_1 | + readonly DB_WEB_PASSWORD
docker_lamp_1 | + readonly DB_WEB_USERNAME
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=jmnyadmin~g' /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env
docker_lamp_1 | + sed -i -E 's~DB_HOST=.*$~DB_HOST=mariadb~g' /home/jiminny/.env
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.root
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.migrate
docker_lamp_1 | + sed -i -E 's~[ENV_SECRET] /home/jiminny/.env.root
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.migrate
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Generating a random key for xpack.security.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.security.encryptionKey in kibana.yml"}
docker_lamp_1 | + sed -i -E 's~DB_USERNAME=.*$~DB_USERNAME=root~g' /home/jiminny/.env.root
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","security","config"],"pid":7,"message":"Session cookies will be transmitted over insecure connections. This is not recommended."}
docker_lamp_1 | + cp -f /home/jiminny/.env /home/jiminny/.env.local
docker_lamp_1 | + echo ''
docker_lamp_1 | + echo '[ENV_SECRET]
docker_lamp_1 | + echo DB_ADMIN_USERNAME=jmnyadmin
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_DEV_USERNAME=jmnydev
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_ROOT_USERNAME=root
docker_lamp_1 | + echo [ENV_SECRET]
docker_lamp_1 | + echo DB_WEB_USERNAME=jmnyweb
docker_lamp_1 | + mv /home/jiminny/.env.local.bak /home/jiminny/.env.local
docker_lamp_1 | + declare IP_ADDRESS
docker_lamp_1 | ++ curl ifconfig.co/
docker_lamp_1 | % Total % Received % Xferd Average Speed Time Time Time Current
docker_lamp_1 | Dload Upload Total Spent Left Speed
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","encryptedSavedObjects","config"],"pid":7,"message":"Generating a random key for xpack.encryptedSavedObjects.encryptionKey. To be able to decrypt encrypted saved objects attributes after restart, please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml"}
kibana | {"type":"log","@timestamp":"2026-04-20T06:40:04Z","tags":["warning","plugins","ingestManager"],"pid":7,"message":"Fleet APIs are disabled due to the Encrypted Saved Objects plugin using an ephemeral encryption key. Please set xpack.encryptedSavedObjects.encryptionKey in kibana.yml."}
100 13 100 13 0 0 221 0 --:--:-- --:--:-- --:--:-- 224
docker_lamp_1 | + IP_ADDRESS=[IP_ADDRESS]
docker_lamp_1 | + readonly IP_ADDRESS
docker_lamp_1 | + sed -i -E 's~X-Forwarded-For "\$remote_addr"~X-Forwarded-For "IP_ADDRESS"~g' /etc/nginx/nginx.conf
docker_lamp_1 | + sed -i -E s~IP_ADDRESS~[IP_ADDRESS]~g /etc/nginx/nginx.conf
docker_lamp_1 | + perl -0777 -i -pe 's/# QA Only Begin \{([^#]*)# QA Only End \}//g' /etc/nginx/nginx.conf
docker_lamp_1 | + storage_init
docker_lamp_1 | + false
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/config.php ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/bootstrap/cache/packages.php ]]
docker_lamp_1 | + [[ ! -f /home/jiminny/...
|
iTerm2
|
EU (ssh)
|
NULL
|
52156
|
|
46616
|
info Visit [URL_WITH_CREDENTIALS] ~/jiminny/app/fr info Visit [URL_WITH_CREDENTIALS] ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yanrn build
zsh: command not found: yanrn
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn build
[dotenv@17.3.1] injecting env (0) from .env -- tip: ⚡️ secrets for agents: [URL_WITH_CREDENTIALS] ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn install
➤ YN0000: · Yarn 4.12.0
➤ YN0000: ┌ Resolution step
➤ YN0000: └ Completed
➤ YN0000: ┌ Post-resolution validation
➤ YN0002: │ root-workspace-0b6124@workspace:. doesn't provide @testing-library/dom (pea6ced), requested by @testing-library/user-event.
➤ YN0002: │ root-workspace-0b6124@workspace:. doesn't provide noty (p5bdb0d), requested by vue-components.
➤ YN0086: │ Some peer dependencies are incorrectly met by your project; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code.
➤ YN0086: │ Some peer dependencies are incorrectly met by dependencies; run yarn explain peer-requirements for details.
➤ YN0000: └ Completed
➤ YN0000: ┌ Fetch step
➤ YN0000: └ Completed in 1s 655ms
➤ YN0000: ┌ Link step
➤ YN0008: │ root-workspace-0b6124@workspace:. must be rebuilt because its dependency tree changed
➤ YN0000: └ Completed in 0s 969ms
➤ YN0000: · Done with warnings in 3s 14ms
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn build
[dotenv@17.3.1] injecting env (0) from .env -- tip: ⚙️ override existing env vars with { override: true }
vite v8.0.0 building client environment for production...
✓ 4656 modules transformed.
[sentry-vite-plugin] Warning: No auth token provided. Will not create release. Please set the `authToken` option. You can find information on how to generate a Sentry auth token here: https://docs.sentry.io/api/auth/
[sentry-vite-plugin] Warning: No auth token provided. Will not upload source maps. Please set the `authToken` option. You can find information on how to generate a Sentry auth token here: https://docs.sentry.io/api/auth/
computing gzip size...
../public/vue-assets/index.html 4.58 kB │ gzip: 1.15 kB
../public/vue-assets/assets/job-adder-B2LcgC4o.png 5.41 kB
../public/vue-assets/assets/dixa-DzT89tKh.png 7.87 kB
../public/vue-assets/assets/planhat-CQycOTMW.svg 8.61 kB │ gzip: 3.48 kB
../public/vue-assets/assets/funding-circle-9iGnDGsz.png 9.43 kB
../public/vue-assets/assets/cision-uOdx_YiN.svg 14.41 kB │ gzip: 6.32 kB
../public/vue-assets/assets/les-mills-DWabpmYc.png 15.73 kB
../public/vue-assets/assets/superside-btT37L9L.png 26.69 kB
../public/vue-assets/assets/flags-a2kmUSbF.webp 28.18 kB
../public/vue-assets/assets/quinyx-RGQNHx6o.png 63.17 kB
../public/vue-assets/assets/[EMAIL] 66.44 kB
../public/vue-assets/assets/bg-marketing-CAIHIJMj.png 70.66 kB
../public/vue-assets/.vite/manifest.json 85.73 kB │ gzip: 8.32 kB
../public/vue-assets/assets/jiminny-score-DtTbKZaD.svg 131.93 kB │ gzip: 86.07 kB
../public/vue-assets/assets/wavy-bg-vnCaCm9m.webm 670.97 kB
../public/vue-assets/assets/wavy-bg-Op6dRIjR.mp4 14,484.63 kB
../public/vue-assets/assets/AppForm-EEGPL5K8.css 0.04 kB │ gzip: 0.06 kB
../public/vue-assets/assets/RecipientsCell-DNOSS8k1.css 0.12 kB │ gzip: 0.12 kB
../public/vue-assets/assets/softphone-coach-Cy_9nuQZ.css 0.13 kB │ gzip: 0.10 kB
../public/vue-assets/assets/locked-CHS8VPur.css 0.15 kB │ gzip: 0.12 kB
../public/vue-assets/assets/mobile-KPnGhnTp.css 0.16 kB │ gzip: 0.14 kB
../public/vue-assets/assets/BuildInfo-CxNQv_OV.css 0.20 kB │ gzip: 0.17 kB
../public/vue-assets/assets/AvatarsStack-C0Nvfh1V.css 0.28 kB │ gzip: 0.20 kB
../public/vue-assets/assets/LogoLong100-ByQTvUY3.css 0.32 kB │ gzip: 0.20 kB
../public/vue-assets/assets/connect--xBDzLbc.css 0.34 kB │ gzip: 0.21 kB
../public/vue-assets/assets/KioskBanner-xU3FRd8g.css 0.41 kB │ gzip: 0.27 kB
../public/vue-assets/assets/extension-installed-BPBhG13P.css 0.41 kB │ gzip: 0.26 kB
../public/vue-assets/assets/FollowModal-A0rJmIOK.css 0.45 kB │ gzip: 0.23 kB
../public/vue-assets/assets/GenericMessage-DutAlObK.css 0.49 kB │ gzip: 0.26 kB
../public/vue-assets/assets/textarea-caret-DCiNUER_.css 0.51 kB │ gzip: 0.32 kB
../public/vue-assets/assets/AppLinks-BznfGcBR.css 0.58 kB │ gzip: 0.33 kB
../public/vue-assets/assets/SeekBtn-_VMPIk6v.css 0.63 kB │ gzip: 0.32 kB
../public/vue-assets/assets/emoji-input-kJWRbWf8.css 0.68 kB │ gzip: 0.38 kB
../public/vue-assets/assets/activity-preview-BDxP2S0-.css 0.70 kB │ gzip: 0.41 kB
../public/vue-assets/assets/ListLoader-DQqAw-8s.css 0.73 kB │ gzip: 0.38 kB
../public/vue-assets/assets/filters-Bgb6cQD6.css 0.75 kB │ gzip: 0.39 kB
../public/vue-assets/assets/AiContext-Bg-zjjbA.css 0.99 kB │ gzip: 0.45 kB
../public/vue-assets/assets/vue3-daterange-picker-izPOBj3P.css 1.02 kB │ gzip: 0.39 kB
../public/vue-assets/assets/PrimaryButton-CJj2FzKS.css 1.08 kB │ gzip: 0.34 kB
../public/vue-assets/assets/other-BTuPXtu_.css 1.15 kB │ gzip: 0.49 kB
../public/vue-assets/assets/jenesius-vue-modal-DDcfejHO.css 1.17 kB │ gzip: 0.48 kB
../public/vue-assets/assets/TabEmptyState-DOlIO68k.css 1.21 kB │ gzip: 0.53 kB
../public/vue-assets/assets/AppAlert-BlSJKiqj.css 1.29 kB │ gzip: 0.41 kB
../public/vue-assets/assets/useActivityCustomerName-CBjfIUfJ.css 1.42 kB │ gzip: 0.45 kB
../public/vue-assets/assets/UserAvatar-co8l9mMk.css 1.51 kB │ gzip: 0.49 kB
../public/vue-assets/assets/basic-modal-Dx7ZBbcA.css 1.55 kB │ gzip: 0.66 kB
../public/vue-assets/assets/login-DYbcBesM.css 1.58 kB │ gzip: 0.61 kB
../public/vue-assets/assets/DrawerWidget-poAmqspk.css 1.67 kB │ gzip: 0.69 kB
../public/vue-assets/assets/ActionItems-5FkZSo3G.css 1.79 kB │ gzip: 0.80 kB
../public/vue-assets/assets/join-conference-DQ6Yl7eU.css 1.82 kB │ gzip: 0.75 kB
../public/vue-assets/assets/InputText-C-DxDNSV.css 1.95 kB │ gzip: 0.71 kB
../public/vue-assets/assets/GoogleLikeButton-DvQHxbqR.css 2.11 kB │ gzip: 0.66 kB
../public/vue-assets/assets/usePusherEventListener-CDj8aSNp.css 2.20 kB │ gzip: 0.73 kB
../public/vue-assets/assets/onboard-IzGiQ-EC.css 2.32 kB │ gzip: 0.95 kB
../public/vue-assets/assets/ai-reports-manage-DaynTl62.css 2.34 kB │ gzip: 0.86 kB
../public/vue-assets/assets/StatusBadge-B4N12dzU.css 2.35 kB │ gzip: 0.86 kB
../public/vue-assets/assets/GridView-CsP1Jk-k.css 2.36 kB │ gzip: 0.90 kB
../public/vue-assets/assets/WelcomeLayout-6AX86p6F.css 2.38 kB │ gzip: 0.91 kB
../public/vue-assets/assets/snackbar-CS_iqvrv.css 2.45 kB │ gzip: 0.80 kB
../public/vue-assets/assets/ai-reports-mtqGqPPq.css 2.47 kB │ gzip: 0.91 kB
../public/vue-assets/assets/Comment-gmAfiuer.css 2.63 kB │ gzip: 1.00 kB
../public/vue-assets/assets/meeting-consent-BasRvxE3.css 2.66 kB │ gzip: 0.96 kB
../public/vue-assets/assets/DealRiskList-CwxmDnSC.css 2.67 kB │ gzip: 0.96 kB
../public/vue-assets/assets/add-to-playlist-modal-C_ad1hht.css 2.84 kB │ gzip: 0.95 kB
../public/vue-assets/assets/live-_zT5qBc1.css 3.60 kB │ gzip: 1.16 kB
../public/vue-assets/assets/RadioField-DDp8Y71G.css 3.64 kB │ gzip: 1.20 kB
../public/vue-assets/assets/vue-multiselect-BkQNV7oL.css 3.80 kB │ gzip: 0.86 kB
../public/vue-assets/assets/activity-preview-result-Db1Da81T.css 3.89 kB │ gzip: 1.15 kB
../public/vue-assets/assets/liquor-tree-CB5oh8v1.css 4.08 kB │ gzip: 1.24 kB
../public/vue-assets/assets/sentry-CSjHcvmn.css 4.47 kB │ gzip: 1.08 kB
../public/vue-assets/assets/AiAutomation-Cj1t8v5y.css 4.61 kB │ gzip: 0.84 kB
../public/vue-assets/assets/export-portal-CyIUVTEe.css 5.34 kB │ gzip: 1.65 kB
../public/vue-assets/assets/InputField-C4hoKJ7Z.css 6.13 kB │ gzip: 1.27 kB
../public/vue-assets/assets/vue-mq-bh4L87Tr.css 6.64 kB │ gzip: 1.95 kB
../public/vue-assets/assets/invitation-EPR1t-bH.css 6.74 kB │ gzip: 2.22 kB
../public/vue-assets/assets/ondemand-C2m0YVGA.css 6.84 kB │ gzip: 2.02 kB
../public/vue-assets/assets/AskAnything-CyvEhHS1.css 8.60 kB │ gzip: 2.60 kB
../public/vue-assets/assets/AppButton-BIeoar_U.css 8.88 kB │ gzip: 2.04 kB
../public/vue-assets/assets/kiosk-BcheyWWL.css 9.56 kB │ gzip: 2.53 kB
../public/vue-assets/assets/AiCrmNotes-nNTe_mOY.css 10.62 kB │ gzip: 1.93 kB
../public/vue-assets/assets/deal-view-DJzG7PLp.css 11.70 kB │ gzip: 3.29 kB
../public/vue-assets/assets/tokens-DZgWrfKd.css 12.44 kB │ gzip: 2.82 kB
../public/vue-assets/assets/tokens-DL2BPbai.css 12.89 kB │ gzip: 2.87 kB
../public/vue-assets/assets/playlists-Cy3t3kYU.css 13.54 kB │ gzip: 3.16 kB
../public/vue-assets/assets/PhoneField-D_kGOah0.css 13.75 kB │ gzip: 3.44 kB
../public/vue-assets/assets/ListView-Dt0qkyuY.css 17.07 kB │ gzip: 4.11 kB
../public/vue-assets/assets/intl-tel-input-DgmgTINs.css 17.11 kB │ gzip: 5.37 kB
../public/vue-assets/assets/AppFormField-DvpfRzi8.css 18.98 kB │ gzip: 4.42 kB
../public/vue-assets/assets/OrgSettingsLayout-BogdXtgY.css 21.98 kB │ gzip: 5.51 kB
../public/vue-assets/assets/dashboard-DQw3TKyk.css 22.74 kB │ gzip: 4.35 kB
../public/vue-assets/assets/deal-insights-CfX3UNzh.css 29.66 kB │ gzip: 6.44 kB
../public/vue-assets/assets/team-insights-CjVhm0JN.css 42.31 kB │ gzip: 8.46 kB
../public/vue-assets/assets/playback-CvJP5tX1.css 44.55 kB │ gzip: 10.36 kB
../public/vue-assets/assets/video-js-skin-DYZluGb-.css 47.73 kB │ gzip: 12.63 kB
../public/vue-assets/assets/assets-CAbfI4CY.css 83.38 kB │ gzip: 51.71 kB
../public/vue-assets/assets/logged-in-layout-1mAqkmnS.css 153.67 kB │ gzip: 27.18 kB
../public/vue-assets/assets/assets-xH6dx_9q.css 259.71 kB │ gzip: 181.71 kB
../public/vue-assets/assets/directives-DJJeJJOP.js 0.53 kB │ gzip: 0.36 kB │ map: 0.32 kB
../public/vue-assets/assets/jenesius-vue-modal-BuBhyl83.js 0.54 kB │ gzip: 0.37 kB │ map: 0.47 kB
../public/vue-assets/assets/spark-D_-Wgfar.js 0.54 kB │ gzip: 0.35 kB │ map: 0.40 kB
../public/vue-assets/assets/url-messenger-_CGQa-lH.js 0.56 kB │ gzip: 0.37 kB │ map: 0.46 kB
../public/vue-assets/assets/wavy-bg-Cec-_GDj.js 0.57 kB │ gzip: 0.36 kB │ map: 0.35 kB
../public/vue-assets/assets/component-css-class-CtO0AVgW.js 0.61 kB │ gzip: 0.40 kB │ map: 0.91 kB
../public/vue-assets/assets/theme-B5VdyMN_.js 0.63 kB │ gzip: 0.41 kB │ map: 0.49 kB
../public/vue-assets/assets/pick-C8Pw7kHx.js 0.71 kB │ gzip: 0.45 kB │ map: 1.59 kB
../public/vue-assets/assets/utils-BjsEoLDB.js 0.77 kB │ gzip: 0.50 kB │ map: 1.68 kB
../public/vue-assets/assets/throttle-siTOKEMt.js 0.78 kB │ gzip: 0.49 kB │ map: 3.24 kB
../public/vue-assets/assets/lastFilters-B_B_O5LY.js 0.80 kB │ gzip: 0.51 kB │ map: 1.12 kB
../public/vue-assets/assets/useAuthState-Bpx8WpBc.js 0.81 kB │ gz...
|
iTerm2
|
APP (-zsh)
|
NULL
|
46616
|
|
46617
|
info Visit [URL_WITH_CREDENTIALS] ~/jiminny/app/fr info Visit [URL_WITH_CREDENTIALS] ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yanrn build
zsh: command not found: yanrn
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn build
[dotenv@17.3.1] injecting env (0) from .env -- tip: ⚡️ secrets for agents: [URL_WITH_CREDENTIALS] ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn install
➤ YN0000: · Yarn 4.12.0
➤ YN0000: ┌ Resolution step
➤ YN0000: └ Completed
➤ YN0000: ┌ Post-resolution validation
➤ YN0002: │ root-workspace-0b6124@workspace:. doesn't provide @testing-library/dom (pea6ced), requested by @testing-library/user-event.
➤ YN0002: │ root-workspace-0b6124@workspace:. doesn't provide noty (p5bdb0d), requested by vue-components.
➤ YN0086: │ Some peer dependencies are incorrectly met by your project; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code.
➤ YN0086: │ Some peer dependencies are incorrectly met by dependencies; run yarn explain peer-requirements for details.
➤ YN0000: └ Completed
➤ YN0000: ┌ Fetch step
➤ YN0000: └ Completed in 1s 655ms
➤ YN0000: ┌ Link step
➤ YN0008: │ root-workspace-0b6124@workspace:. must be rebuilt because its dependency tree changed
➤ YN0000: └ Completed in 0s 969ms
➤ YN0000: · Done with warnings in 3s 14ms
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn build
[dotenv@17.3.1] injecting env (0) from .env -- tip: ⚙️ override existing env vars with { override: true }
vite v8.0.0 building client environment for production...
✓ 4656 modules transformed.
[sentry-vite-plugin] Warning: No auth token provided. Will not create release. Please set the `authToken` option. You can find information on how to generate a Sentry auth token here: https://docs.sentry.io/api/auth/
[sentry-vite-plugin] Warning: No auth token provided. Will not upload source maps. Please set the `authToken` option. You can find information on how to generate a Sentry auth token here: https://docs.sentry.io/api/auth/
computing gzip size...
../public/vue-assets/index.html 4.58 kB │ gzip: 1.15 kB
../public/vue-assets/assets/job-adder-B2LcgC4o.png 5.41 kB
../public/vue-assets/assets/dixa-DzT89tKh.png 7.87 kB
../public/vue-assets/assets/planhat-CQycOTMW.svg 8.61 kB │ gzip: 3.48 kB
../public/vue-assets/assets/funding-circle-9iGnDGsz.png 9.43 kB
../public/vue-assets/assets/cision-uOdx_YiN.svg 14.41 kB │ gzip: 6.32 kB
../public/vue-assets/assets/les-mills-DWabpmYc.png 15.73 kB
../public/vue-assets/assets/superside-btT37L9L.png 26.69 kB
../public/vue-assets/assets/flags-a2kmUSbF.webp 28.18 kB
../public/vue-assets/assets/quinyx-RGQNHx6o.png 63.17 kB
../public/vue-assets/assets/[EMAIL] 66.44 kB
../public/vue-assets/assets/bg-marketing-CAIHIJMj.png 70.66 kB
../public/vue-assets/.vite/manifest.json 85.73 kB │ gzip: 8.32 kB
../public/vue-assets/assets/jiminny-score-DtTbKZaD.svg 131.93 kB │ gzip: 86.07 kB
../public/vue-assets/assets/wavy-bg-vnCaCm9m.webm 670.97 kB
../public/vue-assets/assets/wavy-bg-Op6dRIjR.mp4 14,484.63 kB
../public/vue-assets/assets/AppForm-EEGPL5K8.css 0.04 kB │ gzip: 0.06 kB
../public/vue-assets/assets/RecipientsCell-DNOSS8k1.css 0.12 kB │ gzip: 0.12 kB
../public/vue-assets/assets/softphone-coach-Cy_9nuQZ.css 0.13 kB │ gzip: 0.10 kB
../public/vue-assets/assets/locked-CHS8VPur.css 0.15 kB │ gzip: 0.12 kB
../public/vue-assets/assets/mobile-KPnGhnTp.css 0.16 kB │ gzip: 0.14 kB
../public/vue-assets/assets/BuildInfo-CxNQv_OV.css 0.20 kB │ gzip: 0.17 kB
../public/vue-assets/assets/AvatarsStack-C0Nvfh1V.css 0.28 kB │ gzip: 0.20 kB
../public/vue-assets/assets/LogoLong100-ByQTvUY3.css 0.32 kB │ gzip: 0.20 kB
../public/vue-assets/assets/connect--xBDzLbc.css 0.34 kB │ gzip: 0.21 kB
../public/vue-assets/assets/KioskBanner-xU3FRd8g.css 0.41 kB │ gzip: 0.27 kB
../public/vue-assets/assets/extension-installed-BPBhG13P.css 0.41 kB │ gzip: 0.26 kB
../public/vue-assets/assets/FollowModal-A0rJmIOK.css 0.45 kB │ gzip: 0.23 kB
../public/vue-assets/assets/GenericMessage-DutAlObK.css 0.49 kB │ gzip: 0.26 kB
../public/vue-assets/assets/textarea-caret-DCiNUER_.css 0.51 kB │ gzip: 0.32 kB
../public/vue-assets/assets/AppLinks-BznfGcBR.css 0.58 kB │ gzip: 0.33 kB
../public/vue-assets/assets/SeekBtn-_VMPIk6v.css 0.63 kB │ gzip: 0.32 kB
../public/vue-assets/assets/emoji-input-kJWRbWf8.css 0.68 kB │ gzip: 0.38 kB
../public/vue-assets/assets/activity-preview-BDxP2S0-.css 0.70 kB │ gzip: 0.41 kB
../public/vue-assets/assets/ListLoader-DQqAw-8s.css 0.73 kB │ gzip: 0.38 kB
../public/vue-assets/assets/filters-Bgb6cQD6.css 0.75 kB │ gzip: 0.39 kB
../public/vue-assets/assets/AiContext-Bg-zjjbA.css 0.99 kB │ gzip: 0.45 kB
../public/vue-assets/assets/vue3-daterange-picker-izPOBj3P.css 1.02 kB │ gzip: 0.39 kB
../public/vue-assets/assets/PrimaryButton-CJj2FzKS.css 1.08 kB │ gzip: 0.34 kB
../public/vue-assets/assets/other-BTuPXtu_.css 1.15 kB │ gzip: 0.49 kB
../public/vue-assets/assets/jenesius-vue-modal-DDcfejHO.css 1.17 kB │ gzip: 0.48 kB
../public/vue-assets/assets/TabEmptyState-DOlIO68k.css 1.21 kB │ gzip: 0.53 kB
../public/vue-assets/assets/AppAlert-BlSJKiqj.css 1.29 kB │ gzip: 0.41 kB
../public/vue-assets/assets/useActivityCustomerName-CBjfIUfJ.css 1.42 kB │ gzip: 0.45 kB
../public/vue-assets/assets/UserAvatar-co8l9mMk.css 1.51 kB │ gzip: 0.49 kB
../public/vue-assets/assets/basic-modal-Dx7ZBbcA.css 1.55 kB │ gzip: 0.66 kB
../public/vue-assets/assets/login-DYbcBesM.css 1.58 kB │ gzip: 0.61 kB
../public/vue-assets/assets/DrawerWidget-poAmqspk.css 1.67 kB │ gzip: 0.69 kB
../public/vue-assets/assets/ActionItems-5FkZSo3G.css 1.79 kB │ gzip: 0.80 kB
../public/vue-assets/assets/join-conference-DQ6Yl7eU.css 1.82 kB │ gzip: 0.75 kB
../public/vue-assets/assets/InputText-C-DxDNSV.css 1.95 kB │ gzip: 0.71 kB
../public/vue-assets/assets/GoogleLikeButton-DvQHxbqR.css 2.11 kB │ gzip: 0.66 kB
../public/vue-assets/assets/usePusherEventListener-CDj8aSNp.css 2.20 kB │ gzip: 0.73 kB
../public/vue-assets/assets/onboard-IzGiQ-EC.css 2.32 kB │ gzip: 0.95 kB
../public/vue-assets/assets/ai-reports-manage-DaynTl62.css 2.34 kB │ gzip: 0.86 kB
../public/vue-assets/assets/StatusBadge-B4N12dzU.css 2.35 kB │ gzip: 0.86 kB
../public/vue-assets/assets/GridView-CsP1Jk-k.css 2.36 kB │ gzip: 0.90 kB
../public/vue-assets/assets/WelcomeLayout-6AX86p6F.css 2.38 kB │ gzip: 0.91 kB
../public/vue-assets/assets/snackbar-CS_iqvrv.css 2.45 kB │ gzip: 0.80 kB
../public/vue-assets/assets/ai-reports-mtqGqPPq.css 2.47 kB │ gzip: 0.91 kB
../public/vue-assets/assets/Comment-gmAfiuer.css 2.63 kB │ gzip: 1.00 kB
../public/vue-assets/assets/meeting-consent-BasRvxE3.css 2.66 kB │ gzip: 0.96 kB
../public/vue-assets/assets/DealRiskList-CwxmDnSC.css 2.67 kB │ gzip: 0.96 kB
../public/vue-assets/assets/add-to-playlist-modal-C_ad1hht.css 2.84 kB │ gzip: 0.95 kB
../public/vue-assets/assets/live-_zT5qBc1.css 3.60 kB │ gzip: 1.16 kB
../public/vue-assets/assets/RadioField-DDp8Y71G.css 3.64 kB │ gzip: 1.20 kB
../public/vue-assets/assets/vue-multiselect-BkQNV7oL.css 3.80 kB │ gzip: 0.86 kB
../public/vue-assets/assets/activity-preview-result-Db1Da81T.css 3.89 kB │ gzip: 1.15 kB
../public/vue-assets/assets/liquor-tree-CB5oh8v1.css 4.08 kB │ gzip: 1.24 kB
../public/vue-assets/assets/sentry-CSjHcvmn.css 4.47 kB │ gzip: 1.08 kB
../public/vue-assets/assets/AiAutomation-Cj1t8v5y.css 4.61 kB │ gzip: 0.84 kB
../public/vue-assets/assets/export-portal-CyIUVTEe.css 5.34 kB │ gzip: 1.65 kB
../public/vue-assets/assets/InputField-C4hoKJ7Z.css 6.13 kB │ gzip: 1.27 kB
../public/vue-assets/assets/vue-mq-bh4L87Tr.css 6.64 kB │ gzip: 1.95 kB
../public/vue-assets/assets/invitation-EPR1t-bH.css 6.74 kB │ gzip: 2.22 kB
../public/vue-assets/assets/ondemand-C2m0YVGA.css 6.84 kB │ gzip: 2.02 kB
../public/vue-assets/assets/AskAnything-CyvEhHS1.css 8.60 kB │ gzip: 2.60 kB
../public/vue-assets/assets/AppButton-BIeoar_U.css 8.88 kB │ gzip: 2.04 kB
../public/vue-assets/assets/kiosk-BcheyWWL.css 9.56 kB │ gzip: 2.53 kB
../public/vue-assets/assets/AiCrmNotes-nNTe_mOY.css 10.62 kB │ gzip: 1.93 kB
../public/vue-assets/assets/deal-view-DJzG7PLp.css 11.70 kB │ gzip: 3.29 kB
../public/vue-assets/assets/tokens-DZgWrfKd.css 12.44 kB │ gzip: 2.82 kB
../public/vue-assets/assets/tokens-DL2BPbai.css 12.89 kB │ gzip: 2.87 kB
../public/vue-assets/assets/playlists-Cy3t3kYU.css 13.54 kB │ gzip: 3.16 kB
../public/vue-assets/assets/PhoneField-D_kGOah0.css 13.75 kB │ gzip: 3.44 kB
../public/vue-assets/assets/ListView-Dt0qkyuY.css 17.07 kB │ gzip: 4.11 kB
../public/vue-assets/assets/intl-tel-input-DgmgTINs.css 17.11 kB │ gzip: 5.37 kB
../public/vue-assets/assets/AppFormField-DvpfRzi8.css 18.98 kB │ gzip: 4.42 kB
../public/vue-assets/assets/OrgSettingsLayout-BogdXtgY.css 21.98 kB │ gzip: 5.51 kB
../public/vue-assets/assets/dashboard-DQw3TKyk.css 22.74 kB │ gzip: 4.35 kB
../public/vue-assets/assets/deal-insights-CfX3UNzh.css 29.66 kB │ gzip: 6.44 kB
../public/vue-assets/assets/team-insights-CjVhm0JN.css 42.31 kB │ gzip: 8.46 kB
../public/vue-assets/assets/playback-CvJP5tX1.css 44.55 kB │ gzip: 10.36 kB
../public/vue-assets/assets/video-js-skin-DYZluGb-.css 47.73 kB │ gzip: 12.63 kB
../public/vue-assets/assets/assets-CAbfI4CY.css 83.38 kB │ gzip: 51.71 kB
../public/vue-assets/assets/logged-in-layout-1mAqkmnS.css 153.67 kB │ gzip: 27.18 kB
../public/vue-assets/assets/assets-xH6dx_9q.css 259.71 kB │ gzip: 181.71 kB
../public/vue-assets/assets/directives-DJJeJJOP.js 0.53 kB │ gzip: 0.36 kB │ map: 0.32 kB
../public/vue-assets/assets/jenesius-vue-modal-BuBhyl83.js 0.54 kB │ gzip: 0.37 kB │ map: 0.47 kB
../public/vue-assets/assets/spark-D_-Wgfar.js 0.54 kB │ gzip: 0.35 kB │ map: 0.40 kB
../public/vue-assets/assets/url-messenger-_CGQa-lH.js 0.56 kB │ gzip: 0.37 kB │ map: 0.46 kB
../public/vue-assets/assets/wavy-bg-Cec-_GDj.js 0.57 kB │ gzip: 0.36 kB │ map: 0.35 kB
../public/vue-assets/assets/component-css-class-CtO0AVgW.js 0.61 kB │ gzip: 0.40 kB │ map: 0.91 kB
../public/vue-assets/assets/theme-B5VdyMN_.js 0.63 kB │ gzip: 0.41 kB │ map: 0.49 kB
../public/vue-assets/assets/pick-C8Pw7kHx.js 0.71 kB │ gzip: 0.45 kB │ map: 1.59 kB
../public/vue-assets/assets/utils-BjsEoLDB.js 0.77 kB │ gzip: 0.50 kB │ map: 1.68 kB
../public/vue-assets/assets/throttle-siTOKEMt.js 0.78 kB │ gzip: 0.49 kB │ map: 3.24 kB
../public/vue-assets/assets/lastFilters-B_B_O5LY.js 0.80 kB │ gzip: 0.51 kB │ map: 1.12 kB
../public/vue-assets/assets/useAuthState-Bpx8WpBc.js 0.81 kB │ gz...
|
iTerm2
|
APP (-zsh)
|
NULL
|
46617
|
|
46881
|
info Visit [URL_WITH_CREDENTIALS] ~/jiminny/app/fr info Visit [URL_WITH_CREDENTIALS] ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yanrn build
zsh: command not found: yanrn
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn build
[dotenv@17.3.1] injecting env (0) from .env -- tip: ⚡️ secrets for agents: [URL_WITH_CREDENTIALS] ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn install
➤ YN0000: · Yarn 4.12.0
➤ YN0000: ┌ Resolution step
➤ YN0000: └ Completed
➤ YN0000: ┌ Post-resolution validation
➤ YN0002: │ root-workspace-0b6124@workspace:. doesn't provide @testing-library/dom (pea6ced), requested by @testing-library/user-event.
➤ YN0002: │ root-workspace-0b6124@workspace:. doesn't provide noty (p5bdb0d), requested by vue-components.
➤ YN0086: │ Some peer dependencies are incorrectly met by your project; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code.
➤ YN0086: │ Some peer dependencies are incorrectly met by dependencies; run yarn explain peer-requirements for details.
➤ YN0000: └ Completed
➤ YN0000: ┌ Fetch step
➤ YN0000: └ Completed in 1s 655ms
➤ YN0000: ┌ Link step
➤ YN0008: │ root-workspace-0b6124@workspace:. must be rebuilt because its dependency tree changed
➤ YN0000: └ Completed in 0s 969ms
➤ YN0000: · Done with warnings in 3s 14ms
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn build
[dotenv@17.3.1] injecting env (0) from .env -- tip: ⚙️ override existing env vars with { override: true }
vite v8.0.0 building client environment for production...
✓ 4656 modules transformed.
[sentry-vite-plugin] Warning: No auth token provided. Will not create release. Please set the `authToken` option. You can find information on how to generate a Sentry auth token here: https://docs.sentry.io/api/auth/
[sentry-vite-plugin] Warning: No auth token provided. Will not upload source maps. Please set the `authToken` option. You can find information on how to generate a Sentry auth token here: https://docs.sentry.io/api/auth/
computing gzip size...
../public/vue-assets/index.html 4.58 kB │ gzip: 1.15 kB
../public/vue-assets/assets/job-adder-B2LcgC4o.png 5.41 kB
../public/vue-assets/assets/dixa-DzT89tKh.png 7.87 kB
../public/vue-assets/assets/planhat-CQycOTMW.svg 8.61 kB │ gzip: 3.48 kB
../public/vue-assets/assets/funding-circle-9iGnDGsz.png 9.43 kB
../public/vue-assets/assets/cision-uOdx_YiN.svg 14.41 kB │ gzip: 6.32 kB
../public/vue-assets/assets/les-mills-DWabpmYc.png 15.73 kB
../public/vue-assets/assets/superside-btT37L9L.png 26.69 kB
../public/vue-assets/assets/flags-a2kmUSbF.webp 28.18 kB
../public/vue-assets/assets/quinyx-RGQNHx6o.png 63.17 kB
../public/vue-assets/assets/[EMAIL] 66.44 kB
../public/vue-assets/assets/bg-marketing-CAIHIJMj.png 70.66 kB
../public/vue-assets/.vite/manifest.json 85.73 kB │ gzip: 8.32 kB
../public/vue-assets/assets/jiminny-score-DtTbKZaD.svg 131.93 kB │ gzip: 86.07 kB
../public/vue-assets/assets/wavy-bg-vnCaCm9m.webm 670.97 kB
../public/vue-assets/assets/wavy-bg-Op6dRIjR.mp4 14,484.63 kB
../public/vue-assets/assets/AppForm-EEGPL5K8.css 0.04 kB │ gzip: 0.06 kB
../public/vue-assets/assets/RecipientsCell-DNOSS8k1.css 0.12 kB │ gzip: 0.12 kB
../public/vue-assets/assets/softphone-coach-Cy_9nuQZ.css 0.13 kB │ gzip: 0.10 kB
../public/vue-assets/assets/locked-CHS8VPur.css 0.15 kB │ gzip: 0.12 kB
../public/vue-assets/assets/mobile-KPnGhnTp.css 0.16 kB │ gzip: 0.14 kB
../public/vue-assets/assets/BuildInfo-CxNQv_OV.css 0.20 kB │ gzip: 0.17 kB
../public/vue-assets/assets/AvatarsStack-C0Nvfh1V.css 0.28 kB │ gzip: 0.20 kB
../public/vue-assets/assets/LogoLong100-ByQTvUY3.css 0.32 kB │ gzip: 0.20 kB
../public/vue-assets/assets/connect--xBDzLbc.css 0.34 kB │ gzip: 0.21 kB
../public/vue-assets/assets/KioskBanner-xU3FRd8g.css 0.41 kB │ gzip: 0.27 kB
../public/vue-assets/assets/extension-installed-BPBhG13P.css 0.41 kB │ gzip: 0.26 kB
../public/vue-assets/assets/FollowModal-A0rJmIOK.css 0.45 kB │ gzip: 0.23 kB
../public/vue-assets/assets/GenericMessage-DutAlObK.css 0.49 kB │ gzip: 0.26 kB
../public/vue-assets/assets/textarea-caret-DCiNUER_.css 0.51 kB │ gzip: 0.32 kB
../public/vue-assets/assets/AppLinks-BznfGcBR.css 0.58 kB │ gzip: 0.33 kB
../public/vue-assets/assets/SeekBtn-_VMPIk6v.css 0.63 kB │ gzip: 0.32 kB
../public/vue-assets/assets/emoji-input-kJWRbWf8.css 0.68 kB │ gzip: 0.38 kB
../public/vue-assets/assets/activity-preview-BDxP2S0-.css 0.70 kB │ gzip: 0.41 kB
../public/vue-assets/assets/ListLoader-DQqAw-8s.css 0.73 kB │ gzip: 0.38 kB
../public/vue-assets/assets/filters-Bgb6cQD6.css 0.75 kB │ gzip: 0.39 kB
../public/vue-assets/assets/AiContext-Bg-zjjbA.css 0.99 kB │ gzip: 0.45 kB
../public/vue-assets/assets/vue3-daterange-picker-izPOBj3P.css 1.02 kB │ gzip: 0.39 kB
../public/vue-assets/assets/PrimaryButton-CJj2FzKS.css 1.08 kB │ gzip: 0.34 kB
../public/vue-assets/assets/other-BTuPXtu_.css 1.15 kB │ gzip: 0.49 kB
../public/vue-assets/assets/jenesius-vue-modal-DDcfejHO.css 1.17 kB │ gzip: 0.48 kB
../public/vue-assets/assets/TabEmptyState-DOlIO68k.css 1.21 kB │ gzip: 0.53 kB
../public/vue-assets/assets/AppAlert-BlSJKiqj.css 1.29 kB │ gzip: 0.41 kB
../public/vue-assets/assets/useActivityCustomerName-CBjfIUfJ.css 1.42 kB │ gzip: 0.45 kB
../public/vue-assets/assets/UserAvatar-co8l9mMk.css 1.51 kB │ gzip: 0.49 kB
../public/vue-assets/assets/basic-modal-Dx7ZBbcA.css 1.55 kB │ gzip: 0.66 kB
../public/vue-assets/assets/login-DYbcBesM.css 1.58 kB │ gzip: 0.61 kB
../public/vue-assets/assets/DrawerWidget-poAmqspk.css 1.67 kB │ gzip: 0.69 kB
../public/vue-assets/assets/ActionItems-5FkZSo3G.css 1.79 kB │ gzip: 0.80 kB
../public/vue-assets/assets/join-conference-DQ6Yl7eU.css 1.82 kB │ gzip: 0.75 kB
../public/vue-assets/assets/InputText-C-DxDNSV.css 1.95 kB │ gzip: 0.71 kB
../public/vue-assets/assets/GoogleLikeButton-DvQHxbqR.css 2.11 kB │ gzip: 0.66 kB
../public/vue-assets/assets/usePusherEventListener-CDj8aSNp.css 2.20 kB │ gzip: 0.73 kB
../public/vue-assets/assets/onboard-IzGiQ-EC.css 2.32 kB │ gzip: 0.95 kB
../public/vue-assets/assets/ai-reports-manage-DaynTl62.css 2.34 kB │ gzip: 0.86 kB
../public/vue-assets/assets/StatusBadge-B4N12dzU.css 2.35 kB │ gzip: 0.86 kB
../public/vue-assets/assets/GridView-CsP1Jk-k.css 2.36 kB │ gzip: 0.90 kB
../public/vue-assets/assets/WelcomeLayout-6AX86p6F.css 2.38 kB │ gzip: 0.91 kB
../public/vue-assets/assets/snackbar-CS_iqvrv.css 2.45 kB │ gzip: 0.80 kB
../public/vue-assets/assets/ai-reports-mtqGqPPq.css 2.47 kB │ gzip: 0.91 kB
../public/vue-assets/assets/Comment-gmAfiuer.css 2.63 kB │ gzip: 1.00 kB
../public/vue-assets/assets/meeting-consent-BasRvxE3.css 2.66 kB │ gzip: 0.96 kB
../public/vue-assets/assets/DealRiskList-CwxmDnSC.css 2.67 kB │ gzip: 0.96 kB
../public/vue-assets/assets/add-to-playlist-modal-C_ad1hht.css 2.84 kB │ gzip: 0.95 kB
../public/vue-assets/assets/live-_zT5qBc1.css 3.60 kB │ gzip: 1.16 kB
../public/vue-assets/assets/RadioField-DDp8Y71G.css 3.64 kB │ gzip: 1.20 kB
../public/vue-assets/assets/vue-multiselect-BkQNV7oL.css 3.80 kB │ gzip: 0.86 kB
../public/vue-assets/assets/activity-preview-result-Db1Da81T.css 3.89 kB │ gzip: 1.15 kB
../public/vue-assets/assets/liquor-tree-CB5oh8v1.css 4.08 kB │ gzip: 1.24 kB
../public/vue-assets/assets/sentry-CSjHcvmn.css 4.47 kB │ gzip: 1.08 kB
../public/vue-assets/assets/AiAutomation-Cj1t8v5y.css 4.61 kB │ gzip: 0.84 kB
../public/vue-assets/assets/export-portal-CyIUVTEe.css 5.34 kB │ gzip: 1.65 kB
../public/vue-assets/assets/InputField-C4hoKJ7Z.css 6.13 kB │ gzip: 1.27 kB
../public/vue-assets/assets/vue-mq-bh4L87Tr.css 6.64 kB │ gzip: 1.95 kB
../public/vue-assets/assets/invitation-EPR1t-bH.css 6.74 kB │ gzip: 2.22 kB
../public/vue-assets/assets/ondemand-C2m0YVGA.css 6.84 kB │ gzip: 2.02 kB
../public/vue-assets/assets/AskAnything-CyvEhHS1.css 8.60 kB │ gzip: 2.60 kB
../public/vue-assets/assets/AppButton-BIeoar_U.css 8.88 kB │ gzip: 2.04 kB
../public/vue-assets/assets/kiosk-BcheyWWL.css 9.56 kB │ gzip: 2.53 kB
../public/vue-assets/assets/AiCrmNotes-nNTe_mOY.css 10.62 kB │ gzip: 1.93 kB
../public/vue-assets/assets/deal-view-DJzG7PLp.css 11.70 kB │ gzip: 3.29 kB
../public/vue-assets/assets/tokens-DZgWrfKd.css 12.44 kB │ gzip: 2.82 kB
../public/vue-assets/assets/tokens-DL2BPbai.css 12.89 kB │ gzip: 2.87 kB
../public/vue-assets/assets/playlists-Cy3t3kYU.css 13.54 kB │ gzip: 3.16 kB
../public/vue-assets/assets/PhoneField-D_kGOah0.css 13.75 kB │ gzip: 3.44 kB
../public/vue-assets/assets/ListView-Dt0qkyuY.css 17.07 kB │ gzip: 4.11 kB
../public/vue-assets/assets/intl-tel-input-DgmgTINs.css 17.11 kB │ gzip: 5.37 kB
../public/vue-assets/assets/AppFormField-DvpfRzi8.css 18.98 kB │ gzip: 4.42 kB
../public/vue-assets/assets/OrgSettingsLayout-BogdXtgY.css 21.98 kB │ gzip: 5.51 kB
../public/vue-assets/assets/dashboard-DQw3TKyk.css 22.74 kB │ gzip: 4.35 kB
../public/vue-assets/assets/deal-insights-CfX3UNzh.css 29.66 kB │ gzip: 6.44 kB
../public/vue-assets/assets/team-insights-CjVhm0JN.css 42.31 kB │ gzip: 8.46 kB
../public/vue-assets/assets/playback-CvJP5tX1.css 44.55 kB │ gzip: 10.36 kB
../public/vue-assets/assets/video-js-skin-DYZluGb-.css 47.73 kB │ gzip: 12.63 kB
../public/vue-assets/assets/assets-CAbfI4CY.css 83.38 kB │ gzip: 51.71 kB
../public/vue-assets/assets/logged-in-layout-1mAqkmnS.css 153.67 kB │ gzip: 27.18 kB
../public/vue-assets/assets/assets-xH6dx_9q.css 259.71 kB │ gzip: 181.71 kB
../public/vue-assets/assets/directives-DJJeJJOP.js 0.53 kB │ gzip: 0.36 kB │ map: 0.32 kB
../public/vue-assets/assets/jenesius-vue-modal-BuBhyl83.js 0.54 kB │ gzip: 0.37 kB │ map: 0.47 kB
../public/vue-assets/assets/spark-D_-Wgfar.js 0.54 kB │ gzip: 0.35 kB │ map: 0.40 kB
../public/vue-assets/assets/url-messenger-_CGQa-lH.js 0.56 kB │ gzip: 0.37 kB │ map: 0.46 kB
../public/vue-assets/assets/wavy-bg-Cec-_GDj.js 0.57 kB │ gzip: 0.36 kB │ map: 0.35 kB
../public/vue-assets/assets/component-css-class-CtO0AVgW.js 0.61 kB │ gzip: 0.40 kB │ map: 0.91 kB
../public/vue-assets/assets/theme-B5VdyMN_.js 0.63 kB │ gzip: 0.41 kB │ map: 0.49 kB
../public/vue-assets/assets/pick-C8Pw7kHx.js 0.71 kB │ gzip: 0.45 kB │ map: 1.59 kB
../public/vue-assets/assets/utils-BjsEoLDB.js 0.77 kB │ gzip: 0.50 kB │ map: 1.68 kB
../public/vue-assets/assets/throttle-siTOKEMt.js 0.78 kB │ gzip: 0.49 kB │ map: 3.24 kB
../public/vue-assets/assets/lastFilters-B_B_O5LY.js 0.80 kB │ gzip: 0.51 kB │ map: 1.12 kB
../public/vue-assets/assets/useAuthState-Bpx8WpBc.js 0.81 kB │ gz...
|
iTerm2
|
APP (-zsh)
|
NULL
|
46881
|
|
46882
|
info Visit [URL_WITH_CREDENTIALS] ~/jiminny/app/fr info Visit [URL_WITH_CREDENTIALS] ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yanrn build
zsh: command not found: yanrn
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn build
[dotenv@17.3.1] injecting env (0) from .env -- tip: ⚡️ secrets for agents: [URL_WITH_CREDENTIALS] ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn install
➤ YN0000: · Yarn 4.12.0
➤ YN0000: ┌ Resolution step
➤ YN0000: └ Completed
➤ YN0000: ┌ Post-resolution validation
➤ YN0002: │ root-workspace-0b6124@workspace:. doesn't provide @testing-library/dom (pea6ced), requested by @testing-library/user-event.
➤ YN0002: │ root-workspace-0b6124@workspace:. doesn't provide noty (p5bdb0d), requested by vue-components.
➤ YN0086: │ Some peer dependencies are incorrectly met by your project; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code.
➤ YN0086: │ Some peer dependencies are incorrectly met by dependencies; run yarn explain peer-requirements for details.
➤ YN0000: └ Completed
➤ YN0000: ┌ Fetch step
➤ YN0000: └ Completed in 1s 655ms
➤ YN0000: ┌ Link step
➤ YN0008: │ root-workspace-0b6124@workspace:. must be rebuilt because its dependency tree changed
➤ YN0000: └ Completed in 0s 969ms
➤ YN0000: · Done with warnings in 3s 14ms
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/jiminny/app/front-end (JY-20692-fix-integration-app-[API_KEY]) $ yarn build
[dotenv@17.3.1] injecting env (0) from .env -- tip: ⚙️ override existing env vars with { override: true }
vite v8.0.0 building client environment for production...
✓ 4656 modules transformed.
[sentry-vite-plugin] Warning: No auth token provided. Will not create release. Please set the `authToken` option. You can find information on how to generate a Sentry auth token here: https://docs.sentry.io/api/auth/
[sentry-vite-plugin] Warning: No auth token provided. Will not upload source maps. Please set the `authToken` option. You can find information on how to generate a Sentry auth token here: https://docs.sentry.io/api/auth/
computing gzip size...
../public/vue-assets/index.html 4.58 kB │ gzip: 1.15 kB
../public/vue-assets/assets/job-adder-B2LcgC4o.png 5.41 kB
../public/vue-assets/assets/dixa-DzT89tKh.png 7.87 kB
../public/vue-assets/assets/planhat-CQycOTMW.svg 8.61 kB │ gzip: 3.48 kB
../public/vue-assets/assets/funding-circle-9iGnDGsz.png 9.43 kB
../public/vue-assets/assets/cision-uOdx_YiN.svg 14.41 kB │ gzip: 6.32 kB
../public/vue-assets/assets/les-mills-DWabpmYc.png 15.73 kB
../public/vue-assets/assets/superside-btT37L9L.png 26.69 kB
../public/vue-assets/assets/flags-a2kmUSbF.webp 28.18 kB
../public/vue-assets/assets/quinyx-RGQNHx6o.png 63.17 kB
../public/vue-assets/assets/[EMAIL] 66.44 kB
../public/vue-assets/assets/bg-marketing-CAIHIJMj.png 70.66 kB
../public/vue-assets/.vite/manifest.json 85.73 kB │ gzip: 8.32 kB
../public/vue-assets/assets/jiminny-score-DtTbKZaD.svg 131.93 kB │ gzip: 86.07 kB
../public/vue-assets/assets/wavy-bg-vnCaCm9m.webm 670.97 kB
../public/vue-assets/assets/wavy-bg-Op6dRIjR.mp4 14,484.63 kB
../public/vue-assets/assets/AppForm-EEGPL5K8.css 0.04 kB │ gzip: 0.06 kB
../public/vue-assets/assets/RecipientsCell-DNOSS8k1.css 0.12 kB │ gzip: 0.12 kB
../public/vue-assets/assets/softphone-coach-Cy_9nuQZ.css 0.13 kB │ gzip: 0.10 kB
../public/vue-assets/assets/locked-CHS8VPur.css 0.15 kB │ gzip: 0.12 kB
../public/vue-assets/assets/mobile-KPnGhnTp.css 0.16 kB │ gzip: 0.14 kB
../public/vue-assets/assets/BuildInfo-CxNQv_OV.css 0.20 kB │ gzip: 0.17 kB
../public/vue-assets/assets/AvatarsStack-C0Nvfh1V.css 0.28 kB │ gzip: 0.20 kB
../public/vue-assets/assets/LogoLong100-ByQTvUY3.css 0.32 kB │ gzip: 0.20 kB
../public/vue-assets/assets/connect--xBDzLbc.css 0.34 kB │ gzip: 0.21 kB
../public/vue-assets/assets/KioskBanner-xU3FRd8g.css 0.41 kB │ gzip: 0.27 kB
../public/vue-assets/assets/extension-installed-BPBhG13P.css 0.41 kB │ gzip: 0.26 kB
../public/vue-assets/assets/FollowModal-A0rJmIOK.css 0.45 kB │ gzip: 0.23 kB
../public/vue-assets/assets/GenericMessage-DutAlObK.css 0.49 kB │ gzip: 0.26 kB
../public/vue-assets/assets/textarea-caret-DCiNUER_.css 0.51 kB │ gzip: 0.32 kB
../public/vue-assets/assets/AppLinks-BznfGcBR.css 0.58 kB │ gzip: 0.33 kB
../public/vue-assets/assets/SeekBtn-_VMPIk6v.css 0.63 kB │ gzip: 0.32 kB
../public/vue-assets/assets/emoji-input-kJWRbWf8.css 0.68 kB │ gzip: 0.38 kB
../public/vue-assets/assets/activity-preview-BDxP2S0-.css 0.70 kB │ gzip: 0.41 kB
../public/vue-assets/assets/ListLoader-DQqAw-8s.css 0.73 kB │ gzip: 0.38 kB
../public/vue-assets/assets/filters-Bgb6cQD6.css 0.75 kB │ gzip: 0.39 kB
../public/vue-assets/assets/AiContext-Bg-zjjbA.css 0.99 kB │ gzip: 0.45 kB
../public/vue-assets/assets/vue3-daterange-picker-izPOBj3P.css 1.02 kB │ gzip: 0.39 kB
../public/vue-assets/assets/PrimaryButton-CJj2FzKS.css 1.08 kB │ gzip: 0.34 kB
../public/vue-assets/assets/other-BTuPXtu_.css 1.15 kB │ gzip: 0.49 kB
../public/vue-assets/assets/jenesius-vue-modal-DDcfejHO.css 1.17 kB │ gzip: 0.48 kB
../public/vue-assets/assets/TabEmptyState-DOlIO68k.css 1.21 kB │ gzip: 0.53 kB
../public/vue-assets/assets/AppAlert-BlSJKiqj.css 1.29 kB │ gzip: 0.41 kB
../public/vue-assets/assets/useActivityCustomerName-CBjfIUfJ.css 1.42 kB │ gzip: 0.45 kB
../public/vue-assets/assets/UserAvatar-co8l9mMk.css 1.51 kB │ gzip: 0.49 kB
../public/vue-assets/assets/basic-modal-Dx7ZBbcA.css 1.55 kB │ gzip: 0.66 kB
../public/vue-assets/assets/login-DYbcBesM.css 1.58 kB │ gzip: 0.61 kB
../public/vue-assets/assets/DrawerWidget-poAmqspk.css 1.67 kB │ gzip: 0.69 kB
../public/vue-assets/assets/ActionItems-5FkZSo3G.css 1.79 kB │ gzip: 0.80 kB
../public/vue-assets/assets/join-conference-DQ6Yl7eU.css 1.82 kB │ gzip: 0.75 kB
../public/vue-assets/assets/InputText-C-DxDNSV.css 1.95 kB │ gzip: 0.71 kB
../public/vue-assets/assets/GoogleLikeButton-DvQHxbqR.css 2.11 kB │ gzip: 0.66 kB
../public/vue-assets/assets/usePusherEventListener-CDj8aSNp.css 2.20 kB │ gzip: 0.73 kB
../public/vue-assets/assets/onboard-IzGiQ-EC.css 2.32 kB │ gzip: 0.95 kB
../public/vue-assets/assets/ai-reports-manage-DaynTl62.css 2.34 kB │ gzip: 0.86 kB
../public/vue-assets/assets/StatusBadge-B4N12dzU.css 2.35 kB │ gzip: 0.86 kB
../public/vue-assets/assets/GridView-CsP1Jk-k.css 2.36 kB │ gzip: 0.90 kB
../public/vue-assets/assets/WelcomeLayout-6AX86p6F.css 2.38 kB │ gzip: 0.91 kB
../public/vue-assets/assets/snackbar-CS_iqvrv.css 2.45 kB │ gzip: 0.80 kB
../public/vue-assets/assets/ai-reports-mtqGqPPq.css 2.47 kB │ gzip: 0.91 kB
../public/vue-assets/assets/Comment-gmAfiuer.css 2.63 kB │ gzip: 1.00 kB
../public/vue-assets/assets/meeting-consent-BasRvxE3.css 2.66 kB │ gzip: 0.96 kB
../public/vue-assets/assets/DealRiskList-CwxmDnSC.css 2.67 kB │ gzip: 0.96 kB
../public/vue-assets/assets/add-to-playlist-modal-C_ad1hht.css 2.84 kB │ gzip: 0.95 kB
../public/vue-assets/assets/live-_zT5qBc1.css 3.60 kB │ gzip: 1.16 kB
../public/vue-assets/assets/RadioField-DDp8Y71G.css 3.64 kB │ gzip: 1.20 kB
../public/vue-assets/assets/vue-multiselect-BkQNV7oL.css 3.80 kB │ gzip: 0.86 kB
../public/vue-assets/assets/activity-preview-result-Db1Da81T.css 3.89 kB │ gzip: 1.15 kB
../public/vue-assets/assets/liquor-tree-CB5oh8v1.css 4.08 kB │ gzip: 1.24 kB
../public/vue-assets/assets/sentry-CSjHcvmn.css 4.47 kB │ gzip: 1.08 kB
../public/vue-assets/assets/AiAutomation-Cj1t8v5y.css 4.61 kB │ gzip: 0.84 kB
../public/vue-assets/assets/export-portal-CyIUVTEe.css 5.34 kB │ gzip: 1.65 kB
../public/vue-assets/assets/InputField-C4hoKJ7Z.css 6.13 kB │ gzip: 1.27 kB
../public/vue-assets/assets/vue-mq-bh4L87Tr.css 6.64 kB │ gzip: 1.95 kB
../public/vue-assets/assets/invitation-EPR1t-bH.css 6.74 kB │ gzip: 2.22 kB
../public/vue-assets/assets/ondemand-C2m0YVGA.css 6.84 kB │ gzip: 2.02 kB
../public/vue-assets/assets/AskAnything-CyvEhHS1.css 8.60 kB │ gzip: 2.60 kB
../public/vue-assets/assets/AppButton-BIeoar_U.css 8.88 kB │ gzip: 2.04 kB
../public/vue-assets/assets/kiosk-BcheyWWL.css 9.56 kB │ gzip: 2.53 kB
../public/vue-assets/assets/AiCrmNotes-nNTe_mOY.css 10.62 kB │ gzip: 1.93 kB
../public/vue-assets/assets/deal-view-DJzG7PLp.css 11.70 kB │ gzip: 3.29 kB
../public/vue-assets/assets/tokens-DZgWrfKd.css 12.44 kB │ gzip: 2.82 kB
../public/vue-assets/assets/tokens-DL2BPbai.css 12.89 kB │ gzip: 2.87 kB
../public/vue-assets/assets/playlists-Cy3t3kYU.css 13.54 kB │ gzip: 3.16 kB
../public/vue-assets/assets/PhoneField-D_kGOah0.css 13.75 kB │ gzip: 3.44 kB
../public/vue-assets/assets/ListView-Dt0qkyuY.css 17.07 kB │ gzip: 4.11 kB
../public/vue-assets/assets/intl-tel-input-DgmgTINs.css 17.11 kB │ gzip: 5.37 kB
../public/vue-assets/assets/AppFormField-DvpfRzi8.css 18.98 kB │ gzip: 4.42 kB
../public/vue-assets/assets/OrgSettingsLayout-BogdXtgY.css 21.98 kB │ gzip: 5.51 kB
../public/vue-assets/assets/dashboard-DQw3TKyk.css 22.74 kB │ gzip: 4.35 kB
../public/vue-assets/assets/deal-insights-CfX3UNzh.css 29.66 kB │ gzip: 6.44 kB
../public/vue-assets/assets/team-insights-CjVhm0JN.css 42.31 kB │ gzip: 8.46 kB
../public/vue-assets/assets/playback-CvJP5tX1.css 44.55 kB │ gzip: 10.36 kB
../public/vue-assets/assets/video-js-skin-DYZluGb-.css 47.73 kB │ gzip: 12.63 kB
../public/vue-assets/assets/assets-CAbfI4CY.css 83.38 kB │ gzip: 51.71 kB
../public/vue-assets/assets/logged-in-layout-1mAqkmnS.css 153.67 kB │ gzip: 27.18 kB
../public/vue-assets/assets/assets-xH6dx_9q.css 259.71 kB │ gzip: 181.71 kB
../public/vue-assets/assets/directives-DJJeJJOP.js 0.53 kB │ gzip: 0.36 kB │ map: 0.32 kB
../public/vue-assets/assets/jenesius-vue-modal-BuBhyl83.js 0.54 kB │ gzip: 0.37 kB │ map: 0.47 kB
../public/vue-assets/assets/spark-D_-Wgfar.js 0.54 kB │ gzip: 0.35 kB │ map: 0.40 kB
../public/vue-assets/assets/url-messenger-_CGQa-lH.js 0.56 kB │ gzip: 0.37 kB │ map: 0.46 kB
../public/vue-assets/assets/wavy-bg-Cec-_GDj.js 0.57 kB │ gzip: 0.36 kB │ map: 0.35 kB
../public/vue-assets/assets/component-css-class-CtO0AVgW.js 0.61 kB │ gzip: 0.40 kB │ map: 0.91 kB
../public/vue-assets/assets/theme-B5VdyMN_.js 0.63 kB │ gzip: 0.41 kB │ map: 0.49 kB
../public/vue-assets/assets/pick-C8Pw7kHx.js 0.71 kB │ gzip: 0.45 kB │ map: 1.59 kB
../public/vue-assets/assets/utils-BjsEoLDB.js 0.77 kB │ gzip: 0.50 kB │ map: 1.68 kB
../public/vue-assets/assets/throttle-siTOKEMt.js 0.78 kB │ gzip: 0.49 kB │ map: 3.24 kB
../public/vue-assets/assets/lastFilters-B_B_O5LY.js 0.80 kB │ gzip: 0.51 kB │ map: 1.12 kB
../public/vue-assets/assets/useAuthState-Bpx8WpBc.js 0.81 kB │ gz...
|
iTerm2
|
APP (-zsh)
|
NULL
|
46882
|
|
80568
|
incarcerated meaningIncarcerared means vemo morson incarcerated meaningIncarcerared means vemo morsoned or connnee n alal or ouison.As an adjective: "The incarcerated individuals awaited trial." As a verb (to incarcerate): "Thecourt ordered him to be incarcerated for five years."It comes from the Latin incarcerare — in- (into) + carcer (prison), which also gives us the word"carceral."It's often used in formal, legal, or policy contexts as a more neutral alternative to "imprisoned"or "jailed."Reply..Sonnet 4.6Claude is Al and can make mistakes. Please double-check responses....
|
Control Centre
|
Control Centre
|
NULL
|
80568
|
|
80572
|
incarcerated meaningIncarcerared means vemo morson incarcerated meaningIncarcerared means vemo morsoned or connnee n alal or ouison.As an adjective: "The incarcerated individuals awaited trial." As a verb (to incarcerate): "Thecourt ordered him to be incarcerated for five years."It comes from the Latin incarcerare — in- (into) + carcer (prison), which also gives us the word"carceral."It's often used in formal, legal, or policy contexts as a more neutral alternative to "imprisoned"or "jailed."Reply..Sonnet 4.6Claude is Al and can make mistakes. Please double-check responses....
|
iTerm2
|
sqlite3
|
NULL
|
80572
|
|
70213
|
in:#platform-inner-team datadog
in:
#platform-inne in:#platform-inner-team datadog
in:
#platform-inner-team
datadog
Clear
Show results for:
in:
platform-inner-team
datadog
Select
Give feedback
Close
HomeActivityLaterMoreSlackVIewMistonWindowHelpin: [PASSWORD] lattorm-inner-teanfatadog• Show results for: in:@platform-inner-team datadogeaDack+ + SelectGive feedback# Messages~11 resultsFromv= Filters (1)Sort: Most relevant (default)~Galya Dimitrova A platform-inner-teamSep 3rd, 2025 at 9:20 AMобоо утро, това e catadog чаота с всякакви Аl автоматизациивключително и репортите -[URL_WITH_CREDENTIALS] Batch iniial sync for SFeed — iminny — Sentry8 JiminnyPipelines - /app• Formalize[SRD-6793) Les Mills activity type7 Search results: calendar | Jiminny9 liminni8 Jiminny8 JiminnyE* Edit - Engineering - Confluence(UY-18909) (Part2) Automated rep0) SevenShores|Hubspot|Exceptions)@ CloudWatch | us-east-2ww Usage | WindsurtSevenShores|Hubspot\Exceptions)3 Workers | Datadogf. Pull requests - jiminny/app® Circle cl 8 PRODUS 8 Staging Li SentryInspectSplit GraphCompare TimeX Collapse QueryExperiment with your query. Keep in mind, changes are only saved in Edit modeMetrics• jiminny.sqs.work.count+ Add FormulaDisplay onlytop • 10 • sortby a jiminny.sqs.work.countfrom sjobqueue , senvironment+ Add QueryX sumby jobname x as countTop Jobs10 groups ranging from 34.43 events to 1.93k events.crm-sync -1926 jiminny.jobs.crm.delete.verifyactivitycrmtaskjob917 jiminny.jobs.crm.hubspot.processwebhookeventsjob429 jiminny.jobs.crm.syncprofilemetadata200 jiminny.jobs.crm.salesforce.fetchsalesforceentitiesjob130 jiminny.jobs.crm.synchubspotobjects118 jiminny.jobs.crm.delete.deleteaccountjob101 jiminny.jobs.crm.salesforce.processsalesforceentitybatchjob97 jiminny.jobs.crm.delete.deletecontactjob45 jiminny.jobs.crm.syncobjects34 jiminny.jobs.crm.hubspot.importcontactbatch& Share$0f Support Daily - in 2h 3m100% C4Wed 22 Apr 12:57:121c Past 1 Dav{ Modifyjobname: *L, Sharejiminny.jobs.crm.delete.deleteaccountjob1925.5 eventsVALUE117.6 events428.9 events129.7 event31 events20 795 ovonte69 2)k ovontd10770 events264.37k events 418.0 events$165 events14.11k events5044 events2234 events707 event8.15k events36.46k events 127.0 events585 event161 events10.74k event20.28k event3.0 eventd...
|
Slack
|
Search - Jiminny Inc - 1 new item - Slack
|
NULL
|
70213
|
|
33978
|
in:#jiminny-x-integration-app promise — Search
in: in:#jiminny-x-integration-app promise — Search
in:
#jiminny-x-integration-app
promise
— Search
Clear
Show results in this channel for:
promise
Show results for:
in:
jiminny-x-integration-app
promise
Recent messages in
jiminny-x-integration-app
Lukas Kovalik
jiminny-x-integration-app
Today
…
a
Promise
(
https://console.integration.app/ref/sdk/classes/IntegrationAccessor.html#openNewConnection
) upon
…
Select
Enter
Open
Actions
Give feedback
Close
FirefoxFileEditViewHistoryBookmarksProfilesToolsWindowHelpSupport Daily - in 3h 44 m100% C8Thu 16 Apr 11:16:27-zshDOCKER₫812026-04-16T10:53:52.475287Z2026-04-16T10:56:14.166792Z2026-04-16T10:56:17.065758Z2026-04-16T10:56:20.134540Z2026-04-16T10:56:23.139917Z2026-04-16T10:56:26.244357Z2026-04-16T10:56:27.562449Z2026-04-16T10:56:32.226001Z2026-04-16T10:56:35.318534Z2026-64-16110:56:52.32415772026-04-16T10:56:55.329246Z2026-04-16T10:56:58.356364Z2026-04-16110:57:01143462972026-04-16T10:57:03.322854Z2026-04-16T10:57:10.706619Z2026-04-16T10:57:36.755032Z2026-04-16T10:58:56.119319Z\nFROM\nframes\nWHERE\nd=3.628561542s2026-04-16T10:58:56.121240Z2026-04-16710:59:00.498686Z2026-04-16T10:59:09.377495Z2026-04-16T10:59:36.545512Z2026-04-16T11:04:15.237763Z\nFROM\nframes\nWHERE\n=5.844181583s2026-04-16T11:04:15.238562Z2026-04-16T11:04:24.012117Z2026-04-16T11:04:40.465386Z2026-04-16T11:09:42.625140Z\nFROM\nframes \nWHERE\nd=2.14471325s2026-04-16T11:09:42.630467Z2026-04-16111:09:51.29824622026-04-16T11:10:06.899071Z2026-04-16T11:15:13.835798Z\nFROM\nframes\nWHERE\nd-6.920954875s2026-04-16T11:15:13.840780Z2026-04-16T11:15:21.357742Z2026-04-16711:15:39.238002Z2026-04-16T11:15:39.455620ZnFROM\nframes \nWHERE\n56582333sDEV (docker)282APP (-zsh)83ec2-user@ip-10-30-₴4-zsh86-zsh₴7* Unable to acce...O 88INFOscreenpipe_engine::snapshot_compaction:snapshotcompaction:4613.1MB→ 1.5MB46JPEGSdeletedINFOINFOscreenpipe_engine::event_driven_capture:frames,contentdedup:skipping(8.6x),capture for monitor 2Chash=3616940803251985209,trigger=visual_change)screenpipe_engine::event_driven_capture:contentdedup:skipping capture for monitor 2INFOscreenpipe_engine::event_driven_capture:contentdedup:skipping capture for monitor 2Chash=3616940803251985209,trigger=visual_change)Chash=3616940803251985209,trigger=visual_change)INFOINFOscreenpipe_engine::event_driven_capture:contentdedup:skippingcapture for monitor 2Chash=3616940803251985209,trigger=visual_change)screenpipe_engine::event_driven_capture:contentdedup: skippingcapture for monitor 2 (hash=3616940803251985209,trigger=visual_change)INFOscreenpipe_engine::event_driven_capture:contentdedup: skippingcapture for monitor 1 (hash=3616940803251985209,trigger=click)INFOscreenpipe_engine::event_driven_capture:contentdedup:skipping capture for monitor 2INFOscreenpipe_engine::event_driven_capture:contentdedup:skipping capture for monitor 2Chash=3616940803251985209,INFOscreenpipe_engine::event_driven_capture: contentdedup:skippingcapture for monitor 2Chash=3616940803251985209,trigger=visual_change)trigger=visual_change)(hash=-1884356785177423556, trigger=visual_change)INFOscreenpipe_engine::event_driven_capture:contentdedup:skippingcapture for monitor 2INFOChash=-1884356785177423556,trigger=visual_change)INFOscreenpipe_engine::event_driven_capture: contentdedup:skippingcapture for monitor 2 (hash=-1884356785177423556,trigger=visual_change)screenpipe_engine::event_driven_capture:contentdedup:skippingcapture for monitor 2 (hash=-1884356785177423556,trigger=visual_change)INFOscreenpipe_engine::event,_driven_capture:contentdedup:skippingcapture for monitor 2 (hash=-1884356785177423556, trigger=click)INFOINFOscreenpipe_engine::event_driven_capture:contentdedup:skippingcapture for monitor 2 (hash=-1884356785177423556,trigger=visual_change)screenpipe_engine::event_driven_capture:content dedup:skippingcapturefor monitor 2 (hash=3616940803251985209,trigger=click)WARNsalx::query:summary="SELECT id,snapshot_path, device_name, "db.statement="\n\nSELECT\n id,\nsnapshot_path, \ndevice_name, \ntimestampsnapshot_path IS NOT NULL\nAND timestamp < ?1\nORDER BY\ndevice_name, \ntimestamp ASC\nLIMIT\n5000\n"rows_affected=0 rows_returned=117elapseINFOscreenpipe_engine::snapshot_compaction: snapshot compaction: found 117 eligible framesINFOscreenpipe_engine::snapshot_compaction: snapshotINFOcompaction: 44 frames, 11.OMB → 3.4MB (3.2x), 44 JPEGs deletedscreenpipe_engine::snapshot_compaction: snapshot compaction: 71 frames,11. 6MB → 3.7MB (3.2x), 71 JPEGs deletedINFOscreenpipe_engine::event_driven_capture: content dedup:skipping capture for monitor 2 (hash=-5207847904424027181,trigger=visual_change)WARNsqlx::query:summary="SELECT id, snapshot_path, device_name,db.statement="\n\nSELECT\n id, \nsnapshot_path, In device_name, \ntimestampsnapshot_path IS NOT NULL\nAND timestamp < ?1\nORDER BY\ndevice_name, \ntimestamp ASC\nLIMIT\n 5000\n" rows_affected-0 rows_returned=99 elapsedINFOscreenpipe_engine::snapshot_compaction: snapshotcompaction: found 99eligible framesINFOscreenpipe_engine::snapshot_compaction: snapshotcompaction: 40 frames, 12.0MB → 3.4MB (3.6x), 40 JPEGs deletedINFOscreenpipe_engine::snapshot_compaction: snapshot compaction: 57 frames,10.2MB→ 3.4MB (3.0x), 57 JPEGSdeletedWARNsqlx::query:summary="SELECT id, snapshot_path, device_name, " db.statement="\n\nSELECT\n id,\nsnapshot_path, \ndevice_name, \ntimestampsnapshot_path IS NOT NULL\nAND timestamp < ?1\nORDER BY\ndevice_name, \ntimestamp ASC\nLIMIT\n5000\n" rows_affected=0 rows_returned-132 elapseINFOscreenpipe_engine::snapshot_compaction: snapshot compaction: found 132 eligible framesINFOscreenpipe_engine::snapshot_compaction: snapshot compaction: 51 frames, 18.6MB → 6.6MB (2.8x), 51 JPEGs deletedINFOscreenpipe_engine::snapshot_compaction: snapshot compaction: 79 frames,11.2MB → 3.8MB (2.9x),79 JPEGs deletedWARNsqlx::query:summary="SELECT id, snapshot_path, device_name, _" db.statement="\n\nSELECT\n id,\nsnapshot_path, In device_name, \ntimestampsnapshot_path IS NOT NULL\nAND timestamp < ?1\nORDER BY\n device_name, \ntimestamp ASC\nLIMIT\n5000\n'rows_affected=0 rows_returned=100 elapseINFOscreenpipe_engine::snapshot_compaction: snapshotcompaction: found 100 eligible framesINFOscreenpipe_engine::snapshot_compaction: snapshot compaction: 33 frames, 12.5MB + 1.9MB (6.7x), 33 JPEGs deletedINFOscreenpipe_engine::snapshot_compaction: snapshotWARNcompaction: 65 frames,13.8MB → 5.2MB (2.6x), 65 JPEGs deletedsqlx:: query:summary="SELECT DISTINCT app_name, window_name,db.statement="\n\nSELECT\n DISTINCT app_name, \nwindow_name, \nbrowser_url\timestamp › datetime('now''-30 seconds')\n AND app_name IS NOT NULL\n AND window_name IS NOT NULL\n" rows_affected-0 rows_returned=147 elapsed=2.0...
|
Slack
|
jiminny-x-integration-app (Channel) - Jiminny Inc jiminny-x-integration-app (Channel) - Jiminny Inc - Slack...
|
NULL
|
33978
|
|
40458
|
idx_elements_source_role_text|elements|CREATE INDE idx_elements_source_role_text|elements|CREATE INDEX idx_elements_source_role_text
ON elements(source, role, frame_id)
WHERE text IS NOT NULL
idx_elements_frame_source_role|elements|CREATE INDEX idx_elements_frame_source_role ON elements(frame_id, source, role) WHERE text IS NOT NULL
idx_frames_timestamp|frames|CREATE INDEX idx_frames_timestamp ON frames(timestamp)
idx_frames_video_chunk_id|frames|CREATE INDEX idx_frames_video_chunk_id ON frames(video_chunk_id)
idx_frames_timestamp_device|frames|CREATE INDEX idx_frames_timestamp_device
ON frames(timestamp, device_name)
idx_frames_snapshot_path|frames|CREATE INDEX idx_frames_snapshot_path
ON frames(snapshot_path) WHERE snapshot_path IS NOT NULL
idx_frames_cloud_blob_id|frames|CREATE INDEX idx_frames_cloud_blob_id
ON frames(cloud_blob_id) WHERE cloud_blob_id IS NULL AND snapshot_path IS NOT NULL
idx_frames_sync_id|frames|CREATE INDEX idx_frames_sync_id ON frames(sync_id) WHERE sync_id IS NOT NULL
idx_frames_elements_ref_frame_id|frames|CREATE INDEX idx_frames_elements_ref_frame_id
ON frames(elements_ref_frame_id)
WHERE elements_ref_frame_id IS NOT NULL
idx_meetings_start|meetings|CREATE INDEX idx_meetings_start ON meetings(meeting_start)
idx_meetings_end|meetings|CREATE INDEX idx_meetings_end ON meetings(meeting_end)
idx_memories_created_at|memories|CREATE INDEX idx_memories_created_at ON memories(created_at)
idx_memories_importance|memories|CREATE INDEX idx_memories_importance ON memories(importance DESC)
idx_memories_source|memories|CREATE INDEX idx_memories_source ON memories(source)
idx_memories_frame_id|memories|CREATE INDEX idx_memories_frame_id ON memories(frame_id)
idx_ocr_text_frame_id|ocr_text|CREATE INDEX idx_ocr_text_frame_id ON ocr_text(frame_id)
idx_ocr_text_frame_app_window|ocr_text|CREATE INDEX idx_ocr_text_frame_app_window ON ocr_text(frame_id, app_name, window_name)
idx_ocr_text_length|ocr_text|CREATE INDEX idx_ocr_text_length ON ocr_text (text_length)
idx_ocr_text_sync_id|ocr_text|CREATE INDEX idx_ocr_text_sync_id ON ocr_text(sync_id) WHERE sync_id IS NOT NULL
idx_pipe_exec_name_status|pipe_executions|CREATE INDEX idx_pipe_exec_name_status ON pipe_executions(pipe_name, status)
idx_pipe_exec_running|pipe_executions|CREATE INDEX idx_pipe_exec_running ON pipe_executions(status) WHERE status = 'running'
idx_pipe_exec_name_time|pipe_executions|CREATE INDEX idx_pipe_exec_name_time ON pipe_executions(pipe_name, id DESC)
idx_ui_events_timestamp|ui_events|CREATE INDEX idx_ui_events_timestamp ON ui_events(timestamp)
idx_ui_events_event_type|ui_events|CREATE INDEX idx_ui_events_event_type ON ui_events(event_type)
idx_ui_events_app_name|ui_events|CREATE INDEX idx_ui_events_app_name ON ui_events(app_name)
idx_ui_events_session_id|ui_events|CREATE INDEX idx_ui_events_session_id ON ui_events(session_id)
idx_ui_events_frame_id|ui_events|CREATE INDEX idx_ui_events_frame_id ON ui_events(frame_id)
idx_ui_events_sync_id|ui_events|CREATE INDEX idx_ui_events_sync_id ON ui_events(sync_id)
idx_ui_events_synced_at|ui_events|CREATE INDEX idx_ui_events_synced_at ON ui_events(synced_at)
idx_ui_events_unsynced|ui_events|CREATE INDEX idx_ui_events_unsynced ON ui_events(synced_at) WHERE synced_at IS NULL
idx_video_chunks_device_name|video_chunks|CREATE INDEX idx_video_chunks_device_name ON video_chunks(device_name)
idx_video_chunks_device_name_id|video_chunks|CREATE INDEX idx_video_chunks_device_name_id ON video_chunks(device_name, id DESC)
idx_video_chunks_cloud_blob_id|video_chunks|CREATE INDEX idx_video_chunks_cloud_blob_id
ON video_chunks(cloud_blob_id) WHERE cloud_blob_id IS NULL
idx_vision_tags_vision_id|vision_tags|CREATE INDEX idx_vision_tags_vision_id ON vision_tags(vision_id)
idx_vision_tags_tag_id|vision_tags|CREATE INDEX idx_vision_tags_tag_id ON vision_tags(tag_id)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ clear
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT
'elements_fts' as name, COUNT(*) as rows FROM elements_fts
UNION ALL SELECT 'frames_fts', COUNT(*) FROM frames_fts
UNION ALL SELECT 'ui_events_fts', COUNT(*) FROM ui_events_fts
UNION ALL SELECT 'audio_transcriptions_fts', COUNT(*) FROM audio_transcriptions_fts;"
elements_fts|2595462
frames_fts|40394
ui_events_fts|51394
audio_transcriptions_fts|3
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT name, SUM(payload) as bytes
FROM dbstat
WHERE name IN ('elements_fts_data','frames_fts_data','ui_events_fts_data')
GROUP BY name;"
elements_fts_data|49417125
frames_fts_data|46211048
ui_events_fts_data|1661147
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT name, SUM(payload) as bytes
FROM dbstat
WHERE name IN ('frames', 'ui_events', 'elements_fts_data','frames_fts_data','ui_events_fts_data')
GROUP BY name;"
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
screenpipe"
Close Tab
-zsh
Close Tab
✳ Review screenpipe usage and Boosteroid integration (node)
Close Tab
⌥⌘1
-zsh...
|
iTerm2
|
-zsh
|
NULL
|
40458
|
|
40501
|
idx_elements_source_role_text|elements|CREATE INDE idx_elements_source_role_text|elements|CREATE INDEX idx_elements_source_role_text
ON elements(source, role, frame_id)
WHERE text IS NOT NULL
idx_elements_frame_source_role|elements|CREATE INDEX idx_elements_frame_source_role ON elements(frame_id, source, role) WHERE text IS NOT NULL
idx_frames_timestamp|frames|CREATE INDEX idx_frames_timestamp ON frames(timestamp)
idx_frames_video_chunk_id|frames|CREATE INDEX idx_frames_video_chunk_id ON frames(video_chunk_id)
idx_frames_timestamp_device|frames|CREATE INDEX idx_frames_timestamp_device
ON frames(timestamp, device_name)
idx_frames_snapshot_path|frames|CREATE INDEX idx_frames_snapshot_path
ON frames(snapshot_path) WHERE snapshot_path IS NOT NULL
idx_frames_cloud_blob_id|frames|CREATE INDEX idx_frames_cloud_blob_id
ON frames(cloud_blob_id) WHERE cloud_blob_id IS NULL AND snapshot_path IS NOT NULL
idx_frames_sync_id|frames|CREATE INDEX idx_frames_sync_id ON frames(sync_id) WHERE sync_id IS NOT NULL
idx_frames_elements_ref_frame_id|frames|CREATE INDEX idx_frames_elements_ref_frame_id
ON frames(elements_ref_frame_id)
WHERE elements_ref_frame_id IS NOT NULL
idx_meetings_start|meetings|CREATE INDEX idx_meetings_start ON meetings(meeting_start)
idx_meetings_end|meetings|CREATE INDEX idx_meetings_end ON meetings(meeting_end)
idx_memories_created_at|memories|CREATE INDEX idx_memories_created_at ON memories(created_at)
idx_memories_importance|memories|CREATE INDEX idx_memories_importance ON memories(importance DESC)
idx_memories_source|memories|CREATE INDEX idx_memories_source ON memories(source)
idx_memories_frame_id|memories|CREATE INDEX idx_memories_frame_id ON memories(frame_id)
idx_ocr_text_frame_id|ocr_text|CREATE INDEX idx_ocr_text_frame_id ON ocr_text(frame_id)
idx_ocr_text_frame_app_window|ocr_text|CREATE INDEX idx_ocr_text_frame_app_window ON ocr_text(frame_id, app_name, window_name)
idx_ocr_text_length|ocr_text|CREATE INDEX idx_ocr_text_length ON ocr_text (text_length)
idx_ocr_text_sync_id|ocr_text|CREATE INDEX idx_ocr_text_sync_id ON ocr_text(sync_id) WHERE sync_id IS NOT NULL
idx_pipe_exec_name_status|pipe_executions|CREATE INDEX idx_pipe_exec_name_status ON pipe_executions(pipe_name, status)
idx_pipe_exec_running|pipe_executions|CREATE INDEX idx_pipe_exec_running ON pipe_executions(status) WHERE status = 'running'
idx_pipe_exec_name_time|pipe_executions|CREATE INDEX idx_pipe_exec_name_time ON pipe_executions(pipe_name, id DESC)
idx_ui_events_timestamp|ui_events|CREATE INDEX idx_ui_events_timestamp ON ui_events(timestamp)
idx_ui_events_event_type|ui_events|CREATE INDEX idx_ui_events_event_type ON ui_events(event_type)
idx_ui_events_app_name|ui_events|CREATE INDEX idx_ui_events_app_name ON ui_events(app_name)
idx_ui_events_session_id|ui_events|CREATE INDEX idx_ui_events_session_id ON ui_events(session_id)
idx_ui_events_frame_id|ui_events|CREATE INDEX idx_ui_events_frame_id ON ui_events(frame_id)
idx_ui_events_sync_id|ui_events|CREATE INDEX idx_ui_events_sync_id ON ui_events(sync_id)
idx_ui_events_synced_at|ui_events|CREATE INDEX idx_ui_events_synced_at ON ui_events(synced_at)
idx_ui_events_unsynced|ui_events|CREATE INDEX idx_ui_events_unsynced ON ui_events(synced_at) WHERE synced_at IS NULL
idx_video_chunks_device_name|video_chunks|CREATE INDEX idx_video_chunks_device_name ON video_chunks(device_name)
idx_video_chunks_device_name_id|video_chunks|CREATE INDEX idx_video_chunks_device_name_id ON video_chunks(device_name, id DESC)
idx_video_chunks_cloud_blob_id|video_chunks|CREATE INDEX idx_video_chunks_cloud_blob_id
ON video_chunks(cloud_blob_id) WHERE cloud_blob_id IS NULL
idx_vision_tags_vision_id|vision_tags|CREATE INDEX idx_vision_tags_vision_id ON vision_tags(vision_id)
idx_vision_tags_tag_id|vision_tags|CREATE INDEX idx_vision_tags_tag_id ON vision_tags(tag_id)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ clear
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT
'elements_fts' as name, COUNT(*) as rows FROM elements_fts
UNION ALL SELECT 'frames_fts', COUNT(*) FROM frames_fts
UNION ALL SELECT 'ui_events_fts', COUNT(*) FROM ui_events_fts
UNION ALL SELECT 'audio_transcriptions_fts', COUNT(*) FROM audio_transcriptions_fts;"
elements_fts|2595462
frames_fts|40394
ui_events_fts|51394
audio_transcriptions_fts|3
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT name, SUM(payload) as bytes
FROM dbstat
WHERE name IN ('elements_fts_data','frames_fts_data','ui_events_fts_data')
GROUP BY name;"
elements_fts_data|49417125
frames_fts_data|46211048
ui_events_fts_data|1661147
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT name, SUM(payload) as bytes
FROM dbstat
WHERE name IN ('frames', 'ui_events', 'elements', 'elements_fts_data','frames_fts_data','ui_events_fts_data')
GROUP BY name;"
elements|334884969
elements_fts_data|49472096
frames|689381789
frames_fts_data|46270876
ui_events|8354906
ui_events_fts_data|1663427
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "SELECT sql FROM sqlite_master WHERE type='table' AND name IN ('elements_fts','frames_fts','ui_events_fts','memories_fts') ORDER BY name;"
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
screenpipe"
Close Tab
-zsh
Close Tab
✳ Review screenpipe usage and Boosteroid integration (node)
Close Tab
⌥⌘1
-zsh...
|
iTerm2
|
-zsh
|
NULL
|
40501
|
|
40502
|
idx_elements_source_role_text|elements|CREATE INDE idx_elements_source_role_text|elements|CREATE INDEX idx_elements_source_role_text
ON elements(source, role, frame_id)
WHERE text IS NOT NULL
idx_elements_frame_source_role|elements|CREATE INDEX idx_elements_frame_source_role ON elements(frame_id, source, role) WHERE text IS NOT NULL
idx_frames_timestamp|frames|CREATE INDEX idx_frames_timestamp ON frames(timestamp)
idx_frames_video_chunk_id|frames|CREATE INDEX idx_frames_video_chunk_id ON frames(video_chunk_id)
idx_frames_timestamp_device|frames|CREATE INDEX idx_frames_timestamp_device
ON frames(timestamp, device_name)
idx_frames_snapshot_path|frames|CREATE INDEX idx_frames_snapshot_path
ON frames(snapshot_path) WHERE snapshot_path IS NOT NULL
idx_frames_cloud_blob_id|frames|CREATE INDEX idx_frames_cloud_blob_id
ON frames(cloud_blob_id) WHERE cloud_blob_id IS NULL AND snapshot_path IS NOT NULL
idx_frames_sync_id|frames|CREATE INDEX idx_frames_sync_id ON frames(sync_id) WHERE sync_id IS NOT NULL
idx_frames_elements_ref_frame_id|frames|CREATE INDEX idx_frames_elements_ref_frame_id
ON frames(elements_ref_frame_id)
WHERE elements_ref_frame_id IS NOT NULL
idx_meetings_start|meetings|CREATE INDEX idx_meetings_start ON meetings(meeting_start)
idx_meetings_end|meetings|CREATE INDEX idx_meetings_end ON meetings(meeting_end)
idx_memories_created_at|memories|CREATE INDEX idx_memories_created_at ON memories(created_at)
idx_memories_importance|memories|CREATE INDEX idx_memories_importance ON memories(importance DESC)
idx_memories_source|memories|CREATE INDEX idx_memories_source ON memories(source)
idx_memories_frame_id|memories|CREATE INDEX idx_memories_frame_id ON memories(frame_id)
idx_ocr_text_frame_id|ocr_text|CREATE INDEX idx_ocr_text_frame_id ON ocr_text(frame_id)
idx_ocr_text_frame_app_window|ocr_text|CREATE INDEX idx_ocr_text_frame_app_window ON ocr_text(frame_id, app_name, window_name)
idx_ocr_text_length|ocr_text|CREATE INDEX idx_ocr_text_length ON ocr_text (text_length)
idx_ocr_text_sync_id|ocr_text|CREATE INDEX idx_ocr_text_sync_id ON ocr_text(sync_id) WHERE sync_id IS NOT NULL
idx_pipe_exec_name_status|pipe_executions|CREATE INDEX idx_pipe_exec_name_status ON pipe_executions(pipe_name, status)
idx_pipe_exec_running|pipe_executions|CREATE INDEX idx_pipe_exec_running ON pipe_executions(status) WHERE status = 'running'
idx_pipe_exec_name_time|pipe_executions|CREATE INDEX idx_pipe_exec_name_time ON pipe_executions(pipe_name, id DESC)
idx_ui_events_timestamp|ui_events|CREATE INDEX idx_ui_events_timestamp ON ui_events(timestamp)
idx_ui_events_event_type|ui_events|CREATE INDEX idx_ui_events_event_type ON ui_events(event_type)
idx_ui_events_app_name|ui_events|CREATE INDEX idx_ui_events_app_name ON ui_events(app_name)
idx_ui_events_session_id|ui_events|CREATE INDEX idx_ui_events_session_id ON ui_events(session_id)
idx_ui_events_frame_id|ui_events|CREATE INDEX idx_ui_events_frame_id ON ui_events(frame_id)
idx_ui_events_sync_id|ui_events|CREATE INDEX idx_ui_events_sync_id ON ui_events(sync_id)
idx_ui_events_synced_at|ui_events|CREATE INDEX idx_ui_events_synced_at ON ui_events(synced_at)
idx_ui_events_unsynced|ui_events|CREATE INDEX idx_ui_events_unsynced ON ui_events(synced_at) WHERE synced_at IS NULL
idx_video_chunks_device_name|video_chunks|CREATE INDEX idx_video_chunks_device_name ON video_chunks(device_name)
idx_video_chunks_device_name_id|video_chunks|CREATE INDEX idx_video_chunks_device_name_id ON video_chunks(device_name, id DESC)
idx_video_chunks_cloud_blob_id|video_chunks|CREATE INDEX idx_video_chunks_cloud_blob_id
ON video_chunks(cloud_blob_id) WHERE cloud_blob_id IS NULL
idx_vision_tags_vision_id|vision_tags|CREATE INDEX idx_vision_tags_vision_id ON vision_tags(vision_id)
idx_vision_tags_tag_id|vision_tags|CREATE INDEX idx_vision_tags_tag_id ON vision_tags(tag_id)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ clear
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT
'elements_fts' as name, COUNT(*) as rows FROM elements_fts
UNION ALL SELECT 'frames_fts', COUNT(*) FROM frames_fts
UNION ALL SELECT 'ui_events_fts', COUNT(*) FROM ui_events_fts
UNION ALL SELECT 'audio_transcriptions_fts', COUNT(*) FROM audio_transcriptions_fts;"
elements_fts|2595462
frames_fts|40394
ui_events_fts|51394
audio_transcriptions_fts|3
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT name, SUM(payload) as bytes
FROM dbstat
WHERE name IN ('elements_fts_data','frames_fts_data','ui_events_fts_data')
GROUP BY name;"
elements_fts_data|49417125
frames_fts_data|46211048
ui_events_fts_data|1661147
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT name, SUM(payload) as bytes
FROM dbstat
WHERE name IN ('frames', 'ui_events', 'elements', 'elements_fts_data','frames_fts_data','ui_events_fts_data')
GROUP BY name;"
elements|334884969
elements_fts_data|49472096
frames|689381789
frames_fts_data|46270876
ui_events|8354906
ui_events_fts_data|1663427
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "SELECT sql FROM sqlite_master WHERE type='table' AND name IN ('elements_fts','frames_fts','ui_events_fts','memories_fts') ORDER BY name;"
DOCKER
Close Tab
DEV (-zsh)
Close Tab
APP (-zsh)
Close Tab
screenpipe"
Close Tab
-zsh
Close Tab
✳ Review screenpipe usage and Boosteroid integration (node)
Close Tab
⌥⌘1
-zsh...
|
iTerm2
|
-zsh
|
NULL
|
40502
|
|
40525
|
idx_elements_source_role_text|elements|CREATE INDE idx_elements_source_role_text|elements|CREATE INDEX idx_elements_source_role_text
ON elements(source, role, frame_id)
WHERE text IS NOT NULL
idx_elements_frame_source_role|elements|CREATE INDEX idx_elements_frame_source_role ON elements(frame_id, source, role) WHERE text IS NOT NULL
idx_frames_timestamp|frames|CREATE INDEX idx_frames_timestamp ON frames(timestamp)
idx_frames_video_chunk_id|frames|CREATE INDEX idx_frames_video_chunk_id ON frames(video_chunk_id)
idx_frames_timestamp_device|frames|CREATE INDEX idx_frames_timestamp_device
ON frames(timestamp, device_name)
idx_frames_snapshot_path|frames|CREATE INDEX idx_frames_snapshot_path
ON frames(snapshot_path) WHERE snapshot_path IS NOT NULL
idx_frames_cloud_blob_id|frames|CREATE INDEX idx_frames_cloud_blob_id
ON frames(cloud_blob_id) WHERE cloud_blob_id IS NULL AND snapshot_path IS NOT NULL
idx_frames_sync_id|frames|CREATE INDEX idx_frames_sync_id ON frames(sync_id) WHERE sync_id IS NOT NULL
idx_frames_elements_ref_frame_id|frames|CREATE INDEX idx_frames_elements_ref_frame_id
ON frames(elements_ref_frame_id)
WHERE elements_ref_frame_id IS NOT NULL
idx_meetings_start|meetings|CREATE INDEX idx_meetings_start ON meetings(meeting_start)
idx_meetings_end|meetings|CREATE INDEX idx_meetings_end ON meetings(meeting_end)
idx_memories_created_at|memories|CREATE INDEX idx_memories_created_at ON memories(created_at)
idx_memories_importance|memories|CREATE INDEX idx_memories_importance ON memories(importance DESC)
idx_memories_source|memories|CREATE INDEX idx_memories_source ON memories(source)
idx_memories_frame_id|memories|CREATE INDEX idx_memories_frame_id ON memories(frame_id)
idx_ocr_text_frame_id|ocr_text|CREATE INDEX idx_ocr_text_frame_id ON ocr_text(frame_id)
idx_ocr_text_frame_app_window|ocr_text|CREATE INDEX idx_ocr_text_frame_app_window ON ocr_text(frame_id, app_name, window_name)
idx_ocr_text_length|ocr_text|CREATE INDEX idx_ocr_text_length ON ocr_text (text_length)
idx_ocr_text_sync_id|ocr_text|CREATE INDEX idx_ocr_text_sync_id ON ocr_text(sync_id) WHERE sync_id IS NOT NULL
idx_pipe_exec_name_status|pipe_executions|CREATE INDEX idx_pipe_exec_name_status ON pipe_executions(pipe_name, status)
idx_pipe_exec_running|pipe_executions|CREATE INDEX idx_pipe_exec_running ON pipe_executions(status) WHERE status = 'running'
idx_pipe_exec_name_time|pipe_executions|CREATE INDEX idx_pipe_exec_name_time ON pipe_executions(pipe_name, id DESC)
idx_ui_events_timestamp|ui_events|CREATE INDEX idx_ui_events_timestamp ON ui_events(timestamp)
idx_ui_events_event_type|ui_events|CREATE INDEX idx_ui_events_event_type ON ui_events(event_type)
idx_ui_events_app_name|ui_events|CREATE INDEX idx_ui_events_app_name ON ui_events(app_name)
idx_ui_events_session_id|ui_events|CREATE INDEX idx_ui_events_session_id ON ui_events(session_id)
idx_ui_events_frame_id|ui_events|CREATE INDEX idx_ui_events_frame_id ON ui_events(frame_id)
idx_ui_events_sync_id|ui_events|CREATE INDEX idx_ui_events_sync_id ON ui_events(sync_id)
idx_ui_events_synced_at|ui_events|CREATE INDEX idx_ui_events_synced_at ON ui_events(synced_at)
idx_ui_events_unsynced|ui_events|CREATE INDEX idx_ui_events_unsynced ON ui_events(synced_at) WHERE synced_at IS NULL
idx_video_chunks_device_name|video_chunks|CREATE INDEX idx_video_chunks_device_name ON video_chunks(device_name)
idx_video_chunks_device_name_id|video_chunks|CREATE INDEX idx_video_chunks_device_name_id ON video_chunks(device_name, id DESC)
idx_video_chunks_cloud_blob_id|video_chunks|CREATE INDEX idx_video_chunks_cloud_blob_id
ON video_chunks(cloud_blob_id) WHERE cloud_blob_id IS NULL
idx_vision_tags_vision_id|vision_tags|CREATE INDEX idx_vision_tags_vision_id ON vision_tags(vision_id)
idx_vision_tags_tag_id|vision_tags|CREATE INDEX idx_vision_tags_tag_id ON vision_tags(tag_id)
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ clear
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT
'elements_fts' as name, COUNT(*) as rows FROM elements_fts
UNION ALL SELECT 'frames_fts', COUNT(*) FROM frames_fts
UNION ALL SELECT 'ui_events_fts', COUNT(*) FROM ui_events_fts
UNION ALL SELECT 'audio_transcriptions_fts', COUNT(*) FROM audio_transcriptions_fts;"
elements_fts|2595462
frames_fts|40394
ui_events_fts|51394
audio_transcriptions_fts|3
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT name, SUM(payload) as bytes
FROM dbstat
WHERE name IN ('elements_fts_data','frames_fts_data','ui_events_fts_data')
GROUP BY name;"
elements_fts_data|49417125
frames_fts_data|46211048
ui_events_fts_data|1661147
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "
SELECT name, SUM(payload) as bytes
FROM dbstat
WHERE name IN ('frames', 'ui_events', 'elements', 'elements_fts_data','frames_fts_data','ui_events_fts_data')
GROUP BY name;"
elements|334884969
elements_fts_data|49472096
frames|689381789
frames_fts_data|46270876
ui_events|8354906
ui_events_fts_data|1663427
lukas@Lukas-Kovaliks-MacBook-Pro-Jiminny ~/.screenpipe $ sqlite3 ~/.screenpipe/db.sqlite "SELECT sql FROM sqlite_master WHERE type='table' AND name IN ('elements_fts','frames_fts','ui_events_fts','memories_fts') ORDER BY name;"
CREATE VIRTUAL TABLE elements_fts USING fts5(
text,
role,
frame_id UNINDEXED,
content='elements',
content_rowid='id',
tokenize='unicode61'
)
CREATE VIRTUAL TABLE frames_fts USING fts5(
full_text,
app_name,
window_name,
browser_url,
UW PICO 5.09 New Buffer
[ Read 1 line ]
^G Get Help ^O WriteOut ^R Read File ^Y Prev Pg ^K Cut Text ^C Cur Pos
^X Exit ^J Justify ^W Where is ^V Next Pg ^U UnCut Text ^T To Spell
UW PICO 5.09 File: screenpipe_sync.sh Modified
#!/bin/bash
# screenpipe_sync.sh
# Syncs Screenpipe SQLite data to a NAS archive database (append-only, no deletions).
#
# Usage:
# ./screenpipe_sync.sh # syncs yesterday (default)
# ./screenpipe_sync.sh 2026-04-15 # syncs a specific date
# ./screenpipe_sync.sh today # syncs today so far
#
# Cron example (runs at 3am daily):
# 0 3 * * * /Users/lukas/scripts/screenpipe_sync.sh >> |