WARNING: THIS SITE IS A MIRROR OF GITHUB.COM / IT CANNOT LOGIN OR REGISTER ACCOUNTS / THE CONTENTS ARE PROVIDED AS-IS / THIS SITE ASSUMES NO RESPONSIBILITY FOR ANY DISPLAYED CONTENT OR LINKS / IF YOU FOUND SOMETHING MAY NOT GOOD FOR EVERYONE, CONTACT ADMIN AT ilovescratch@foxmail.com
Skip to content

Main Container not starting services in kubernetes / openshift #476

@MWurzinger

Description

@MWurzinger

Support guidelines

I've found a bug and checked that ...

  • ... the documentation does not mention anything about my problem
  • ... there are no open or closed issues that are related to my problem

Description

We have deployed the helm-chart to a openshift cluster.
The service account has the right to use privileged rights.

Expected behaviour

The container should start the needed services and serve the gui.

Actual behaviour

The log says:
[services.d] starting services
[services.d] done.

But the container fails the readiness probe.

When i start the terminal and execute "/init" by hand, the fpm process starts and the gui is accessable.

Steps to reproduce

Create service aaccount with roleref "system:openshift:scc:privileged".
Deploy helm chart and set service account.

Stateful set starts pod, but container fails readiness probe.

Docker info

Not possible because of openshift

Docker Compose config

No response

Logs

[s6-init] making user provided files available at /var/run/s6/etc...exited 0.
[s6-init] ensuring user provided files have correct perms...exited 0.
[fix-attrs.d] applying ownership & permissions fixes...
[fix-attrs.d] done.
[cont-init.d] executing container initialization scripts...
[cont-init.d] 00-fix-logs.sh: executing... 
chown: changing ownership of '/proc/self/fd/1': Permission denied
chown: changing ownership of '/proc/self/fd/2': Permission denied
[cont-init.d] 00-fix-logs.sh: exited 0.
[cont-init.d] 01-fix-uidgid.sh: executing... 
[cont-init.d] 01-fix-uidgid.sh: exited 0.
[cont-init.d] 02-fix-perms.sh: executing... 
Fixing perms...
[cont-init.d] 02-fix-perms.sh: exited 0.
[cont-init.d] 03-config.sh: executing... 
Setting timezone to UTC...
Setting PHP-FPM configuration...
Setting PHP INI configuration...
Setting OpCache configuration...
Setting Nginx configuration...
Updating SNMP community...
Initializing LibreNMS files / folders...
Setting LibreNMS configuration...
Checking LibreNMS plugins...
Fixing perms...
Checking additional Monitoring plugins...
Checking alert templates...
[cont-init.d] 03-config.sh: exited 0.
[cont-init.d] 04-svc-main.sh: executing... 
Waiting 60s for database to be ready...
Database ready!
Updating database schema...

   INFO  Nothing to migrate.  


   INFO  Seeding database.  

  Database\Seeders\DefaultAlertTemplateSeeder ........................ RUNNING  
  Database\Seeders\DefaultAlertTemplateSeeder ...................... 1 ms DONE  

  Database\Seeders\ConfigSeeder ...................................... RUNNING  
  Database\Seeders\ConfigSeeder .................................... 1 ms DONE  

  Database\Seeders\RolesSeeder ....................................... RUNNING  
  Database\Seeders\RolesSeeder .................................... 13 ms DONE  

Clear cache

   INFO  Application cache cleared successfully.  


   INFO  Configuration cached successfully.  

[cont-init.d] 04-svc-main.sh: exited 0.
[cont-init.d] 05-svc-dispatcher.sh: executing... 
[cont-init.d] 05-svc-dispatcher.sh: exited 0.
[cont-init.d] 06-svc-syslogng.sh: executing... 
[cont-init.d] 06-svc-syslogng.sh: exited 0.
[cont-init.d] 07-svc-cron.sh: executing... 
Creating LibreNMS daily.sh cron task with the following period fields: 15 0 * * *
Creating LibreNMS cron artisan schedule:run
Fixing crontabs permissions...
[cont-init.d] 07-svc-cron.sh: exited 0.
[cont-init.d] 08-svc-snmptrapd.sh: executing... 
[cont-init.d] 08-svc-snmptrapd.sh: exited 0.
[cont-init.d] ~-socklog: executing... 
[cont-init.d] ~-socklog: exited 0.
[cont-init.d] done.
[services.d] starting services
[services.d] done.

Additional info

We first had other problems but get them solved by starting the pod with the privileged service account.

When i go to the pod-terminal and execute "/init" then the following log appears AND we are able to open the gui.
When we close the terminal window the service dissapears.

[s6-init] making user provided files available at /var/run/s6/etc...exited 0.
[s6-init] ensuring user provided files have correct perms...exited 0.
[fix-attrs.d] applying ownership & permissions fixes...
[fix-attrs.d] done.
[cont-init.d] executing container initialization scripts...
[cont-init.d] 00-fix-logs.sh: executing...
[cont-init.d] 00-fix-logs.sh: exited 0.
[cont-init.d] 01-fix-uidgid.sh: executing...
[cont-init.d] 01-fix-uidgid.sh: exited 0.
[cont-init.d] 02-fix-perms.sh: executing...
Fixing perms...
[cont-init.d] 02-fix-perms.sh: exited 0.
[cont-init.d] 03-config.sh: executing...
Setting timezone to UTC...
Setting PHP-FPM configuration...
Setting PHP INI configuration...
Setting OpCache configuration...
Setting Nginx configuration...
Updating SNMP community...
Initializing LibreNMS files / folders...
Setting LibreNMS configuration...
Checking LibreNMS plugins...
Fixing perms...
Checking additional Monitoring plugins...
Checking alert templates...
[cont-init.d] 03-config.sh: exited 0.
[cont-init.d] 04-svc-main.sh: executing...
Waiting 60s for database to be ready...
Database ready!
Updating database schema...

INFO Nothing to migrate.

INFO Seeding database.

Database\Seeders\DefaultAlertTemplateSeeder .............................................................................................. RUNNING
Database\Seeders\DefaultAlertTemplateSeeder ............................................................................................ 1 ms DONE

Database\Seeders\ConfigSeeder ............................................................................................................ RUNNING
Database\Seeders\ConfigSeeder .......................................................................................................... 1 ms DONE

Database\Seeders\RolesSeeder ............................................................................................................. RUNNING
Database\Seeders\RolesSeeder .......................................................................................................... 12 ms DONE

Clear cache

INFO Application cache cleared successfully.

INFO Configuration cached successfully.

[cont-init.d] 04-svc-main.sh: exited 0.
[cont-init.d] 05-svc-dispatcher.sh: executing...
[cont-init.d] 05-svc-dispatcher.sh: exited 0.
[cont-init.d] 06-svc-syslogng.sh: executing...
[cont-init.d] 06-svc-syslogng.sh: exited 0.
[cont-init.d] 07-svc-cron.sh: executing...
Creating LibreNMS daily.sh cron task with the following period fields: 15 0 * * *
Creating LibreNMS cron artisan schedule:run
Fixing crontabs permissions...
[cont-init.d] 07-svc-cron.sh: exited 0.
[cont-init.d] 08-svc-snmptrapd.sh: executing...
[cont-init.d] 08-svc-snmptrapd.sh: exited 0.
[cont-init.d] ~-socklog: executing...
[cont-init.d] ~-socklog: exited 0.
[cont-init.d] done.
[services.d] starting services
crond: crond (busybox 1.36.1) started, log level 8
[services.d] done.
2024/12/05 13:12:45 [notice] 1523#1523: using the "epoll" event method
2024/12/05 13:12:45 [notice] 1523#1523: nginx/1.24.0
2024/12/05 13:12:45 [notice] 1523#1523: OS: Linux 5.14.0-284.73.1.el9_2.x86_64
2024/12/05 13:12:45 [notice] 1523#1523: getrlimit(RLIMIT_NOFILE): 1048576:1048576
2024/12/05 13:12:45 [notice] 1523#1523: start worker processes
2024/12/05 13:12:45 [notice] 1523#1523: start worker process 1557
2024/12/05 13:12:45 [notice] 1523#1523: start worker process 1559
2024/12/05 13:12:45 [notice] 1523#1523: start worker process 1560
2024/12/05 13:12:45 [notice] 1523#1523: start worker process 1561
2024/12/05 13:12:45 [notice] 1523#1523: start worker process 1562
2024/12/05 13:12:45 [notice] 1523#1523: start worker process 1563
s6-log: fatal: unable to lock /var/log/socklog/cron/lock: Resource busy
[05-Dec-2024 13:12:45] NOTICE: fpm is running, pid 1519
[05-Dec-2024 13:12:45] NOTICE: ready to handle connections
s6-log: fatal: unable to lock /var/log/socklog/cron/lock: Resource busy
s6-log: fatal: unable to lock /var/log/socklog/cron/lock: Resource busy
s6-log: fatal: unable to lock /var/log/socklog/cron/lock: Resource busy
s6-log: fatal: unable to lock /var/log/socklog/cron/lock: Resource busy
s6-log: fatal: unable to lock /var/log/socklog/cron/lock: Resource busy
s6-log: fatal: unable to lock /var/log/socklog/cron/lock: Resource busy
s6-log: fatal: unable to lock /var/log/socklog/cron/lock: Resource busy

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions