# Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # Default values for monasca-api. # This is a YAML-formatted file. # Declare name/value pairs to be passed into your templates. # name: value --- images: tags: ks_user: docker.io/openstackhelm/heat:2024.1-ubuntu_jammy ks_service: docker.io/openstackhelm/heat:2024.1-ubuntu_jammy ks_endpoints: docker.io/openstackhelm/heat:2024.1-ubuntu_jammy influxdb_init: radial/busyboxplus:curl db_init: docker.io/openstackhelm/heat:2024.1-ubuntu_jammy monasca_db_sync: docker.io/openstackhelm/monasca:stein-ubuntu_bionic db_drop: docker.io/openstackhelm/heat:2024.1-ubuntu_jammy monasca_api: docker.io/openstackhelm/monasca_api:stein-ubuntu_bionic monasca_collector: docker.io/openstackhelm/monasca_agent:stein-ubuntu_bionic monasca_forwarder: docker.io/openstackhelm/monasca_forwarder:stein-ubuntu_bionic monasca_statsd: docker.io/openstackhelm/monasca_statsd:stein-ubuntu_bionic monasca_notification: docker.io/openstackhelm/monasca_notification:stein-ubuntu_bionic monasca_persister: docker.io/openstackhelm/monasca_persister:stein-ubuntu_bionic monasca_thresh: docker.io/openstackhelm/monasca_thresh:stein-ubuntu_bionic test: docker.io/openstackhelm/osh-selenium:latest-ubuntu_focal dep_check: quay.io/airshipit/kubernetes-entrypoint:v1.0.0 image_repo_sync: docker.io/docker:17.07.0 pull_policy: "IfNotPresent" local_registry: active: false exclude: - dep_check - image_repo_sync release_group: null labels: agent: node_selector_key: openstack-compute-node node_selector_value: enabled api: node_selector_key: openstack-control-plane node_selector_value: enabled notification: node_selector_key: openstack-control-plane node_selector_value: enabled persister: node_selector_key: openstack-control-plane node_selector_value: enabled thresh: node_selector_key: openstack-control-plane node_selector_value: enabled job: node_selector_key: openstack-control-plane node_selector_value: enabled network: api: ingress: public: true classes: namespace: "nginx" cluster: "nginx-cluster" annotations: nginx.ingress.kubernetes.io/rewrite-target: / external_policy_local: false node_port: enabled: false port: 31000 conf: agent: Api: amplifier: 0 backlog_send_rate: 1000 ca_file: null insecure: false max_batch_size: 0 max_buffer_size: 1000 max_measurement_buffer_size: -1 service_type: monitoring Logging: disable_file_logging: false collector_log_file: /var/log/monasca/agent/collector.log forwarder_log_file: /var/log/monasca/agent/forwarder.log statsd_log_file: /var/log/monasca/agent/statsd.log enable_logrotate: true log_level: WARN log_to_event_viewer: false log_to_syslog: false syslog_host: None syslog_port: None Main: check_freq: 15 autorestart: false collector_restart_interval: 24 dimensions: service: monitoring # %AGENT_HOSTNAME% is replaced with the correct value in the init container. hostname: "%AGENT_HOSTNAME%" # %FORWARDER_IP% is replaced with the correct value in the init container. forwarder_url: "http://%FORWARDER_IP%:17123" num_collector_threads: 1 pool_full_max_retries: 4 sub_collection_warn: 6 non_local_traffic: true Statsd: monasca_statsd_interval: 20 monasca_statsd_port: 8125 # monasca_statsd_forward_host: None # monasca_statsd_forward_port: 8125 agent_plugins: # Plugin config has 2 fields; # auto_detect: specify if it is auto configured by monasca-setup or not # config: plugin specific configurations. # when auto_detect is false, it is rendered as config yaml file # when auto_detect is true, it is used as args in monasca-setup libvirt: auto_detect: false config: init_config: {} # auto_detect: true # config: # nova_refresh: # vm_probation: # Custom check plugins. Refer values_overrides/nvidia.yaml. check_scripts: {} # Custom detection plugins. Refer values_overrides/nvidia.yaml. detection_scripts: {} monasca_api: DEFAULT: enable_logs_api: true region: RegionOne log_config_append: /etc/monasca/api-logging.conf messaging: driver: monasca_api.common.messaging.kafka_publisher:KafkaPublisher kafka: logs_topics: log uri: 127.0.0.1:9092 repositories: metrics_driver: monasca_api.common.repositories.influxdb.metrics_repository:MetricsRepository cassandra: contact_points: 127.0.0.1 influxdb: db_per_tenant: false database_name: monasca database: max_retries: -1 keystone_authtoken: auth_type: password auth_version: v3 memcache_security_strategy: ENCRYPT security: default_authorized_roles: monasca-user agent_authorized_roles: monasca-agent read_only_authorized_roles: monasca-read-only-user delegate_authorized_roles: monasca-agent monasca_api_paste: DEFAULT: name: monasca_api pipeline:main: pipeline: request_id auth api app:api: paste.app_factory: monasca_api.api.server:launch filter:auth: paste.filter_factory: monasca_api.healthcheck.keystone_protocol:filter_factory filter:request_id: paste.filter_factory: oslo_middleware.request_id:RequestId.factory server:main: use: egg:gunicorn#main host: 0.0.0.0 port: 8070 workers: 8 worker-connections: 2000 worker-class: eventlet timeout: 30 backlog: 2048 keepalive: 2 proc_name: monasca-api loglevel: DEBUG logging: loggers: keys: - root - sqlalchemy - kafka - kafkalib handlers: keys: - stdout - stderr - "null" formatters: keys: - context - default logger_root: level: INFO handlers: 'stdout' logger_sqlalchemy: qualname: sqlalchemy.engine level: DEBUG handlers: 'stdout' propagate: 0 logger_kafka: qualname: kafka level: DEBUG handlers: 'stdout' propagate: 0 logger_kafkalib: qualname: monasca_common.kafka_lib level: INFO handlers: 'stdout' propagate: 0 handler_null: class: logging.NullHandler formatter: default args: () handler_stdout: class: StreamHandler args: (sys.stdout,) formatter: context handler_stderr: class: StreamHandler args: (sys.stderr,) formatter: context formatter_context: class: oslo_log.formatters.ContextFormatter datefmt: "%Y-%m-%d %H:%M:%S" formatter_default: format: "%(message)s" datefmt: "%Y-%m-%d %H:%M:%S" wsgi_monasca_api: | {{- $portInt := tuple "monitoring" "internal" "api" $ | include "helm-toolkit.endpoints.endpoint_port_lookup" }} Listen 0.0.0.0:{{ $portInt }} LogFormat "%h %l %u %t \"%r\" %>s %b \"%{Referer}i\" \"%{User-Agent}i\"" combined LogFormat "%{X-Forwarded-For}i %l %u %t \"%r\" %>s %b \"%{Referer}i\" \"%{User-Agent}i\"" proxy SetEnvIf X-Forwarded-For "^.*\..*\..*\..*" forwarded CustomLog /dev/stdout combined env=!forwarded CustomLog /dev/stdout proxy env=forwarded WSGIDaemonProcess monasca-api processes=1 threads=1 user=monasca-api group=monasca-api display-name=%{GROUP} WSGIProcessGroup monasca-api WSGIScriptAlias / /var/www/cgi-bin/monasca/monasca-api-wsgi WSGIApplicationGroup %{GLOBAL} WSGIPassAuthorization On = 2.4> ErrorLogFormat "%{cu}t %M" ErrorLog /dev/stdout SetEnvIf X-Forwarded-For "^.*\..*\..*\..*" forwarded CustomLog /dev/stdout combined env=!forwarded CustomLog /dev/stdout proxy env=forwarded thresh_config: metricSpoutThreads: 2 metricSpoutTasks: 2 statsdConfig: host: "127.0.0.1" port: 8125 prefix: monasca.storm. dimensions: !!map service: monitoring component: storm metricSpoutConfig: kafkaConsumerConfiguration: # See http://kafka.apache.org/documentation.html#api for semantics and defaults. topic: "metrics" numThreads: 1 groupId: "thresh-metric" zookeeperConnect: "127.0.0.1:2181" consumerId: 1 socketTimeoutMs: 30000 socketReceiveBufferBytes: 65536 fetchMessageMaxBytes: 1048576 autoCommitEnable: true autoCommitIntervalMs: 60000 queuedMaxMessageChunks: 10 rebalanceMaxRetries: 4 fetchMinBytes: 1 fetchWaitMaxMs: 100 rebalanceBackoffMs: 2000 refreshLeaderBackoffMs: 200 autoOffsetReset: largest consumerTimeoutMs: -1 clientId: 1 zookeeperSessionTimeoutMs: 60000 zookeeperConnectionTimeoutMs: 60000 zookeeperSyncTimeMs: 2000 eventSpoutConfig: kafkaConsumerConfiguration: # See http://kafka.apache.org/documentation.html#api for semantics and defaults. topic: "events" numThreads: 1 groupId: "thresh-event" zookeeperConnect: "127.0.0.1:2181" consumerId: 1 socketTimeoutMs: 30000 socketReceiveBufferBytes: 65536 fetchMessageMaxBytes: 1048576 autoCommitEnable: true autoCommitIntervalMs: 60000 queuedMaxMessageChunks: 10 rebalanceMaxRetries: 4 fetchMinBytes: 1 fetchWaitMaxMs: 100 rebalanceBackoffMs: 2000 refreshLeaderBackoffMs: 200 autoOffsetReset: largest consumerTimeoutMs: -1 clientId: 1 zookeeperSessionTimeoutMs: 60000 zookeeperConnectionTimeoutMs: 60000 zookeeperSyncTimeMs: 2000 kafkaProducerConfig: # See http://kafka.apache.org/documentation.html#api for semantics and defaults. topic: "alarm-state-transitions" metadataBrokerList: "kafka:9092" serializerClass: kafka.serializer.StringEncoder partitionerClass: "" requestRequiredAcks: 1 requestTimeoutMs: 10000 producerType: sync keySerializerClass: "" compressionCodec: none compressedTopics: "" messageSendMaxRetries: 3 retryBackoffMs: 100 topicMetadataRefreshIntervalMs: 600000 queueBufferingMaxMs: 5000 queueBufferingMaxMessages: 10000 queueEnqueueTimeoutMs: -1 batchNumMessages: 200 sendBufferBytes: 102400 clientId: Threshold_Engine sporadicMetricNamespaces: - foo database: driverClass: org.mariadb.jdbc.Driver url: "jdbc:mariadb://%THRESH_DB_URL%" user: "%THRESH_DB_USER%" password: "%THRESH_DB_PASSWORD%" properties: ssl: false # the maximum amount of time to wait on an empty pool before throwing an exception maxWaitForConnection: 1s # the SQL query to run when validating a connection's liveness validationQuery: "/* mysql Health Check */ SELECT 1" # the minimum number of connections to keep open minSize: 8 # the maximum number of connections to keep open maxSize: 41 notification_config: kafka: url: kafka:9092 database: repo_driver: monasca_notification.common.repositories.mysql.mysql_repo:MysqlRepo email_notifier: grafana_url: grafana:3000 # from_addr: string # server: email_server # port: email_port # timeout: 5 # min value is 1 # user: email_user # password: email_password jira_notifier: {} # timeout: 5 # min value is 1 # user: # password: # custom_formatter: # proxy: pagerduty_notifier: {} # timeout: # url: slack_notifier: {} # timeout: # insecure: # ca_certs: # proxy: # message_template: webhook_notifier: {} # timeout: 5 keystone: {} # auth_url: 199.204.45.231/identity/v3 notification_types: enabled: - email - pagerduty - webhook - jira - slack mysql: {} zookeeper: url: ["127.0.0.1:2181"] persister_config: DEFAULT: debug: "True" default_log_levels: "monasca_common.kafka_lib.client=INFO" logging_exception_prefix: "ERROR %(name)s %(instance)s" logging_default_format_string: "%(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s" logging_context_format_string": "%(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s" logging_debug_format_suffix: "{{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}}" use_syslog: "False" influxdb: database_name: monasca db_per_tenant: false kafka: num_processors: 1 kafka_alarm_history: batch_size: 1 topic: alarm-state-transitions group_id: 1_alarm-state-transitions uri: kafka:9092 kafka_events: {} kafka_metrics: batch_size: 30 topic: metrics group_id: 1_metrics uri: kafka:9092 repositories: alarm_state_history_driver: monasca_persister.repositories.influxdb.alarm_state_history_repository:AlarmStateHistInfluxdbRepository metrics_driver: monasca_persister.repositories.influxdb.metrics_repository:MetricInfluxdbRepository zookeeper: uri: zookeeper:2181 storm: java.library.path: "/usr/local/lib:/opt/local/lib:/usr/lib" storm.local.dir: "/var/storm" storm.zookeeper.servers: - "127.0.0.1" storm.zookeeper.port: 2181 storm.zookeeper.retry.interval: 5000 storm.zookeeper.retry.times: 60 storm.zookeeper.root: /storm storm.zookeeper.session.timeout: 3000 supervisor.slots.ports: - 6701 - 6702 supervisor.childopts: -Xmx256m worker.childopts: -Xmx1280m -XX:+UseConcMarkSweepGC -Dcom.sun.management.jmxremote nimbus.seeds: ["127.0.0.1"] nimbus.thrift.port: 6627 nimbus.childopts: -Xmx256m ui.host: localhost ui.port: 8089 ui.childopts: -Xmx768m logviewer.port: 8090 logviewer.childopts: -Xmx128m transactional.zookeeper.servers: - "127.0.0.1" transactional.zookeeper.port: 2181 transactional.zookeeper.root: /storm-transactional topology.acker.executors: 1 topology.debug: False software: apache2: binary: apache2 start_parameters: -DFOREGROUND conf_dir: /etc/apache2/conf-enabled site_dir: /etc/apache2/sites-enable mods_dir: /etc/apache2/mods-available a2enmod: null a2dismod: null pod: probes: rpc_timeout: 60 rpc_retries: 2 api: default: liveness: enabled: true params: initialDelaySeconds: 120 periodSeconds: 90 timeoutSeconds: 70 readiness: enabled: true params: initialDelaySeconds: 80 periodSeconds: 90 timeoutSeconds: 70 forwarder: default: liveness: enabled: true params: initialDelaySeconds: 120 periodSeconds: 90 timeoutSeconds: 70 readiness: enabled: true params: initialDelaySeconds: 80 periodSeconds: 90 timeoutSeconds: 70 statsd: default: {} security_context: agent: pod: runAsUser: 42424 container: agent_init: readOnlyRootFilesystem: false allowPrivilegeEscalation: false runAsUser: 0 monasca_collector: readOnlyRootFilesystem: false allowPrivilegeEscalation: false runAsUser: 0 monasca_forwarder: readOnlyRootFilesystem: false allowPrivilegeEscalation: false runAsUser: 0 monasca_statsd: readOnlyRootFilesystem: false allowPrivilegeEscalation: false runAsUser: 0 api: pod: runAsUser: 42424 container: monasca-api: readOnlyRootFilesystem: false allowPrivilegeEscalation: false runAsUser: 0 notification: pod: runAsUser: 42424 container: monasca-notification: readOnlyRootFilesystem: false allowPrivilegeEscalation: false runAsUser: 0 persister: pod: runAsUser: 42424 container: monasca-persister: readOnlyRootFilesystem: false allowPrivilegeEscalation: false runAsUser: 0 thresh: pod: runAsUser: 42424 container: monasca-thresh: readOnlyRootFilesystem: false allowPrivilegeEscalation: false runAsUser: 0 db_sync: pod: runAsUser: 42424 container: monasca_db_sync: readOnlyRootFilesystem: false allowPrivilegeEscalation: false runAsUser: 0 influxdb_init: pod: runAsUser: 42424 container: influxdb_init: readOnlyRootFilesystem: false allowPrivilegeEscalation: false runAsUser: 0 test: pod: runAsUser: 42424 container: monasca_test: readOnlyRootFilesystem: true allowPrivilegeEscalation: false affinity: anti: type: default: preferredDuringSchedulingIgnoredDuringExecution topologyKey: default: kubernetes.io/hostname weight: default: 10 dns_policy: "ClusterFirstWithHostNet" mounts: monasca_agent: init_container: null monasca_collector: volumeMounts: - name: varliblibvirt mountPath: /var/lib/libvirt readOnly: true - mountPath: /lib/modules name: libmodules readOnly: true - name: varlibnova mountPath: /var/lib/nova volumes: - name: libmodules hostPath: path: /lib/modules - name: varliblibvirt hostPath: path: /var/lib/libvirt - name: varlibnova hostPath: path: /var/lib/nova monasca_forwarder: volumeMounts: volumes: monasca_statsd: volumeMounts: volumes: monasca_db_init: init_container: null monasca_db_init: volumeMounts: volumes: monasca_db_sync: init_container: null monasca_db_sync: volumeMounts: volumes: monasca_api: init_container: null monasca_api: volumeMounts: volumes: monasca_notification: init_container: null monasca_notification: volumeMounts: volumes: monasca_persister: init_container: null monasca_persister: volumeMounts: volumes: monasca_thresh: init_container: null monasca_thresh: volumeMounts: volumes: monasca_tests: init_container: null monasca_tests: volumeMounts: volumes: replicas: agent: 1 api: 1 notification: 1 persister: 1 thresh: 1 lifecycle: upgrades: deployments: revision_history: 3 pod_replacement_strategy: RollingUpdate rolling_update: max_unavailable: 1 max_surge: 3 daemonsets: pod_replacement_strategy: RollingUpdate agent: enabled: true min_ready_seconds: 0 max_unavailable: 1 disruption_budget: monasca: min_available: 0 termination_grace_period: monasca: timeout: 30 resources: enabled: false agent_init: requests: memory: "128Mi" cpu: "100m" limits: memory: "512Mi" cpu: "500m" api: requests: memory: "128Mi" cpu: "100m" limits: memory: "1024Mi" cpu: "2000m" collector: requests: memory: "128Mi" cpu: "100m" limits: memory: "512Mi" cpu: "500m" forwarder: requests: memory: "128Mi" cpu: "100m" limits: memory: "512Mi" cpu: "500m" statsd: requests: memory: "128Mi" cpu: "100m" limits: memory: "512Mi" cpu: "500m" notification: requests: memory: "128Mi" cpu: "100m" limits: memory: "512Mi" cpu: "500m" persister: requests: memory: "128Mi" cpu: "100m" limits: memory: "512Mi" cpu: "500m" thresh: requests: memory: "128Mi" cpu: "100m" limits: memory: "512Mi" cpu: "500m" jobs: db_init: requests: memory: "128Mi" cpu: "100m" limits: memory: "1024Mi" cpu: "2000m" db_sync: requests: memory: "128Mi" cpu: "100m" limits: memory: "1024Mi" cpu: "2000m" db_drop: requests: memory: "128Mi" cpu: "100m" limits: memory: "1024Mi" cpu: "2000m" image_repo_sync: requests: memory: "128Mi" cpu: "100m" limits: memory: "1024Mi" cpu: "2000m" influxdb_init: requests: memory: "128Mi" cpu: "100m" limits: memory: "1024Mi" cpu: "2000m" tests: requests: memory: "128Mi" cpu: "100m" limits: memory: "1024Mi" cpu: "2000m" bootstrap: enabled: false ks_user: monasca script: null # Names of secrets used by bootstrap and environmental checks secrets: identity: admin: monasca-keystone-admin monasca: monasca-keystone-user monasca_agent: monasca-keystone-agent monasca_read_only_user: monasca-keystone-ro test: monasca-keystone-test metrics_db: monasca-influxdb oslo_db: admin: monasca-db-admin monasca: monasca-db-user oslo_messaging: admin: monasca-rabbitmq-admin monasca: monasca-rabbitmq-user tls: monitoring: api: public: monasca-tls-public internal: monasca-tls-api statsd: public: monasca-statsd-tls-public internal: monasca-tls-statsd events_api: api: public: monasca-events-tls-public internal: monasca-events-tls-proxy # typically overridden by environmental # values, but should include all endpoints # required by this chart endpoints: cluster_domain_suffix: cluster.local monitoring: name: monasca hosts: default: monasca-api public: monasca-api host_fqdn_override: default: null path: default: /v2.0 scheme: default: 'http' port: api: default: 8070 public: 80 statsd: default: 8125 public: 8125 forwarder: default: 17123 public: 17123 logs_search: name: monasca hosts: default: kibana public: kibana host_fqdn_override: default: null path: default: / scheme: default: 'http' port: api: default: 5601 public: 80 logs: name: monasca hosts: default: monasca-api public: monasca-api host_fqdn_override: default: null path: default: /v2.0 scheme: default: 'http' port: api: default: 8070 public: 80 local_image_registry: name: docker-registry namespace: docker-registry hosts: default: localhost internal: docker-registry node: localhost host_fqdn_override: default: null port: registry: node: 5000 identity: name: keystone auth: admin: region_name: RegionOne username: admin password: password project_name: admin user_domain_name: default project_domain_name: default monasca: role: - admin - monasca-user region_name: RegionOne username: monasca password: password project_name: admin user_domain_name: default project_domain_name: default monasca_agent: role: monasca-agent region_name: RegionOne username: monasca-agent password: password project_name: service user_domain_name: service project_domain_name: service interface: private monasca_read_only_user: role: monasca-read-only-user region_name: RegionOne username: monasca-read-only-user password: password project_name: service user_domain_name: service project_domain_name: service hosts: default: keystone internal: keystone-api host_fqdn_override: default: null path: default: /v3 scheme: default: http port: api: default: 80 internal: 5000 oslo_cache: auth: # NOTE(portdirect): this is used to define the value for keystone # authtoken cache encryption key, if not set it will be populated # automatically with a random value, but to take advantage of # this feature all services should be set to use the same key, # and memcache service. memcache_secret_key: null hosts: default: memcached host_fqdn_override: default: null port: memcache: default: 11211 oslo_db: auth: admin: username: root password: password secret: tls: internal: mariadb-tls-direct monasca: username: monasca password: password hosts: default: mariadb host_fqdn_override: default: null path: /monasca scheme: mysql+pymysql port: mysql: default: 3306 metrics_db: auth: admin: username: admin password: password monasca_api: username: monasca-api password: password monasca_persister: username: monasca-persister password: password hosts: default: influxdb host_fqdn_override: default: null path: / scheme: http port: influxdb: default: 8086 # NOTE(tp6510): these endpoints allow for things like DNS lookups and ingress # They are using to enable the Egress K8s network policy. kube_dns: namespace: kube-system name: kubernetes-dns hosts: default: kube-dns host_fqdn_override: default: null path: default: null scheme: http port: dns: default: 53 protocol: UDP ingress: namespace: null name: ingress hosts: default: ingress port: ingress: default: 80 network_policy: monasca: ingress: - {} egress: - {} dependencies: dynamic: common: local_image_registry: jobs: - image-repo-sync services: - endpoint: node service: local_image_registry static: agent: services: - endpoint: internal service: monitoring api: jobs: - monasca-db-sync - monasca-ks-service - monasca-ks-user - monasca-ks-endpoints services: - endpoint: internal service: oslo_db - endpoint: internal service: identity notification: jobs: - monasca-db-sync services: - endpoint: internal service: oslo_db # - kafka persister: jobs: - monasca-influxdb-init services: - endpoint: internal service: metrics_db thresh: services: - endpoint: internal service: oslo_db # - storm ks_endpoints: jobs: - monasca-ks-service services: - endpoint: internal service: identity ks_service: services: - endpoint: internal service: identity ks_user: services: - endpoint: internal service: identity db_drop: services: - endpoint: internal service: oslo_db db_init: services: - endpoint: internal service: oslo_db db_sync: jobs: - monasca-db-init services: - endpoint: internal service: oslo_db influxdb_init: services: - endpoint: internal service: metrics_db manifests: certificates: false configmap_bin: true configmap_etc: true daemonset_agent: true deployment_agent: false deployment_api: true deployment_notification: true deployment_persister: true ingress_api: true job_db_init: true job_db_sync: true job_db_drop: false job_influxdb_init: true job_thresh: true job_image_repo_sync: true job_rabbit_init: false job_ks_endpoints: true job_ks_service: true job_ks_user: true network_policy: false secret_db: true secret_ingress_tls: true secret_influxdb: true secret_keystone: true service_ingress: true service: true ...