Implement alarm parsing
Configure tox+content to fetch event and convert alarms and logs to rst for use in build. Handle non-existant tmp dir in zuul builds Add static events.yaml for CI/CD testingx Generalize label construction to prevent namespace conflicts Consume events directly from fm repo (required changes merged) Update logs template for legibility. Add clean up for temporary rst files. Point parser at dynamically downloaded events file Restore logs template Note: This review deletes static alarm and log files Note: This review excludes alarm files from git as they are now build-time temp files. Note: This review uses a static copy of events.yaml to pass tox until the dep. below is met. It will need reconfiguration at that time. Depends-On: https://review.opendev.org/c/starlingx/fault/+/863574 Signed-off-by: Ron Stone <ronald.stone@windriver.com> Change-Id: I0bb8d0a77b9d3cf22b33f8930c569b3e70b7291c
This commit is contained in:
parent
2192558b91
commit
c3444e384d
4
.gitignore
vendored
4
.gitignore
vendored
@ -66,6 +66,10 @@ instance/
|
||||
# Sphinx documentation
|
||||
docs/_build/
|
||||
.pickOptions.sh
|
||||
tmp/
|
||||
# templates/events.yaml
|
||||
*-series-log-messages.rst
|
||||
*-series-alarm-messages.rst
|
||||
|
||||
|
||||
# API Reference Guide
|
||||
|
@ -1,6 +1,7 @@
|
||||
sphinx==4.2.0
|
||||
openstackdocstheme>=2.2.1,<=2.3.1 # Apache-2.0
|
||||
docutils==0.17.1
|
||||
PyYAML==6.0
|
||||
|
||||
# API Reference Guide
|
||||
os-api-ref>=1.5.0 # Apache-2.0
|
||||
|
@ -17,7 +17,7 @@ If you want to use the Horizon Web interface, see
|
||||
Management-affecting alarms cannot be ignored using relaxed alarm rules
|
||||
during an orchestrated Kubernetes version upgrade operation. For a list of
|
||||
management-affecting alarms, see |fault-doc|:
|
||||
:ref:`Alarm Messages <100-series-alarm-messages>`. To display
|
||||
:ref:`Alarm Messages <100-series-alarm-messages-starlingx>`. To display
|
||||
management-affecting active alarms, use the following command:
|
||||
|
||||
.. code-block:: none
|
||||
|
@ -1,368 +0,0 @@
|
||||
|
||||
.. jsy1579701868527
|
||||
.. _100-series-alarm-messages:
|
||||
|
||||
=========================
|
||||
100 Series Alarm Messages
|
||||
=========================
|
||||
|
||||
The system inventory and maintenance service reports system changes with
|
||||
different degrees of severity. Use the reported alarms to monitor the overall
|
||||
health of the system.
|
||||
|
||||
.. include:: /_includes/x00-series-alarm-messages.rest
|
||||
|
||||
.. _100-series-alarm-messages-table-zrd-tg5-v5:
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.101**
|
||||
- Platform CPU threshold exceeded; threshold x%, actual y%.
|
||||
CRITICAL @ 95%
|
||||
|
||||
MAJOR @ 90%
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- Critical
|
||||
* - Severity:
|
||||
- C/M\*
|
||||
* - Proposed Repair Action
|
||||
- Monitor and if condition persists, contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.103**
|
||||
- Memory threshold exceeded; threshold x%, actual y% .
|
||||
|
||||
CRITICAL @ 90%
|
||||
|
||||
MAJOR @ 80%
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- Critical
|
||||
* - Severity:
|
||||
- C/M
|
||||
* - Proposed Repair Action
|
||||
- Monitor and if condition persists, contact next level of support; may
|
||||
require additional memory on Host.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.104**
|
||||
- host=<hostname>.filesystem=<mount-dir>
|
||||
File System threshold exceeded; threshold x%, actual y%.
|
||||
|
||||
CRITICAL @ 90%
|
||||
|
||||
MAJOR @ 80%
|
||||
|
||||
OR
|
||||
|
||||
host=<hostname>.volumegroup=<volumegroup-name>
|
||||
Monitor and if condition persists, consider adding additional
|
||||
physical volumes to the volume group.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.filesystem=<mount-dir>
|
||||
|
||||
OR
|
||||
|
||||
host=<hostname>.volumegroup=<volumegroup-name>
|
||||
* - Degrade Affecting Severity:
|
||||
- Critical
|
||||
* - Severity:
|
||||
- C\*/M
|
||||
* - Proposed Repair Action
|
||||
- Reduce usage or resize filesystem.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.105**
|
||||
- <fs\_name\> filesystem is not added on both controllers and/or does not
|
||||
have the same size: <hostname\>.
|
||||
* - Entity Instance
|
||||
- fs\_name=<image-conversion>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C/M\*
|
||||
* - Proposed Repair Action
|
||||
- Add image-conversion filesystem on both controllers.
|
||||
|
||||
Consult the System Administration Manual for more details.
|
||||
|
||||
If problem persists, contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.106**
|
||||
- 'OAM' Port failed.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.port=<port-name>
|
||||
* - Degrade Affecting Severity:
|
||||
- Major
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.107**
|
||||
- 'OAM' Interface degraded.
|
||||
|
||||
or
|
||||
|
||||
'OAM' Interface failed.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.interface=<if-name>
|
||||
* - Degrade Affecting Severity:
|
||||
- Major
|
||||
* - Severity:
|
||||
- C or M\*
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.108**
|
||||
- 'MGMT' Port failed.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.port=<port-name>
|
||||
* - Degrade Affecting Severity:
|
||||
- Major
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.109**
|
||||
- 'MGMT' Interface degraded.
|
||||
|
||||
or
|
||||
|
||||
'MGMT' Interface failed.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.interface=<if-name>
|
||||
* - Degrade Affecting Severity:
|
||||
- Major
|
||||
* - Severity:
|
||||
- C or M\*
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.110**
|
||||
- 'CLUSTER-HOST' Port failed.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.port=<port-name>
|
||||
* - Degrade Affecting Severity:
|
||||
- Major
|
||||
* - Severity:
|
||||
- C or M\*
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.111**
|
||||
- 'CLUSTER-HOST' Interface degraded.
|
||||
|
||||
OR
|
||||
|
||||
'CLUSTER-HOST' Interface failed.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.interface=<if-name>
|
||||
* - Degrade Affecting Severity:
|
||||
- Major
|
||||
* - Severity:
|
||||
- C or M\*
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.112**
|
||||
- 'DATA-VRS' Port down.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.port=<port-name>
|
||||
* - Degrade Affecting Severity:
|
||||
- Major
|
||||
* - Severity:
|
||||
- M
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.113**
|
||||
- 'DATA-VRS' Interface degraded.
|
||||
|
||||
or
|
||||
|
||||
'DATA-VRS' Interface down.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.interface=<if-name>
|
||||
* - Degrade Affecting Severity:
|
||||
- Major
|
||||
* - Severity:
|
||||
- C or M\*
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.114**
|
||||
- NTP configuration does not contain any valid or reachable NTP servers.
|
||||
The alarm is raised regardless of NTP enabled/disabled status.
|
||||
|
||||
NTP address <IP address> is not a valid or a reachable NTP server.
|
||||
|
||||
Connectivity to external PTP Clock Synchronization is lost.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.ntp
|
||||
|
||||
host=<hostname>.ntp=<IP address>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M or m
|
||||
* - Proposed Repair Action
|
||||
- Monitor and if condition persists, contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.118**
|
||||
- Controller cannot establish connection with remote logging server.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- m
|
||||
* - Proposed Repair Action
|
||||
- Ensure Remote Log Server IP is reachable from Controller through OAM
|
||||
interface; otherwise contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 100.119**
|
||||
- <hostname> does not support the provisioned PTP mode
|
||||
|
||||
OR
|
||||
|
||||
<hostname> PTP clocking is out-of-tolerance
|
||||
|
||||
OR
|
||||
|
||||
<hostname> is not locked to remote PTP Grand Master (|PTS|)
|
||||
|
||||
OR
|
||||
|
||||
<hostname> GNSS signal loss state:<state>
|
||||
|
||||
OR
|
||||
|
||||
<hostname> 1PPS signal loss state:<state>
|
||||
* - Entity Instance
|
||||
- host=<hostname>.ptp OR host=<hostname>.ptp=no-lock
|
||||
|
||||
OR
|
||||
|
||||
host=<hostname>.ptp=<interface>.unsupported=hardware-timestamping
|
||||
|
||||
OR
|
||||
|
||||
host=<hostname>.ptp=<interface>.unsupported=software-timestamping
|
||||
|
||||
OR
|
||||
|
||||
host=<hostname>.ptp=<interface>.unsupported=legacy-timestamping
|
||||
|
||||
OR
|
||||
|
||||
host=<hostname>.ptp=out-of-tolerance
|
||||
|
||||
OR
|
||||
|
||||
host=<hostname>.instance=<instance>.ptp=out-of-tolerance
|
||||
|
||||
OR
|
||||
|
||||
host=<hostname>.interface=<interface>.ptp=signal-loss
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M or m
|
||||
* - Proposed Repair Action
|
||||
- Monitor and, if condition persists, contact next level of support.
|
@ -1,389 +0,0 @@
|
||||
|
||||
.. uof1579701912856
|
||||
.. _200-series-alarm-messages:
|
||||
|
||||
=========================
|
||||
200 Series Alarm Messages
|
||||
=========================
|
||||
|
||||
The system inventory and maintenance service reports system changes with
|
||||
different degrees of severity. Use the reported alarms to monitor the overall
|
||||
health of the system.
|
||||
|
||||
.. include:: /_includes/x00-series-alarm-messages.rest
|
||||
|
||||
.. _200-series-alarm-messages-table-zrd-tg5-v5:
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.001**
|
||||
- <hostname> was administratively locked to take it out-of-service.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- W\*
|
||||
* - Proposed Repair Action
|
||||
- Administratively unlock Host to bring it back in-service.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.004**
|
||||
- <hostname> experienced a service-affecting failure.
|
||||
|
||||
Host is being auto recovered by Reboot.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- If auto-recovery is consistently unable to recover host to the
|
||||
unlocked-enabled state contact next level of support or lock and replace
|
||||
failing host.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.005**
|
||||
- Degrade:
|
||||
|
||||
<hostname> is experiencing an intermittent 'Management Network'
|
||||
communication failures that have exceeded its lower alarming threshold.
|
||||
|
||||
Failure:
|
||||
|
||||
<hostname> is experiencing a persistent Critical 'Management Network'
|
||||
communication failure.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\* (Degrade) or C\* (Failure)
|
||||
* - Proposed Repair Action
|
||||
- Check 'Management Network' connectivity and support for multicast
|
||||
messaging. If problem consistently occurs after that and Host is reset,
|
||||
then contact next level of support or lock and replace failing host.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.006**
|
||||
- Main Process Monitor Daemon Failure \(Major\)
|
||||
|
||||
<hostname> 'Process Monitor' \(pmond\) process is not running or
|
||||
functioning properly. The system is trying to recover this process.
|
||||
|
||||
Monitored Process Failure \(Critical/Major/Minor\)
|
||||
|
||||
Critical: <hostname> Critical '<processname>' process has failed and
|
||||
could not be auto-recovered gracefully. Auto-recovery progression by
|
||||
host reboot is required and in progress.
|
||||
|
||||
Major: <hostname> is degraded due to the failure of its '<processname>'
|
||||
process. Auto recovery of this Major process is in progress.
|
||||
|
||||
Minor:
|
||||
|
||||
<hostname> '<processname>' process has failed. Auto recovery of this
|
||||
Minor process is in progress.
|
||||
|
||||
<hostname> '<processname>' process has failed. Manual recovery is required.
|
||||
|
||||
tp4l/phc2sys process failure. Manual recovery is required.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.process=<processname>
|
||||
* - Degrade Affecting Severity:
|
||||
- Major
|
||||
* - Severity:
|
||||
- C/M/m\*
|
||||
* - Proposed Repair Action
|
||||
- If this alarm does not automatically clear after some time and continues
|
||||
to be asserted after Host is locked and unlocked then contact next level
|
||||
of support for root cause analysis and recovery.
|
||||
|
||||
If problem consistently occurs after Host is locked and unlocked then
|
||||
contact next level of support for root cause analysis and recovery.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.007**
|
||||
- Critical: \(with host degrade\):
|
||||
|
||||
Host is degraded due to a 'Critical' out-of-tolerance reading from the
|
||||
'<sensorname>' sensor
|
||||
|
||||
Major: \(with host degrade\)
|
||||
|
||||
Host is degraded due to a 'Major' out-of-tolerance reading from the
|
||||
'<sensorname>' sensor
|
||||
|
||||
Minor:
|
||||
|
||||
Host is reporting a 'Minor' out-of-tolerance reading from the
|
||||
'<sensorname>' sensor
|
||||
* - Entity Instance
|
||||
- host=<hostname>.sensor=<sensorname>
|
||||
* - Degrade Affecting Severity:
|
||||
- Critical
|
||||
* - Severity:
|
||||
- C/M/m
|
||||
* - Proposed Repair Action
|
||||
- If problem consistently occurs after Host is power cycled and or reset,
|
||||
contact next level of support or lock and replace failing host.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.009**
|
||||
- Degrade:
|
||||
|
||||
<hostname> is experiencing an intermittent 'Cluster-host Network'
|
||||
communication failures that have exceeded its lower alarming threshold.
|
||||
|
||||
Failure:
|
||||
|
||||
<hostname> is experiencing a persistent Critical 'Cluster-host Network'
|
||||
communication failure.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\* (Degrade) or C\* (Critical)
|
||||
* - Proposed Repair Action
|
||||
- Check 'Cluster-host Network' connectivity and support for multicast
|
||||
messaging. If problem consistently occurs after that and Host is reset,
|
||||
then contact next level of support or lock and replace failing host.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.010**
|
||||
- <hostname> access to board management module has failed.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- W
|
||||
* - Proposed Repair Action
|
||||
- Check Host's board management configuration and connectivity.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.011**
|
||||
- <hostname> experienced a configuration failure during initialization.
|
||||
Host is being re-configured by Reboot.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- If auto-recovery is consistently unable to recover host to the
|
||||
unlocked-enabled state contact next level of support or lock and
|
||||
replace failing host.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.013**
|
||||
- <hostname> compute service of the only available controller is not
|
||||
operational. Auto-recovery is disabled. Degrading host instead.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- Major
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Enable second controller and Switch Activity \(Swact\) over to it as
|
||||
soon as possible. Then Lock and Unlock host to recover its local compute
|
||||
service.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.014**
|
||||
- The Hardware Monitor was unable to load, configure and monitor one
|
||||
or more hardware sensors.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- m
|
||||
* - Proposed Repair Action
|
||||
- Check Board Management Controller provisioning. Try reprovisioning the
|
||||
BMC. If problem persists try power cycling the host and then the entire
|
||||
server including the BMC power. If problem persists then contact next
|
||||
level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 200.015**
|
||||
- Unable to read one or more sensor groups from this host's board
|
||||
management controller.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M
|
||||
* - Proposed Repair Action
|
||||
- Check board management connectivity and try rebooting the board
|
||||
management controller. If problem persists contact next level of
|
||||
support or lock and replace failing host.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 210.001**
|
||||
- System Backup in progress.
|
||||
* - Entity Instance
|
||||
- host=controller
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- m\*
|
||||
* - Proposed Repair Action
|
||||
- No action required.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 250.001**
|
||||
- <hostname> Configuration is out-of-date.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Administratively lock and unlock <hostname> to update config.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 250.003**
|
||||
- Kubernetes certificates rotation failed on host <hostname>[, reason =
|
||||
<reason_text>].
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M/w
|
||||
* - Proposed Repair Action
|
||||
- Lock and unlock the host to update services with new certificates
|
||||
(Manually renew kubernetes certificates first if renewal failed).
|
||||
|
||||
-----
|
||||
|
||||
.. only:: partner
|
||||
|
||||
.. include:: /_includes/200-series-alarm-messages.rest
|
||||
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 270.001**
|
||||
- Host <host\_name> compute services failure\[, reason = <reason\_text>\]
|
||||
* - Entity Instance
|
||||
- host=<host\_name>.services=compute
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- Wait for host services recovery to complete; if problem persists contact
|
||||
next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 280.001**
|
||||
- <subcloud> is offline.
|
||||
* - Entity Instance
|
||||
- subcloud=<subcloud>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- Wait for subcloud to become online; if problem persists contact next
|
||||
level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 280.002**
|
||||
- <subcloud><resource> sync status is out-of-sync.
|
||||
* - Entity Instance
|
||||
- \[subcloud=<subcloud>.resource=<compute> \| <network> \| <platform>
|
||||
\| <volumev2>\]
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- If problem persists contact next level of support.
|
@ -1,120 +0,0 @@
|
||||
|
||||
.. lzz1579291773073
|
||||
.. _200-series-maintenance-customer-log-messages:
|
||||
|
||||
============================================
|
||||
200 Series Maintenance Customer Log Messages
|
||||
============================================
|
||||
|
||||
The Customer Logs include events that do not require immediate user action.
|
||||
|
||||
The following types of events are included in the Customer Logs. The severity
|
||||
of the events is represented in the table by one or more letters, as follows:
|
||||
|
||||
.. _200-series-maintenance-customer-log-messages-ul-jsd-jkg-vp:
|
||||
|
||||
- C: Critical
|
||||
|
||||
- M: Major
|
||||
|
||||
- m: Minor
|
||||
|
||||
- W: Warning
|
||||
|
||||
- NA: Not applicable
|
||||
|
||||
.. _200-series-maintenance-customer-log-messages-table-zgf-jvw-v5:
|
||||
|
||||
|
||||
.. table:: Table 1. Customer Log Messages
|
||||
:widths: auto
|
||||
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| Log ID | Description | Severity |
|
||||
+ +------------------------------------------------------------------+----------+
|
||||
| | Entity Instance ID | |
|
||||
+=================+==================================================================+==========+
|
||||
| 200.020 | <hostname> has been 'discovered' on the network | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.event=discovered | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.020 | <hostname> has been 'added' to the system | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.event=add | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.020 | <hostname> has 'entered' multi-node failure avoidance | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.event=mnfa\_enter | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.020 | <hostname> has 'exited' multi-node failure avoidance | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.event=mnfa\_exit | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> board management controller has been 'provisioned' | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=provision | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> board management controller has been 're-provisioned' | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=reprovision | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> board management controller has been 'de-provisioned' | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=deprovision | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> manual 'unlock' request | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=unlock | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> manual 'reboot' request | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=reboot | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> manual 'reset' request | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=reset | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> manual 'power-off' request | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=power-off | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> manual 'power-on' request | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=power-on | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> manual 'reinstall' request | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=reinstall | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> manual 'force-lock' request | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=force-lock | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> manual 'delete' request | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=delete | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.021 | <hostname> manual 'controller switchover' request | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.command=swact | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.022 | <hostname> is now 'disabled' | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.state=disabled | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.022 | <hostname> is now 'enabled' | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.state=enabled | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.022 | <hostname> is now 'online' | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.status=online | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.022 | <hostname> is now 'offline' | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.status=offline | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
||||
| 200.022 | <hostname> is 'disabled-failed' to the system | NA |
|
||||
| | | |
|
||||
| | host=<hostname>.status=failed | |
|
||||
+-----------------+------------------------------------------------------------------+----------+
|
@ -1,53 +0,0 @@
|
||||
|
||||
.. zwe1579701930425
|
||||
.. _300-series-alarm-messages:
|
||||
|
||||
=========================
|
||||
300 Series Alarm Messages
|
||||
=========================
|
||||
|
||||
The system inventory and maintenance service reports system changes with
|
||||
different degrees of severity. Use the reported alarms to monitor the
|
||||
overall health of the system.
|
||||
|
||||
.. include:: /_includes/x00-series-alarm-messages.rest
|
||||
|
||||
.. _300-series-alarm-messages-table-zrd-tg5-v5:
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 300.001**
|
||||
- 'Data' Port failed.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.port=<port-uuid>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 300.002**
|
||||
- 'Data' Interface degraded.
|
||||
|
||||
or
|
||||
|
||||
'Data' Interface failed.
|
||||
* - Entity Instance
|
||||
- host=<hostname>.interface=<if-uuid>
|
||||
* - Degrade Affecting Severity:
|
||||
- Critical
|
||||
* - Severity:
|
||||
- C/M\*
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
@ -1,69 +0,0 @@
|
||||
|
||||
.. ots1579702138430
|
||||
.. _400-series-alarm-messages:
|
||||
|
||||
=========================
|
||||
400 Series Alarm Messages
|
||||
=========================
|
||||
|
||||
The system inventory and maintenance service reports system changes with
|
||||
different degrees of severity. Use the reported alarms to monitor the overall
|
||||
health of the system.
|
||||
|
||||
.. include:: /_includes/x00-series-alarm-messages.rest
|
||||
|
||||
.. _400-series-alarm-messages-table-zrd-tg5-v5:
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 400.003**
|
||||
- License key is not installed; a valid license key is required for
|
||||
operation.
|
||||
|
||||
or
|
||||
|
||||
License key has expired or is invalid; a valid license key is required
|
||||
for operation.
|
||||
|
||||
or
|
||||
|
||||
Evaluation license key will expire on <date>; there are <num\_days> days
|
||||
remaining in this evaluation.
|
||||
|
||||
or
|
||||
|
||||
Evaluation license key will expire on <date>; there is only 1 day
|
||||
remaining in this evaluation.
|
||||
* - Entity Instance:
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- Contact next level of support to obtain a new license key.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 400.005**
|
||||
- Communication failure detected with peer over port <linux-ifname>.
|
||||
|
||||
or
|
||||
|
||||
Communication failure detected with peer over port <linux-ifname>
|
||||
within the last 30 seconds.
|
||||
* - Entity Instance:
|
||||
- host=<hostname>.network=<mgmt \| oam \| cluster-host>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Check cabling and far-end port configuration and status on adjacent
|
||||
equipment.
|
@ -1,81 +0,0 @@
|
||||
|
||||
.. pgb1579292662158
|
||||
.. _400-series-customer-log-messages:
|
||||
|
||||
================================
|
||||
400 Series Customer Log Messages
|
||||
================================
|
||||
|
||||
The Customer Logs include events that do not require immediate user action.
|
||||
|
||||
The following types of events are included in the Customer Logs. The severity
|
||||
of the events is represented in the table by one or more letters, as follows:
|
||||
|
||||
.. _400-series-customer-log-messages-ul-jsd-jkg-vp:
|
||||
|
||||
- C: Critical
|
||||
|
||||
- M: Major
|
||||
|
||||
- m: Minor
|
||||
|
||||
- W: Warning
|
||||
|
||||
- NA: Not applicable
|
||||
|
||||
.. _400-series-customer-log-messages-table-zgf-jvw-v5:
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 15
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 400.003**
|
||||
- License key has expired or is invalid
|
||||
|
||||
or
|
||||
|
||||
Evaluation license key will expire on <date>
|
||||
|
||||
or
|
||||
|
||||
License key is valid
|
||||
* - Entity Instance
|
||||
- host=<host\_name>
|
||||
* - Severity:
|
||||
- C
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 15
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 400.005**
|
||||
- Communication failure detected with peer over port <port> on host
|
||||
<host name>
|
||||
|
||||
or
|
||||
|
||||
Communication failure detected with peer over port <port> on host
|
||||
<host name> within the last <X> seconds
|
||||
|
||||
or
|
||||
|
||||
Communication established with peer over port <port> on host <host name>
|
||||
* - Entity Instance
|
||||
- host=<host\_name>.network=<network>
|
||||
* - Severity:
|
||||
- C
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 15
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 400.007**
|
||||
- Swact or swact-force
|
||||
* - Entity Instance
|
||||
- host=<host\_name>
|
||||
* - Severity:
|
||||
- C
|
@ -1,91 +0,0 @@
|
||||
|
||||
.. xpx1579702157578
|
||||
.. _500-series-alarm-messages:
|
||||
|
||||
=========================
|
||||
500 Series Alarm Messages
|
||||
=========================
|
||||
|
||||
The system inventory and maintenance service reports system changes with
|
||||
different degrees of severity. Use the reported alarms to monitor the overall
|
||||
health of the system.
|
||||
|
||||
.. include:: /_includes/x00-series-alarm-messages.rest
|
||||
|
||||
.. _500-series-alarm-messages-table-zrd-tg5-v5:
|
||||
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 500.101**
|
||||
- Developer patch certificate enabled.
|
||||
* - Entity Instance
|
||||
- host=controller
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C
|
||||
* - Proposed Repair Action
|
||||
- Reinstall system to disable developer certificate and remove untrusted
|
||||
patches.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 500.200**
|
||||
- Certificate ‘system certificate-show <uuid>' (mode=<ssl/ssl_ca/docker_registry/openstack/openstack_ca>) expiring soon on <date>.
|
||||
OR
|
||||
Certificate ‘<Namespace>/<Certificate/Secret>’ expiring soon on <date>.
|
||||
OR
|
||||
Certificate ‘<k8sRootCA/EtcdCA>’ expiring soon on <date>.
|
||||
system.certificate.k8sRootCA
|
||||
* - Entity Instance
|
||||
- system.certificate.mode=<mode>.uuid=<uuid>
|
||||
OR
|
||||
namespace=<namespace-name>.certificate=<certificate-name>
|
||||
OR
|
||||
namespace=<namespace-name>.secret=<secret-name>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M
|
||||
* - Proposed Repair Action
|
||||
- Renew certificate for the entity identified.
|
||||
* - Management_Affecting_Severity:
|
||||
- none
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 500.210**
|
||||
- Certificate ‘system certificate-show <uuid>' (mode=<ssl/ssl_ca/docker_registry/openstack/openstack_ca>) expired.
|
||||
OR
|
||||
Certificate ‘<Namespace>/<Certificate/Secret>’ expired.
|
||||
OR
|
||||
Certificate ‘<k8sRootCA/EtcdRootCA>’ expired.
|
||||
* - Entity Instance
|
||||
- system.certificate.mode=<mode>.uuid=<uuid>
|
||||
OR
|
||||
namespace=<namespace-name>.certificate=<certificate-name>
|
||||
OR
|
||||
namespace=<namespace-name>.secret=<secret-name>
|
||||
OR
|
||||
system.certificate.k8sRootCA
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C
|
||||
* - Proposed Repair Action
|
||||
- Renew certificate for the entity identified.
|
||||
* - Management_Affecting_Severity:
|
||||
- none
|
@ -1,118 +0,0 @@
|
||||
|
||||
.. cta1579702173704
|
||||
.. _750-series-alarm-messages:
|
||||
|
||||
=========================
|
||||
750 Series Alarm Messages
|
||||
=========================
|
||||
|
||||
The system inventory and maintenance service reports system changes with
|
||||
different degrees of severity. Use the reported alarms to monitor the overall
|
||||
health of the system.
|
||||
|
||||
.. include:: /_includes/x00-series-alarm-messages.rest
|
||||
|
||||
.. _750-series-alarm-messages-table-zrd-tg5-v5:
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 750.001**
|
||||
- Application upload failure.
|
||||
* - Entity Instance
|
||||
- k8s\_application=<appname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- W
|
||||
* - Proposed Repair Action
|
||||
- Check the system inventory log for the cause.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 750.002**
|
||||
- Application apply failure.
|
||||
* - Entity Instance
|
||||
- k8s\_application=<appname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M
|
||||
* - Proposed Repair Action
|
||||
- Retry applying the application. If the issue persists, please check the
|
||||
system inventory log for cause.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 750.003**
|
||||
- Application remove failure.
|
||||
* - Entity Instance
|
||||
- k8s\_application=<appname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M
|
||||
* - Proposed Repair Action
|
||||
- Retry removing the application. If the issue persists, please the check
|
||||
system inventory log for cause.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 750.004**
|
||||
- Application apply in progress.
|
||||
* - Entity Instance
|
||||
- k8s\_application=<appname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- W
|
||||
* - Proposed Repair Action
|
||||
- No action is required.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 750.005**
|
||||
- Application update in progress.
|
||||
* - Entity Instance
|
||||
- k8s\_application=<appname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- W
|
||||
* - Proposed Repair Action
|
||||
- No action is required.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 750.006**
|
||||
- Automatic application re-apply is pending.
|
||||
* - Entity Instance
|
||||
- k8s\_application=<appname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- W
|
||||
* - Proposed Repair Action
|
||||
- Ensure all hosts are either locked or unlocked. When the system is
|
||||
stable the application will automatically be reapplied.
|
@ -1,152 +0,0 @@
|
||||
|
||||
.. rww1579702317136
|
||||
.. _800-series-alarm-messages:
|
||||
|
||||
=========================
|
||||
800 Series Alarm Messages
|
||||
=========================
|
||||
|
||||
The system inventory and maintenance service reports system changes with
|
||||
different degrees of severity. Use the reported alarms to monitor the overall
|
||||
health of the system.
|
||||
|
||||
.. include:: /_includes/x00-series-alarm-messages.rest
|
||||
|
||||
.. _800-series-alarm-messages-table-zrd-tg5-v5:
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 800.001**
|
||||
- Storage Alarm Condition:
|
||||
|
||||
1 mons down, quorum 1,2 controller-1,storage-0
|
||||
* - Entity Instance
|
||||
- cluster=<dist-fs-uuid>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C/M\*
|
||||
* - Proposed Repair Action
|
||||
- If problem persists, contact next level of support.
|
||||
|
||||
.. -----
|
||||
..
|
||||
.. .. list-table::
|
||||
.. :widths: 6 25
|
||||
.. :header-rows: 0
|
||||
..
|
||||
.. * - **Alarm ID: 800.003**
|
||||
.. - Storage Alarm Condition: Quota/Space mismatch for the <tiername> tier.
|
||||
.. The sum of Ceph pool quotas does not match the tier size.
|
||||
.. * - Entity Instance
|
||||
.. - cluster=<dist-fs-uuid>.tier=<tiername>
|
||||
.. * - Degrade Affecting Severity:
|
||||
.. - None
|
||||
.. * - Severity:
|
||||
.. - m
|
||||
.. * - Proposed Repair Action
|
||||
.. - Update ceph storage pool quotas to use all available tier space.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 800.010**
|
||||
- Potential data loss. No available OSDs in storage replication group.
|
||||
* - Entity Instance
|
||||
- cluster=<dist-fs-uuid>.peergroup=<group-x>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- Ensure storage hosts from replication group are unlocked and available.
|
||||
Check if OSDs of each storage host are up and running. If problem
|
||||
persists contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 800.011**
|
||||
- Loss of replication in peergroup.
|
||||
* - Entity Instance
|
||||
- cluster=<dist-fs-uuid>.peergroup=<group-x>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Ensure storage hosts from replication group are unlocked and available.
|
||||
Check if OSDs of each storage host are up and running. If problem
|
||||
persists contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 800.102**
|
||||
- Storage Alarm Condition:
|
||||
|
||||
PV configuration <error/failed to apply\> on <hostname>.
|
||||
Reason: <detailed reason\>.
|
||||
* - Entity Instance
|
||||
- pv=<pv\_uuid>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C/M\*
|
||||
* - Proposed Repair Action
|
||||
- Remove failed PV and associated Storage Device then recreate them.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 800.103**
|
||||
- Storage Alarm Condition:
|
||||
|
||||
\[ Metadata usage for LVM thin pool <VG name>/<Pool name> exceeded
|
||||
threshold and automatic extension failed.
|
||||
|
||||
Metadata usage for LVM thin pool <VG name>/<Pool name> exceeded
|
||||
threshold \]; threshold x%, actual y%.
|
||||
* - Entity Instance
|
||||
- <hostname>.lvmthinpool=<VG name>/<Pool name>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- Increase Storage Space Allotment for Cinder on the 'lvm' backend.
|
||||
Consult the user documentation for more details. If problem persists,
|
||||
contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 800.104**
|
||||
- Storage Alarm Condition:
|
||||
|
||||
<storage-backend-name> configuration failed to apply on host: <host-uuid>.
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- Update backend setting to reapply configuration. Consult the user
|
||||
documentation for more details. If problem persists, contact next level
|
||||
of support.
|
@ -1,356 +0,0 @@
|
||||
|
||||
.. pti1579702342696
|
||||
.. _900-series-alarm-messages:
|
||||
|
||||
=========================
|
||||
900 Series Alarm Messages
|
||||
=========================
|
||||
|
||||
The system inventory and maintenance service reports system changes with
|
||||
different degrees of severity. Use the reported alarms to monitor the overall
|
||||
health of the system.
|
||||
|
||||
.. include:: /_includes/x00-series-alarm-messages.rest
|
||||
|
||||
.. _900-series-alarm-messages-table-zrd-tg5-v5:
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.001**
|
||||
- Patching operation in progress.
|
||||
* - Entity Instance
|
||||
- host=controller
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- m\*
|
||||
* - Proposed Repair Action
|
||||
- Complete reboots of affected hosts.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.002**
|
||||
- Patch host install failure.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Undo patching operation.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.003**
|
||||
- Obsolete patch in system.
|
||||
* - Entity Instance
|
||||
- host=controller
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- W\*
|
||||
* - Proposed Repair Action
|
||||
- Remove and delete obsolete patches.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.004**
|
||||
- Host version mismatch.
|
||||
* - Entity Instance
|
||||
- host=<hostname>
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Reinstall host to update applied load.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.005**
|
||||
- System Upgrade in progress.
|
||||
* - Entity Instance
|
||||
- host=controller
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- m\*
|
||||
* - Proposed Repair Action
|
||||
- No action required.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.008**
|
||||
- Kubernetes rootca update in progress.
|
||||
* - Entity Instance
|
||||
- host=controller
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- m
|
||||
* - Proposed Repair Action
|
||||
- Wait for kubernetes rootca procedure to complete.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.009**
|
||||
- Kubernetes root CA update aborted, certificates may not be fully updated.
|
||||
* - Entity Instance
|
||||
- host=controller
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- m
|
||||
* - Management Affecting Severity:
|
||||
- w
|
||||
* - Proposed Repair Action
|
||||
- Fully update certificates by a new root CA update.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.101**
|
||||
- Software patch auto-apply in progress.
|
||||
* - Entity Instance
|
||||
- orchestration=sw-patch
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Wait for software patch auto-apply to complete; if problem persists
|
||||
contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.102**
|
||||
- Software patch auto-apply aborting.
|
||||
* - Entity Instance
|
||||
- orchestration=sw-patch
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Wait for software patch auto-apply abort to complete; if problem
|
||||
persists contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.103**
|
||||
- Software patch auto-apply failed.
|
||||
* - Entity Instance
|
||||
- orchestration=sw-patch
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Attempt to apply software patches manually; if problem persists contact
|
||||
next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.201**
|
||||
- Software upgrade auto-apply in progress.
|
||||
* - Entity Instance
|
||||
- orchestration=sw-upgrade
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Wait for software upgrade auto-apply to complete; if problem persists
|
||||
contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.202**
|
||||
- Software upgrade auto-apply aborting.
|
||||
* - Entity Instance
|
||||
- orchestration=sw-upgrade
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Wait for software upgrade auto-apply abort to complete; if problem
|
||||
persists contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.203**
|
||||
- Software update auto-apply failed.
|
||||
* - Entity Instance
|
||||
- orchestration=sw-upgrade
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Attempt to apply software upgrade manually; if problem persists contact
|
||||
next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.301**
|
||||
- Firmware Update auto-apply in progress.
|
||||
* - Entity Instance
|
||||
- orchestration=fw-upgrade
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Wait for firmware update auto-apply to complete; if problem persists
|
||||
contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.302**
|
||||
- Firmware Update auto-apply aborting.
|
||||
* - Entity Instance
|
||||
- orchestration=fw-upgrade
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Wait for firmware update auto-apply abort to complete; if problem
|
||||
persists contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.303**
|
||||
- Firmware Update auto-apply failed.
|
||||
* - Entity Instance
|
||||
- orchestration=fw-upgrade
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Attempt to apply firmware update manually; if problem persists
|
||||
contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.501**
|
||||
- Kubernetes rootca update auto-apply in progress
|
||||
* - Entity Instance
|
||||
- orchestration=kube-rootca-update
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M
|
||||
* - Management Affecting Severity:
|
||||
- w
|
||||
* - Proposed Repair Action
|
||||
- Wait for kubernetes rootca update auto-apply to complete; if problem
|
||||
persists contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.502**
|
||||
- Kubernetes rootca update auto-apply aborting.
|
||||
* - Entity Instance
|
||||
- orchestration=kube-rootca-update
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- M
|
||||
* - Management Affecting Severity:
|
||||
- w
|
||||
* - Proposed Repair Action
|
||||
- Wait for kubernetes rootca update auto-apply abort to complete; if
|
||||
problem persists contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: 900.503**
|
||||
- Kubernetes rootca update auto-apply failed.
|
||||
* - Entity Instance
|
||||
- orchestration=kube-rootca-update
|
||||
* - Degrade Affecting Severity:
|
||||
- None
|
||||
* - Severity:
|
||||
- C
|
||||
* - Management Affecting Severity:
|
||||
- w
|
||||
* - Proposed Repair Action
|
||||
- Attempt to apply kubernetes rootca update manually; if problem persists
|
||||
contact next level of support.
|
@ -1,211 +0,0 @@
|
||||
|
||||
.. bdq1579700719122
|
||||
.. _900-series-orchestration-customer-log-messages:
|
||||
|
||||
==============================================
|
||||
900 Series Orchestration Customer Log Messages
|
||||
==============================================
|
||||
|
||||
The Customer Logs include events that do not require immediate user action.
|
||||
|
||||
The following types of events are included in the Customer Logs. The severity
|
||||
of the events is represented in the table by one or more letters, as follows:
|
||||
|
||||
.. _900-series-orchestration-customer-log-messages-ul-jsd-jkg-vp:
|
||||
|
||||
- C: Critical
|
||||
|
||||
- M: Major
|
||||
|
||||
- m: Minor
|
||||
|
||||
- W: Warning
|
||||
|
||||
- NA: Not applicable
|
||||
|
||||
.. _900-series-orchestration-customer-log-messages-table-zgf-jvw-v5:
|
||||
|
||||
.. table:: Table 1. Customer Log Messages
|
||||
:widths: auto
|
||||
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| Log ID | Description | Severity |
|
||||
+ +-------------------------------------------------------+----------+
|
||||
| | Entity Instance ID | |
|
||||
+===================+============================================+==========+==========+
|
||||
| 900.111 | Software update auto-apply start | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.112 | Software update auto-apply in progress | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.113 | Software update auto-apply rejected | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.114 | Software update auto-apply canceled | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.115 | Software update auto-apply failed | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.116 | Software update auto-apply completed | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.117 | Software update auto-apply abort | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.118 | Software update auto-apply aborting | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.119 | Software update auto-apply abort rejected | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.120 | Software update auto-apply abort failed | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.121 | Software update auto-apply aborted | C |
|
||||
| | | |
|
||||
| | orchestration=sw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.211 | Software upgrade auto-apply start | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.212 | Software upgrade auto-apply in progress | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.213 | Software upgrade auto-apply rejected | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.214 | Software upgrade auto-apply canceled | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.215 | Software upgrade auto-apply failed | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.216 | Software upgrade auto-apply completed | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.217 | Software upgrade auto-apply abort | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.218 | Software upgrade auto-apply aborting | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.219 | Software upgrade auto-apply abort rejected | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.220 | Software upgrade auto-apply abort failed | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.221 | Software upgrade auto-apply aborted | C |
|
||||
| | | |
|
||||
| | orchestration=sw-upgrade | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.311 | Firmware update auto-apply | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.312 | Firmware update auto-apply in progress | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.313 | Firmware update auto-apply rejected | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.314 | Firmware update auto-apply canceled | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.315 | Firmware update auto-apply failed | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.316 | Firmware update auto-apply completed | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.317 | Firmware update auto-apply aborted | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.318 | Firmware update auto-apply aborting | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.319 | Firmware update auto-apply abort rejected | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.320 | Firmware update auto-apply abort failed | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.321 | Firmware update auto-apply aborted | C |
|
||||
| | | |
|
||||
| | orchestration=fw-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.511 | Kubernetes rootca update auto-apply start | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.512 | Kubernetes rootca update auto-apply in progress | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.513 | Firmware Kubernetes rootca update auto-apply rejected | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.514 | Kubernetes rootca update auto-apply cancelled | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.515 | Kubernetes rootca update auto-apply failed | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.516 | Kubernetes rootca update auto-apply completed | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.517 | Kubernetes rootca update auto-apply abort | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.518 | Kubernetes rootca update auto-apply aborting | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.519 | Kubernetes rootca update auto-apply abort rejected | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.520 | Kubernetes rootca update auto-apply abort failed | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
||||
| 900.521 | Kubernetes rootca update auto-apply aborted | C |
|
||||
| | | |
|
||||
| | orchestration=kube-rootca-update | |
|
||||
+-------------------+-------------------------------------------------------+----------+
|
@ -0,0 +1,7 @@
|
||||
.. _alarm-messages-overview-19c242d3d151:
|
||||
|
||||
=======================
|
||||
Alarm Messages Overview
|
||||
=======================
|
||||
|
||||
.. include:: /shared/_includes/alarm-messages-overview-9d087b0170cf.rest
|
@ -9,19 +9,17 @@ Fault Management Overview
|
||||
An admin user can view |prod-long| fault management alarms and logs in order
|
||||
to monitor and respond to fault conditions.
|
||||
|
||||
See :ref:`Alarm Messages <100-series-alarm-messages>` for the list of
|
||||
See :ref:`Alarm Messages <100-series-alarm-messages-starlingx>` for the list of
|
||||
alarms and :ref:`Customer Log Messages
|
||||
<200-series-maintenance-customer-log-messages>`
|
||||
<200-series-log-messages-starlingx>`
|
||||
for the list of customer logs reported by |prod|.
|
||||
|
||||
You can access active and historical alarms, and customer logs using the CLI,
|
||||
GUI, REST APIs and |SNMP|.
|
||||
|
||||
To use the CLI, see
|
||||
:ref:`Viewing Active Alarms Using the CLI
|
||||
<viewing-active-alarms-using-the-cli>`
|
||||
and :ref:`Viewing the Event Log Using the CLI
|
||||
<viewing-the-event-log-using-the-cli>`.
|
||||
To use the CLI, see :ref:`Viewing Active Alarms Using the CLI
|
||||
<viewing-active-alarms-using-the-cli>` and :ref:`Viewing the Event Log Using the
|
||||
CLI <viewing-the-event-log-using-the-cli>`.
|
||||
|
||||
Using the GUI, you can obtain fault management information in a number of
|
||||
places.
|
||||
@ -38,11 +36,9 @@ places.
|
||||
|
||||
- **Admin** \> **Fault Management** \> **Active Alarms**—Alarms that are
|
||||
currently set, and require user action to clear them. For more
|
||||
information about active alarms, see
|
||||
:ref:`Viewing Active Alarms Using the CLI
|
||||
<viewing-active-alarms-using-the-cli>`
|
||||
and :ref:`Deleting an Alarm Using the CLI
|
||||
<deleting-an-alarm-using-the-cli>`.
|
||||
information about active alarms, see :ref:`Viewing Active Alarms Using
|
||||
the CLI <viewing-active-alarms-using-the-cli>` and :ref:`Deleting an
|
||||
Alarm Using the CLI <deleting-an-alarm-using-the-cli>`.
|
||||
|
||||
- **Admin** \> **Fault Management** \> **Events**—The event log
|
||||
consolidates historical alarms that have occurred in the past, that
|
||||
@ -50,8 +46,7 @@ places.
|
||||
logs.
|
||||
|
||||
For more about the event log, which includes historical alarms and
|
||||
customer logs, see
|
||||
:ref:`Viewing the Event Log Using Horizon
|
||||
customer logs, see :ref:`Viewing the Event Log Using Horizon
|
||||
<viewing-the-event-log-using-horizon>`.
|
||||
|
||||
- **Admin** \> **Fault Management** \> **Events Suppression**—Individual
|
||||
|
@ -108,14 +108,13 @@ Alarm messages
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
|
||||
100-series-alarm-messages
|
||||
200-series-alarm-messages
|
||||
300-series-alarm-messages
|
||||
400-series-alarm-messages
|
||||
500-series-alarm-messages
|
||||
750-series-alarm-messages
|
||||
800-series-alarm-messages
|
||||
900-series-alarm-messages
|
||||
alarm-messages-overview-19c242d3d151
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
:glob:
|
||||
|
||||
*-series-alarm-messages
|
||||
|
||||
************
|
||||
Log messages
|
||||
@ -124,6 +123,10 @@ Log messages
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
|
||||
200-series-maintenance-customer-log-messages
|
||||
400-series-customer-log-messages
|
||||
900-series-orchestration-customer-log-messages
|
||||
log-entries-overview-597c2c453680
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
:glob:
|
||||
|
||||
*-series-log-messages
|
||||
|
@ -0,0 +1,9 @@
|
||||
.. _log-entries-overview-597c2c453680:
|
||||
|
||||
====================
|
||||
Log Entries Overview
|
||||
====================
|
||||
|
||||
.. include:: /shared/_includes/log-entries-overview-6728006a298f.rest
|
||||
|
||||
|
@ -0,0 +1,7 @@
|
||||
.. _alarm-messages-overview-a250f214f65a:
|
||||
|
||||
=======================
|
||||
Alarm Messages Overview
|
||||
=======================
|
||||
|
||||
.. include:: /shared/_includes/alarm-messages-overview-9d087b0170cf.rest
|
@ -21,10 +21,14 @@ OpenStack alarm messages
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
|
||||
openstack-alarm-messages-300s
|
||||
openstack-alarm-messages-400s
|
||||
openstack-alarm-messages-700s
|
||||
openstack-alarm-messages-800s
|
||||
alarm-messages-overview-a250f214f65a
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
:glob:
|
||||
|
||||
*-series-alarm-messages
|
||||
|
||||
|
||||
*******************************
|
||||
OpenStack customer log messages
|
||||
@ -33,6 +37,10 @@ OpenStack customer log messages
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
|
||||
openstack-customer-log-messages-270s-virtual-machines
|
||||
openstack-customer-log-messages-401s-services
|
||||
openstack-customer-log-messages-700s-virtual-machines
|
||||
log-entries-overview-4f1dde7286c2
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
:glob:
|
||||
|
||||
*-series-log-messages
|
||||
|
@ -0,0 +1,8 @@
|
||||
.. _log-entries-overview-4f1dde7286c2:
|
||||
|
||||
====================
|
||||
Log Entries Overview
|
||||
====================
|
||||
|
||||
.. include:: /shared/_includes/log-entries-overview-6728006a298f.rest
|
||||
|
@ -1,110 +0,0 @@
|
||||
|
||||
.. slf1579788051430
|
||||
.. _alarm-messages-300s:
|
||||
|
||||
=====================
|
||||
Alarm Messages - 300s
|
||||
=====================
|
||||
|
||||
The system inventory and maintenance service reports system changes with
|
||||
different degrees of severity. Use the reported alarms to monitor the overall
|
||||
health of the system.
|
||||
|
||||
For more information, see :ref:`Overview
|
||||
<openstack-fault-management-overview>`.
|
||||
|
||||
In the following tables, the severity of the alarms is represented by one or
|
||||
more letters, as follows:
|
||||
|
||||
.. _alarm-messages-300s-ul-jsd-jkg-vp:
|
||||
|
||||
- C: Critical
|
||||
|
||||
- M: Major
|
||||
|
||||
- m: Minor
|
||||
|
||||
- W: Warning
|
||||
|
||||
A slash-separated list of letters is used when the alarm can be triggered with
|
||||
one of several severity levels.
|
||||
|
||||
An asterisk \(\*\) indicates the management-affecting severity, if any. A
|
||||
management-affecting alarm is one that cannot be ignored at the indicated
|
||||
severity level or higher by using relaxed alarm rules during an orchestrated
|
||||
patch or upgrade operation.
|
||||
|
||||
Differences exist between the terminology emitted by some alarms and that used
|
||||
in the |CLI|, GUI, and elsewhere in the documentation:
|
||||
|
||||
.. _alarm-messages-300s-ul-dsf-dxn-bhb:
|
||||
|
||||
- References to provider networks in alarms refer to data networks.
|
||||
|
||||
- References to data networks in alarms refer to physical networks.
|
||||
|
||||
- References to tenant networks in alarms refer to project networks.
|
||||
|
||||
|
||||
.. _alarm-messages-300s-table-zrd-tg5-v5:
|
||||
|
||||
.. table:: Table 1. Alarm Messages
|
||||
:widths: auto
|
||||
|
||||
+----------+-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| Alarm ID | Description | Severity | Proposed Repair Action |
|
||||
+ +-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| | Entity Instance ID |
|
||||
+==========+=====================================================================================+==========+===================================================================================================+
|
||||
| 300.003 | Networking Agent not responding. | M\* | If condition persists, attempt to clear issue by administratively locking and unlocking the Host. |
|
||||
+ +-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| | host=<hostname>.agent=<agent-uuid> |
|
||||
+----------+-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| 300.004 | No enabled compute host with connectivity to provider network. | M\* | Enable compute hosts with required provider network connectivity. |
|
||||
+ +-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| | service=networking.providernet=<pnet-uuid> |
|
||||
+----------+-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| 300.005 | Communication failure detected over provider network x% for ranges y% on host z%. | M\* | Check neighbour switch port VLAN assignments. |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Communication failure detected over provider network x% on host z%. | | |
|
||||
+ +-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| | host=<hostname>.service=networking.providernet=<pnet-uuid> |
|
||||
+----------+-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| 300.010 | ML2 Driver Agent non-reachable | M\* | Monitor and if condition persists, contact next level of support. |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | ML2 Driver Agent reachable but non-responsive | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | ML2 Driver Agent authentication failure | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | ML2 Driver Agent is unable to sync Neutron database | | |
|
||||
+ +-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| | host=<hostname>.ml2driver=<driver> |
|
||||
+----------+-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| 300.012 | Openflow Controller connection failed. | M\* | Check cabling and far-end port configuration and status on adjacent equipment. |
|
||||
+ +-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| | host=<hostname>.openflow-controller=<uri> |
|
||||
+----------+-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| 300.013 | No active Openflow controller connections found for this network. | C, M\* | Check cabling and far-end port configuration and status on adjacent equipment. |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | One or more Openflow controller connections in disconnected state for this network. | | |
|
||||
+ +-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| | host=<hostname>.openflow-network=<name> |
|
||||
+----------+-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| 300.015 | No active OVSDB connections found. | C\* | Check cabling and far-end port configuration and status on adjacent equipment. |
|
||||
+ +-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| | host=<hostname> |
|
||||
+----------+-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| 300.016 | Dynamic routing agent x% lost connectivity to peer y% | M\* | If condition persists, fix connectivity to peer. |
|
||||
+ +-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
||||
| | host=<hostname>,agent=<agent-uuid>,bgp-peer=<bgp-peer> |
|
||||
+----------+-------------------------------------------------------------------------------------+----------+---------------------------------------------------------------------------------------------------+
|
@ -1,47 +0,0 @@
|
||||
|
||||
.. msm1579788069384
|
||||
.. _alarm-messages-400s:
|
||||
|
||||
=====================
|
||||
Alarm Messages - 400s
|
||||
=====================
|
||||
|
||||
.. include:: /_includes/openstack-alarm-messages-xxxs.rest
|
||||
|
||||
.. _alarm-messages-400s-table-zrd-tg5-v5:
|
||||
|
||||
.. table:: Table 1. Alarm Messages
|
||||
:widths: auto
|
||||
|
||||
+----------+------------------------------------------------------------------------------------------------------------------+----------+----------------------------------------------------------------------------------------------------+
|
||||
| Alarm ID | Description | Severity | Proposed Repair Action |
|
||||
+ +------------------------------------------------------------------------------------------------------------------+----------+----------------------------------------------------------------------------------------------------+
|
||||
| | Entity Instance ID |
|
||||
+==========+==================================================================================================================+==========+====================================================================================================+
|
||||
| 400.001 | Service group failure; <list of affected services>. | C/M/m\* | Contact next level of support. |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Service group degraded; <list of affected services>. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Service group warning; <list of affected services>. | | |
|
||||
| | | | |
|
||||
+ +------------------------------------------------------------------------------------------------------------------+----------+----------------------------------------------------------------------------------------------------+
|
||||
| | service_domain=<domain_name>.service_group=<group_name>.host=<hostname> |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------+----------+----------------------------------------------------------------------------------------------------+
|
||||
| 400.002 | Service group loss of redundancy; expected <num> standby member<s> but only <num> standby member<s> available. | M\* | Bring a controller node back in to service, otherwise contact next level of support. |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Service group loss of redundancy; expected <num> active member<s> but no active members available. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Service group loss of redundancy; expected <num> active member<s> but only <num> active member<s> available. | | |
|
||||
| | | | |
|
||||
+ +------------------------------------------------------------------------------------------------------------------+----------+----------------------------------------------------------------------------------------------------+
|
||||
| | service_domain=<domain_name>.service_group=<group_name> |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------+----------+----------------------------------------------------------------------------------------------------+
|
||||
|
@ -1,87 +0,0 @@
|
||||
|
||||
.. uxo1579788086872
|
||||
.. _alarm-messages-700s:
|
||||
|
||||
=====================
|
||||
Alarm Messages - 700s
|
||||
=====================
|
||||
|
||||
.. include:: /_includes/openstack-alarm-messages-xxxs.rest
|
||||
|
||||
.. _alarm-messages-700s-table-zrd-tg5-v5:
|
||||
|
||||
.. table:: Table 1. Alarm Messages
|
||||
:widths: auto
|
||||
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| Alarm ID | Description | Severity | Proposed Repair Action |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | Entity Instance ID |
|
||||
+==========+=====================================================================================================================+==========+===========================================================================================================+
|
||||
| 700.001 | Instance <instance_name> owned by <tenant_name> has failed on host <host_name>. | C\* | The system will attempt recovery; no repair action required. |
|
||||
| | | | |
|
||||
| | Instance <instance_name> owned by <tenant_name> has failed to schedule. | | |
|
||||
| | | | |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.002 | Instance <instance_name> owned by <tenant_name> is paused on host <host_name>. | C\* | Unpause the instance. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.003 | Instance <instance_name> owned by <tenant_name> is suspended on host <host_name>. | C\* | Resume the instance. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.004 | Instance <instance_name> owned by <tenant_name> is stopped on host <host_name>. | C\* | Start the instance. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.005 | Instance <instance_name> owned by <tenant_name> is rebooting on host <host_name>. | C\* | Wait for reboot to complete; if problem persists contact next level of support. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.006 | Instance <instance_name> owned by <tenant_name> is rebuilding on host <host_name>. | C\* | Wait for rebuild to complete; if problem persists contact next level of support. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.007 | Instance <instance_name> owned by <tenant_name> is evacuating from host <host_name>. | C\* | Wait for evacuate to complete; if problem persists contact next level of support. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.008 | Instance <instance_name> owned by <tenant_name> is live migrating from host <host_name>. | W\* | Wait for live migration to complete; if problem persists contact next level of support. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.009 | Instance <instance_name> owned by <tenant_name> is cold migrating from host <host_name>. | C\* | Wait for cold migration to complete; if problem persists contact next level of support. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.010 | Instance <instance_name> owned by <tenant_name> has been cold-migrated to host <host_name> waiting for confirmation.| C\* | Confirm or revert cold-migrate of instance. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.011 | Instance <instance_name> owned by <tenant_name> is reverting cold migrate to host <host_name>. | C\* | Wait for cold migration revert to complete; if problem persists contact next level of support. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.012 | Instance <instance_name> owned by <tenant_name> is resizing on host <host_name>. | C\* | Wait for resize to complete; if problem persists contact next level of support. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.013 | Instance <instance_name> owned by <tenant_name> has been resized on host <host_name> waiting for confirmation. | C\* | Confirm or revert resize of instance. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.014 | Instance <instance_name> owned by <tenant_name> is reverting resize on host <host_name>. | C\* | Wait for resize revert to complete; if problem persists contact next level of support. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.016 | Multi-Node Recovery Mode. | m\* |Wait for the system to exit out of this mode. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | subsystem=vim |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 700.017 | Server group <server_group_name> <policy> policy was not satisfied. | m\* | Migrate instances in an attempt to satisfy the policy; if problem persists contact next level of support. |
|
||||
+ +---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | server-group<server-group-uuid> |
|
||||
+----------+---------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
@ -1,81 +0,0 @@
|
||||
|
||||
.. tsh1579788106505
|
||||
.. _alarm-messages-800s:
|
||||
|
||||
=====================
|
||||
Alarm Messages - 800s
|
||||
=====================
|
||||
|
||||
.. include:: /_includes/openstack-alarm-messages-xxxs.rest
|
||||
|
||||
.. _alarm-messages-800s-table-zrd-tg5-v5:
|
||||
|
||||
.. table:: Table 1. Alarm Messages
|
||||
:widths: auto
|
||||
|
||||
+----------+-------------------------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| Alarm ID | Description | Severity | Proposed Repair Action |
|
||||
+ +-------------------------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | Entity Instance ID |
|
||||
+==========+=====================================================================================================================================+==========+===========================================================================================================+
|
||||
| 800.002 | Image storage media is full: There is not enough disk space on the image storage media. | W\* | If problem persists, contact next level of support. |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Instance <instance name> snapshot failed: There is not enough disk space on the image storage media. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Supplied <attrs> (<supplied>) and <attrs> generated from uploaded image (<actual>) did not match. Setting image status to 'killed'. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Error in store configuration. Adding images to store is disabled. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Forbidden upload attempt: <exception>. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Insufficient permissions on image storage media: <exception>. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Denying attempt to upload image larger than <size> bytes. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Denying attempt to upload image because it exceeds the quota: <exception>. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Received HTTP error while uploading image <image_id>. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Client disconnected before sending all data to backend. | | |
|
||||
| | | | |
|
||||
| | or | | |
|
||||
| | | | |
|
||||
| | Failed to upload image <image_id>. | | |
|
||||
| | | | |
|
||||
+ +-------------------------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | image=<image-uuid>, instance=<instance-uuid> |
|
||||
| | |
|
||||
| | or |
|
||||
| | |
|
||||
| | tenant=<tenant-uuid>, instance=<instance-uuid> |
|
||||
+----------+-------------------------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 800.100 | Storage Alarm Condition: | M\* | Reduce the I/O load on the Cinder LVM backend. Use Cinder QoS mechanisms on high usage volumes. |
|
||||
| | | | |
|
||||
| | Cinder I/O Congestion is above normal range and is building. | | |
|
||||
+ +-------------------------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | cinder_io_monitor |
|
||||
+----------+-------------------------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| 800.101 | Storage Alarm Condition: | C\* | Reduce the I/O load on the Cinder LVM backend. Cinder actions may fail until congestion is reduced. |
|
||||
| | | | Use Cinder QoS mechanisms on high usage volumes. |
|
||||
| | Cinder I/O Congestion is high and impacting guest performance. | | |
|
||||
+ +-------------------------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
||||
| | cinder_io_monitor |
|
||||
+----------+-------------------------------------------------------------------------------------------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------+
|
@ -1,38 +0,0 @@
|
||||
|
||||
.. ftb1579789103703
|
||||
.. _customer-log-messages-270s-virtual-machines:
|
||||
|
||||
=============================================
|
||||
Customer Log Messages 270s - Virtual Machines
|
||||
=============================================
|
||||
|
||||
.. include:: /_includes/openstack-customer-log-messages-xxxs.rest
|
||||
|
||||
.. _customer-log-messages-270s-virtual-machines-table-zgf-jvw-v5:
|
||||
|
||||
.. table:: Table 1. Customer Log Messages - Virtual Machines
|
||||
:widths: auto
|
||||
|
||||
+-----------+----------------------------------------------------------------------------------+----------+
|
||||
| Log ID | Description | Severity |
|
||||
+ +----------------------------------------------------------------------------------+----------+
|
||||
| | Entity Instance ID | |
|
||||
+===========+==================================================================================+==========+
|
||||
| 270.101 | Host <host\_name> compute services failure\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+-----------+----------------------------------------------------------------------------------+----------+
|
||||
| 270.102 | Host <host\_name> compute services enabled | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+-----------+----------------------------------------------------------------------------------+----------+
|
||||
| 270.103 | Host <host\_name> compute services disabled | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+-----------+----------------------------------------------------------------------------------+----------+
|
||||
| 275.001 | Host <host\_name> hypervisor is now <administrative\_state>-<operational\_state> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+-----------+----------------------------------------------------------------------------------+----------+
|
||||
|
||||
See also :ref:`Customer Log Messages 700s - Virtual Machines <customer-log-messages-700s-virtual-machines>`
|
@ -1,40 +0,0 @@
|
||||
|
||||
.. hwr1579789203684
|
||||
.. _customer-log-messages-401s-services:
|
||||
|
||||
=====================================
|
||||
Customer Log Messages 401s - Services
|
||||
=====================================
|
||||
|
||||
.. include:: /_includes/openstack-customer-log-messages-xxxs.rest
|
||||
|
||||
.. _customer-log-messages-401s-services-table-zgf-jvw-v5:
|
||||
|
||||
.. table:: Table 1. Customer Log Messages - Virtual Machines
|
||||
:widths: auto
|
||||
|
||||
+-----------+--------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| Log ID | Description | Severity |
|
||||
+ +--------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| | Entity Instance ID | |
|
||||
+===========+====================================================================================================================+==========+
|
||||
| 401.001 | Service group <group> state change from <state> to <state> on host <host_name> | C |
|
||||
| | | |
|
||||
| | service_domain=<domain>.service_group=<group>.host=<host_name> | |
|
||||
+-----------+--------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 401.002 | Service group <group> loss of redundancy; expected <X> standby member but no standby members available | C |
|
||||
| | | |
|
||||
| | or | |
|
||||
| | | |
|
||||
| | Service group <group> loss of redundancy; expected <X> standby member but only <Y> standby member(s) available | |
|
||||
| | | |
|
||||
| | or | |
|
||||
| | | |
|
||||
| | Service group <group> has no active members available; expected <X> active member(s). | |
|
||||
| | | |
|
||||
| | or | |
|
||||
| | | |
|
||||
| | Service group <group> loss of redundancy; expected <X> active member(s) but only <Y> active member(s) available. | |
|
||||
| | | |
|
||||
| | service_domain=<domain>.service_group=<group> | |
|
||||
+-----------+--------------------------------------------------------------------------------------------------------------------+----------+
|
@ -1,480 +0,0 @@
|
||||
|
||||
.. qfy1579789227230
|
||||
.. _customer-log-messages-700s-virtual-machines:
|
||||
|
||||
=============================================
|
||||
Customer Log Messages 700s - Virtual Machines
|
||||
=============================================
|
||||
|
||||
.. include:: /_includes/openstack-customer-log-messages-xxxs.rest
|
||||
|
||||
.. _customer-log-messages-700s-virtual-machines-table-zgf-jvw-v5:
|
||||
|
||||
.. table:: Table 1. Customer Log Messages
|
||||
:widths: auto
|
||||
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| Log ID | Description | Severity |
|
||||
+ +------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| | Entity Instance ID | |
|
||||
+==========+====================================================================================================================================================================================+==========+
|
||||
| 700.101 | Instance <instance\_name> is enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.102 | Instance <instance\_name> owned by <tenant\_name> has failed\[, reason = <reason\_text>\]. | C |
|
||||
| | Instance <instance\_name> owned by <tenant\_name> has failed to schedule\[, reason = <reason\_text>\] | |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.103 | Create issued by <tenant\_name\> or by the system against <instance\_name> owned by <tenant\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.104 | Creating instance <instance\_name> owned by <tenant\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.105 | Create rejected for instance <instance\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.106 | Create cancelled for instance <instance\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.107 | Create failed for instance <instance\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.108 | Instance <instance\_name> owned by <tenant\_name> has been created | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.109 | Delete issued by <tenant\_name\> or by the system against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.110 | Deleting instance <instance\_name> owned by <tenant\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.111 | Delete rejected for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.112 | Delete cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.113 | Delete failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.114 | Deleted instance <instance\_name> owned by <tenant\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.115 | Pause issued by <tenant\_name\> or by the system against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.116 | Pause inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.117 | Pause rejected for instance <instance\_name> enabled on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.118 | Pause cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.119 | Pause failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.120 | Pause complete for instance <instance\_name> now paused on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.121 | Unpause issued by <tenant\_name\> or by the system against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.122 | Unpause inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.123 | Unpause rejected for instance <instance\_name> paused on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.124 | Unpause cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.125 | Unpause failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.126 | Unpause complete for instance <instance\_name> now enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.127 | Suspend issued by <tenant\_name\> or by the system> against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.128 | Suspend inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.129 | Suspend rejected for instance <instance\_name> enabled on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.130 | Suspend cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.131 | Suspend failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.132 | Suspend complete for instance <instance\_name> now suspended on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.133 | Resume issued by <tenant\_name\> or by the system against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.134 | Resume inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.135 | Resume rejected for instance <instance\_name> suspended on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.136 | Resume cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.137 | Resume failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.138 | Resume complete for instance <instance\_name> now enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.139 | Start issued by <tenant\_name\> or by the system against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.140 | Start inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.141 | Start rejected for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.142 | Start cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.143 | Start failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.144 | Start complete for instance <instance\_name> now enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.145 | Stop issued by <tenant\_name>\ or by the system or by the instance against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.146 | Stop inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.147 | Stop rejected for instance <instance\_name> enabled on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.148 | Stop cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.149 | Stop failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.150 | Stop complete for instance <instance\_name> now disabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.151 | Live-Migrate issued by <tenant\_name> or by the system against instance <instance\_name> owned by <tenant\_name> from host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.152 | Live-Migrate inprogress for instance <instance\_name> from host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.153 | Live-Migrate rejected for instance <instance\_name> now on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.154 | Live-Migrate cancelled for instance <instance\_name> now on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.155 | Live-Migrate failed for instance <instance\_name> now on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.156 | Live-Migrate complete for instance <instance\_name> now enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.157 | Cold-Migrate issued by <tenant\_name> or by the system against instance <instance\_name> owned by <tenant\_name> from host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.158 | Cold-Migrate inprogress for instance <instance\_name> from host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.159 | Cold-Migrate rejected for instance <instance\_name> now on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.160 | Cold-Migrate cancelled for instance <instance\_name> now on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.161 | Cold-Migrate failed for instance <instance\_name> now on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.162 | Cold-Migrate complete for instance <instance\_name> now enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.163 | Cold-Migrate-Confirm issued by <tenant\_name> or by the system against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.164 | Cold-Migrate-Confirm inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.165 | Cold-Migrate-Confirm rejected for instance <instance\_name> now enabled on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.166 | Cold-Migrate-Confirm cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.167 | Cold-Migrate-Confirm failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.168 | Cold-Migrate-Confirm complete for instance <instance\_name> enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.169 | Cold-Migrate-Revert issued by <tenant\_name> or by the system\> against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.170 | Cold-Migrate-Revert inprogress for instance <instance\_name> from host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.171 | Cold-Migrate-Revert rejected for instance <instance\_name> now on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.172 | Cold-Migrate-Revert cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.173 | Cold-Migrate-Revert failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.174 | Cold-Migrate-Revert complete for instance <instance\_name> now enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.175 | Evacuate issued by <tenant\_name> or by the system against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.176 | Evacuating instance <instance\_name> owned by <tenant\_name> from host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.177 | Evacuate rejected for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.178 | Evacuate cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.179 | Evacuate failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.180 | Evacuate complete for instance <instance\_name> now enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.181 | Reboot <\(soft-reboot\) or \(hard-reboot\)> issued by <tenant\_name> or by the system or by the instance against instance <instance\_name> owned by | C |
|
||||
| | <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.182 | Reboot inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.183 | Reboot rejected for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.184 | Reboot cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.185 | Reboot failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.186 | Reboot complete for instance <instance\_name> now enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.187 | Rebuild issued by <tenant\_name> or by the system against instance <instance\_name> using image <image\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.188 | Rebuild inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.189 | Rebuild rejected for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.190 | Rebuild cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.191 | Rebuild failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.192 | Rebuild complete for instance <instance\_name> now enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.193 | Resize issued by <tenant\_name\> or by the system against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.194 | Resize inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.195 | Resize rejected for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.196 | Resize cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.197 | Resize failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.198 | Resize complete for instance <instance\_name> enabled on host <host\_name> waiting for confirmation | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.199 | Resize-Confirm issued by <tenant\_name> or by the system against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.200 | Resize-Confirm inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.201 | Resize-Confirm rejected for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.202 | Resize-Confirm cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.203 | Resize-Confirm failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.204 | Resize-Confirm complete for instance <instance\_name> enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.205 | Resize-Revert issued by <tenant\_name> or by the system against instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.206 | Resize-Revert inprogress for instance <instance\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.207 | Resize-Revert rejected for instance <instance\_name> owned by <tenant\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.208 | Resize-Revert cancelled for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.209 | Resize-Revert failed for instance <instance\_name> on host <host\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.210 | Resize-Revert complete for instance <instance\_name> enabled on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.214 | Instance <instance\_name> has been renamed to <new\_instance\_name> owned by <tenant\_name> on host <host\_name> | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.215 | Guest Health Check failed for instance <instance\_name>\[, reason = <reason\_text>\] | C |
|
||||
| | | |
|
||||
| | tenant=<tenant-uuid>.instance=<instance-uuid> | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.216 | Entered Multi-Node Recovery Mode | C |
|
||||
| | | |
|
||||
| | subsystem-vim | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
| 700.217 | Exited Multi-Node Recovery Mode | C |
|
||||
| | | |
|
||||
| | subsystem-vim | |
|
||||
+----------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------+
|
||||
|
||||
See also :ref:`Customer Log Messages 270s - Virtual Machines <customer-log-messages-270s-virtual-machines>`
|
@ -147,4 +147,4 @@ issues with cert-manager auto-renewal of a certificate with :command:`kubectl
|
||||
|
||||
.. seealso::
|
||||
|
||||
:ref:`500-series-alarm-messages`
|
||||
:ref:`500-series-alarm-messages-starlingx`
|
||||
|
@ -0,0 +1,18 @@
|
||||
The system inventory and maintenance service reports system changes with
|
||||
different degrees of severity. Use the reported alarms to monitor the overall
|
||||
health of the system.
|
||||
|
||||
Alarm messages are numerically coded by the type of alarm.
|
||||
|
||||
A management-affecting alarm is one that cannot be ignored at the indicated
|
||||
severity level or higher by using relaxed alarm rules during an orchestrated
|
||||
patch or upgrade operation.
|
||||
|
||||
For more information, see :ref:`fault-management-overview`.
|
||||
|
||||
.. note::
|
||||
|
||||
**Degrade Affecting Severity**: Critical indicates a node will be degraded if
|
||||
the alarm reaches a critical level.
|
||||
|
||||
|
@ -0,0 +1,16 @@
|
||||
The Customer Logs include events that do not require immediate user action.
|
||||
|
||||
The following types of events are included in the Customer Logs. The severity
|
||||
of the events can be one of:
|
||||
|
||||
* Critical
|
||||
|
||||
* Major
|
||||
|
||||
* Minor
|
||||
|
||||
* Warning
|
||||
|
||||
* NA (Not applicable)
|
||||
|
||||
|
@ -13,7 +13,7 @@ You can configure *Firmware Update Orchestration Strategy* using the
|
||||
Management-affecting alarms cannot be ignored using relaxed alarm rules
|
||||
during an orchestrated firmware update operation. For a list of
|
||||
management-affecting alarms, see |fault-doc|:
|
||||
:ref:`Alarm Messages <100-series-alarm-messages>`. To display
|
||||
:ref:`Alarm Messages <100-series-alarm-messages-starlingx>`. To display
|
||||
management-affecting active alarms, use the following command:
|
||||
|
||||
.. code-block:: none
|
||||
|
@ -19,7 +19,7 @@ You can configure *Kubernetes Version Upgrade Orchestration Strategy* using the
|
||||
Management-affecting alarms cannot be ignored using relaxed alarm rules
|
||||
during an orchestrated Kubernetes version upgrade operation. For a list of
|
||||
management-affecting alarms, see |fault-doc|: :ref:`Alarm Messages
|
||||
<100-series-alarm-messages>`. To display management-affecting active
|
||||
<100-series-alarm-messages-starlingx>`. To display management-affecting active
|
||||
alarms, use the following command:
|
||||
|
||||
.. code-block:: none
|
||||
|
@ -18,7 +18,7 @@ Management** in the left-hand pane.
|
||||
Management-affecting alarms cannot be ignored at the indicated severity
|
||||
level or higher by using relaxed alarm rules during an orchestrated update
|
||||
operation. For a list of management-affecting alarms, see |fault-doc|:
|
||||
:ref:`Alarm Messages <100-series-alarm-messages>`. To display
|
||||
:ref:`Alarm Messages <100-series-alarm-messages-starlingx>`. To display
|
||||
management-affecting active alarms, use the following command:
|
||||
|
||||
.. code-block:: none
|
||||
|
@ -25,7 +25,7 @@ About this task
|
||||
level or higher by using relaxed alarm rules during an orchestrated
|
||||
firmware update operation. For a list of management-affecting alarms, see
|
||||
|fault-doc|: :ref:`Alarm Messages
|
||||
<100-series-alarm-messages>`. To display management-affecting active
|
||||
<100-series-alarm-messages-starlingx>`. To display management-affecting active
|
||||
alarms, use the following command:
|
||||
|
||||
.. code-block:: none
|
||||
|
@ -42,7 +42,7 @@ controller host has been manually upgraded and returned to a stability state.
|
||||
Management-affecting alarms cannot be ignored at the indicated severity
|
||||
level or higher by using relaxed alarm rules during an orchestrated upgrade
|
||||
operation. For a list of management-affecting alarms, see |fault-doc|:
|
||||
:ref:`Alarm Messages <alarm-messages-overview>`. To display
|
||||
:ref:`alarm-messages-overview-19c242d3d151`. To display
|
||||
management-affecting active alarms, use the following command:
|
||||
|
||||
.. code-block:: none
|
||||
|
@ -17,7 +17,7 @@ remaining nodes of the |prod|.
|
||||
Management-affecting alarms cannot be ignored at the indicated severity
|
||||
level or higher by using relaxed alarm rules during an orchestrated upgrade
|
||||
operation. For a list of management-affecting alarms, see |fault-doc|:
|
||||
:ref:`Alarm Messages <alarm-messages-overview>`. To display
|
||||
:ref:`alarm-messages-overview-19c242d3d151`. To display
|
||||
management-affecting active alarms, use the following command:
|
||||
|
||||
.. code-block:: none
|
||||
|
@ -23,7 +23,7 @@ interface dialog, described in :ref:`Configuring Update Orchestration
|
||||
Management-affecting alarms cannot be ignored at the indicated severity
|
||||
level or higher by using relaxed alarm rules during an orchestrated update
|
||||
operation. For a list of management-affecting alarms, see |fault-doc|:
|
||||
:ref:`Alarm Messages <100-series-alarm-messages>`. To display
|
||||
:ref:`Alarm Messages <100-series-alarm-messages-starlingx>`. To display
|
||||
management-affecting active alarms, use the following command:
|
||||
|
||||
.. code-block:: none
|
||||
|
@ -1,15 +1,14 @@
|
||||
#!/usr/bin/env bash
|
||||
|
||||
# Fetch arbitrary files from a remote repo for processing/
|
||||
# Fetch arbitrary files from a remote location for processing/
|
||||
# inclusion in local build.
|
||||
|
||||
message () { echo -e "$@" 1>&2; }
|
||||
|
||||
usage_error () {
|
||||
message "\nUsage: $0 config-file -o <file|stdout> [-f -b]
|
||||
<config-file> contains fetch and save locations for files
|
||||
message "\nUsage: $0 -c config-file -o <file|stdout> [-f -b]
|
||||
-c <config-file> contains fetch and save locations for files
|
||||
-o sets the output path to the save locations in <config-file> or to STDOUT
|
||||
**Note** Do not set to \"stdout\" if you are downloading a binary file.
|
||||
-f optionally forces existing output files to be overwritten
|
||||
-b skips branch lookup. Use this if downloading from a non-git URL\n"
|
||||
exit 1
|
||||
@ -65,6 +64,7 @@ fetch_files () {
|
||||
|
||||
"file")
|
||||
_outfile="$common_target${remote_files[$f]}"
|
||||
if [ ! -d $(dirname $_outfile) ]; then mkdir -p `dirname $_outfile`; fi
|
||||
;;
|
||||
"stdout")
|
||||
_outfile="-"
|
||||
|
246
parser.py
Executable file
246
parser.py
Executable file
@ -0,0 +1,246 @@
|
||||
|
||||
import argparse
|
||||
import yaml
|
||||
from yaml.loader import SafeLoader
|
||||
import re
|
||||
import os
|
||||
|
||||
def parseLayoutFile():
|
||||
layoutFile = open(args.layout,"r")
|
||||
data = layoutFile.read()
|
||||
layoutFile.close()
|
||||
regex= r'(?<=\")([\s\S]*?)(?=\")'
|
||||
data = re.findall(regex, data)
|
||||
if(len(data) != 5):
|
||||
raise Exception('layout file has a different structure from expected')
|
||||
return [data[0], data[2], data[4]]
|
||||
|
||||
|
||||
def parseEventFile():
|
||||
eventFile = open(args.event, "r")
|
||||
data = yaml.load(eventFile, Loader=SafeLoader)
|
||||
eventFile.close()
|
||||
return data
|
||||
|
||||
def alignMultiLine(characterCounter, multilineString):
|
||||
index = 0
|
||||
strings = multilineString.split('\n')
|
||||
for string in strings:
|
||||
if(index !=0):
|
||||
string = string.lstrip()
|
||||
string = "\n"+"".join(" " for i in range(characterCounter)) + string
|
||||
strings[index] = string
|
||||
index+=1
|
||||
return "\n".join(strings)
|
||||
|
||||
def replaceSymbols(text, oldSymbols, newSymbols):
|
||||
counter = 0
|
||||
for oldSymbol in oldSymbols:
|
||||
if(len(newSymbols[counter]) - len(oldSymbols[counter]) > 0):
|
||||
text = str(text).replace(oldSymbol, " "+newSymbols[counter]+" ")
|
||||
else:
|
||||
text = str(text).replace(oldSymbol, newSymbols[counter])
|
||||
counter+=1
|
||||
return text
|
||||
|
||||
def getTitleLength(header):
|
||||
idx1 = header.find("<over_score>")+14
|
||||
idx2 = header.find("<under_score>")
|
||||
return idx2 - idx1
|
||||
|
||||
# def getContext(context):
|
||||
# if(len(context) == 3):
|
||||
# return "starlingx-openstack-empty"
|
||||
# elif(len(context) == 2):
|
||||
# if("starlingx" in context):
|
||||
# if("openstack" in context):
|
||||
# return "starlingx-openstack"
|
||||
# else:
|
||||
# return "starlingx-empty"
|
||||
# else:
|
||||
# return "openstack-empty"
|
||||
# else:
|
||||
# return context[0]
|
||||
|
||||
# RS - 11-16-22 - generalize getContext to all runtime scenarios
|
||||
def getContext(context):
|
||||
return '-'.join(map(str, context))
|
||||
|
||||
def seriesFilter(key, serie):
|
||||
if(float(serie)%100 > 1):
|
||||
return (float(key)//10) == (float(serie)//10)
|
||||
else:
|
||||
return (float(key)//100) == (float(serie)//100)
|
||||
|
||||
def seriesFile(layout, events, types, fileExtension, oldSymbols, newSymbols, products, sort ):
|
||||
series = args.series.split(",")
|
||||
for serie in series:
|
||||
matchingKeys = [key for key in events.keys() if seriesFilter(key, serie) and events[key]["Type"] in types and events[key]["Context"] in products ]
|
||||
if(sort):
|
||||
matchingKeys.sort()
|
||||
if(len(matchingKeys) > 0):
|
||||
serieFile = open(args.outputPath+str(serie)+"-series-"+'-'.join(types).lower()+"-messages"+fileExtension, "w")
|
||||
header = layout[0]
|
||||
header = header.replace("<series_number>",serie).replace("<series-number>",serie)
|
||||
score = "".join(args.titleSymbol for i in range(getTitleLength(header)))
|
||||
header = header.replace("<over_score>", score)
|
||||
header = header.replace("<Context>", getContext(products))
|
||||
header = header.replace("<under_score>", score)
|
||||
serieFile.write(header)
|
||||
for matchingKey in matchingKeys:
|
||||
body = layout[1]
|
||||
body = body.replace("<alarm_id>", format(matchingKey, '.3f'))
|
||||
fields = re.findall('(?<=\<)(.*?)(?=\>)', body)
|
||||
for field in fields:
|
||||
if(field in events[matchingKey]):
|
||||
if(type(events[matchingKey][field]) == type(events)):
|
||||
value = []
|
||||
for subkey in events[matchingKey][field]:
|
||||
value.append(events[matchingKey][field][subkey])
|
||||
events[matchingKey][field] = "\n".join(value)
|
||||
else:
|
||||
events[matchingKey][field] = (re.sub(r'\n\s*\n','\n',str(events[matchingKey][field]),re.MULTILINE))
|
||||
if(oldSymbols != None and newSymbols != None):
|
||||
events[matchingKey][field]= replaceSymbols(events[matchingKey][field], oldSymbols,newSymbols)
|
||||
if('\n' in events[matchingKey][field]):
|
||||
index = body.index('<'+field+'>')
|
||||
characterCounter= 0
|
||||
while(body[index] != '\n'):
|
||||
index-=1
|
||||
characterCounter+=1
|
||||
body = body.replace('<'+field+'>',alignMultiLine(characterCounter-1, events[matchingKey][field]))
|
||||
else:
|
||||
body = body.replace('<'+field+'>',events[matchingKey][field])
|
||||
else:
|
||||
body = body.replace('<'+field+'>','N/A')
|
||||
serieFile.write(body)
|
||||
footer = layout[2]
|
||||
serieFile.write(footer)
|
||||
serieFile.close
|
||||
|
||||
def recordsFile(layout, events, fileExtension, oldSymbols, newSymbols, sort):
|
||||
records = args.records.split(",")
|
||||
if(len(records) > 0):
|
||||
matchingKeys = [float(record) for record in records for key in events.keys() if float(key) == float(record)]
|
||||
if(sort):
|
||||
matchingKeys.sort()
|
||||
if(len(matchingKeys) > 0):
|
||||
serieFile = open(args.outputPath+args.fileName+fileExtension, "w")
|
||||
header = layout[0]
|
||||
score = "".join(args.titleSymbol for i in range(getTitleLength(header)))
|
||||
header = header.replace("<over_score>", score)
|
||||
header = header.replace("<under_score>", score)
|
||||
serieFile.write(header)
|
||||
for matchingKey in matchingKeys:
|
||||
body = layout[1]
|
||||
body = body.replace("<alarm_id>", format(matchingKey, '.3f'))
|
||||
fields = re.findall('(?<=\<)(.*?)(?=\>)', body)
|
||||
for field in fields:
|
||||
if(field in events[matchingKey]):
|
||||
if(type(events[matchingKey][field]) == type(events)):
|
||||
value = []
|
||||
for subkey in events[matchingKey][field]:
|
||||
value.append(events[matchingKey][field][subkey])
|
||||
events[matchingKey][field] = "\n".join(value)
|
||||
else:
|
||||
events[matchingKey][field] = (re.sub(r'\n\s*\n','\n',str(events[matchingKey][field]),re.MULTILINE))
|
||||
if(oldSymbols != None and newSymbols != None):
|
||||
events[matchingKey][field]= replaceSymbols(events[matchingKey][field], oldSymbols,newSymbols)
|
||||
if('\n' in events[matchingKey][field]):
|
||||
index = body.index('<'+field+'>')
|
||||
characterCounter= 0
|
||||
while(body[index] != '\n'):
|
||||
index-=1
|
||||
characterCounter+=1
|
||||
body = body.replace('<'+field+'>',alignMultiLine(characterCounter-1, events[matchingKey][field]))
|
||||
else:
|
||||
body = body.replace('<'+field+'>',events[matchingKey][field])
|
||||
else:
|
||||
body = body.replace('<'+field+'>','N/A')
|
||||
serieFile.write(body)
|
||||
footer = layout[2]
|
||||
serieFile.write(footer)
|
||||
serieFile.close
|
||||
|
||||
|
||||
parser = argparse.ArgumentParser()
|
||||
invalidArguments = False
|
||||
parser.add_argument("-l", "--layout", required=True, help = "path for the layout file")
|
||||
parser.add_argument("-e", "--event", required=True, help = "path for the events.yaml file")
|
||||
parser.add_argument("-s", "--series", help = "list of the desired series")
|
||||
parser.add_argument("-ts", "--titleSymbol", required=True, help = "Symbol used between the title")
|
||||
parser.add_argument("-replace", "--replace", required=False, help = "replaces a symbol with another")
|
||||
parser.add_argument("-type", "--type", help = "type can be Alarm or Log, it also can be both")
|
||||
parser.add_argument("-outputPath", "--outputPath", required=True, help="Path where the output will be saved")
|
||||
parser.add_argument("-records", "--records", help="list of the desired records")
|
||||
parser.add_argument("-fileName", "--fileName", help="file name for the output file")
|
||||
parser.add_argument("-product", "--product", help="product type for filtering")
|
||||
parser.add_argument("-sort", "--sort", help="argument that defines if the output will be sorted")
|
||||
|
||||
|
||||
args = parser.parse_args()
|
||||
oldSymbol = None
|
||||
newSymbol = None
|
||||
types = []
|
||||
|
||||
if(args.series == None and args.records == None):
|
||||
invalidArguments = True
|
||||
print("Expected either series or records as an argument")
|
||||
if(args.series != None and args.product == None):
|
||||
invalidArguments = True
|
||||
print("Expected product as an argument")
|
||||
if(args.series != None and args.type == None):
|
||||
invalidArguments=True
|
||||
print("Expected type as an argument")
|
||||
if(args.replace != None):
|
||||
replaceItems =args.replace.split(",")
|
||||
oldSymbols = []
|
||||
newSymbols = []
|
||||
for replaceItem in replaceItems:
|
||||
replace = replaceItem.lstrip().split(" ")
|
||||
if(len(replace) == 2):
|
||||
oldSymbols.append(replace[0])
|
||||
newSymbols.append(replace[1])
|
||||
if(args.type != None):
|
||||
types = args.type.split(",")
|
||||
counter = 0
|
||||
for recordtype in types:
|
||||
types[counter] = recordtype.lstrip().rstrip().capitalize()
|
||||
if(types[counter] != "Alarm" and types[counter]!= "Log"):
|
||||
invalidArguments = True
|
||||
print("Invalid type argument")
|
||||
counter +=1
|
||||
if(args.records != None):
|
||||
if(args.fileName == None):
|
||||
invalidArguments = True
|
||||
print("Expected fileName as an argument")
|
||||
if(args.product != None):
|
||||
products = args.product.split(",")
|
||||
counter = 0
|
||||
for product in products:
|
||||
products[counter] = product.lstrip().rstrip().lower()
|
||||
if(products[counter] != "openstack" and products[counter] != "starlingx"):
|
||||
if(products[counter] == "empty"):
|
||||
products[counter] = None
|
||||
else:
|
||||
print("Invalid product argument")
|
||||
invalidArguments= True
|
||||
counter +=1
|
||||
if(args.sort != None):
|
||||
if(args.sort.upper() == 'TRUE'):
|
||||
sort = True
|
||||
else:
|
||||
sort = False
|
||||
else:
|
||||
sort = False
|
||||
|
||||
|
||||
|
||||
if(invalidArguments == False):
|
||||
fileName, fileExtension = os.path.splitext(args.layout)
|
||||
layout = parseLayoutFile()
|
||||
events = parseEventFile()
|
||||
if(args.records != None):
|
||||
recordsFile(layout,events,fileExtension, oldSymbols, newSymbols,sort)
|
||||
else:
|
||||
seriesFile(layout, events, types, fileExtension, oldSymbols, newSymbols, products,sort)
|
40
templates/alarms_template.rst
Normal file
40
templates/alarms_template.rst
Normal file
@ -0,0 +1,40 @@
|
||||
head_format="
|
||||
|
||||
.. _<series-number>-series-alarm-messages-<Context>:
|
||||
|
||||
<over_score>
|
||||
<series_number> Series Alarm Messages
|
||||
<under_score>
|
||||
|
||||
"
|
||||
|
||||
record_format="
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Alarm ID: <alarm_id>**
|
||||
- <Description>
|
||||
* - Entity Instance
|
||||
- <Entity_Instance_ID>
|
||||
* - Degrade Affecting Severity:
|
||||
- <Degrade_Affecting_Severity>
|
||||
* - Severity:
|
||||
- <Severity>
|
||||
* - Proposed Repair Action
|
||||
- <Proposed_Repair_Action>
|
||||
* - Management Affecting Severity
|
||||
- <Management_Affecting_Severity>
|
||||
|
||||
-----
|
||||
"
|
||||
|
||||
foot_format="
|
||||
|
||||
.. raw:: html
|
||||
|
||||
 
|
||||
|
||||
"
|
||||
|
@ -1,5 +1,5 @@
|
||||
# If set, will be prepended to output paths
|
||||
common_target="doc/source/fault-mgmt/kubernetes/"
|
||||
common_target="tmp/"
|
||||
|
||||
# The repo to download from
|
||||
remote_repo="opendev.org/starlingx/fault/raw/branch"
|
||||
@ -9,5 +9,5 @@ remote_repo="opendev.org/starlingx/fault/raw/branch"
|
||||
# value is path and name to be saved, relative to CWD and common_target if set
|
||||
|
||||
remote_files=(
|
||||
["fm-doc/fm_doc/events.yaml"]="test-events.yaml"
|
||||
["fm-doc/fm_doc/events.yaml"]="events.yaml"
|
||||
)
|
||||
|
35
templates/logs_template.rst
Normal file
35
templates/logs_template.rst
Normal file
@ -0,0 +1,35 @@
|
||||
head_format="
|
||||
.. _<series-number>-series-log-messages-<Context>:
|
||||
|
||||
<over_score>
|
||||
<series_number> Log Messages
|
||||
<under_score>
|
||||
|
||||
"
|
||||
|
||||
record_format="
|
||||
|
||||
.. list-table::
|
||||
:widths: 6 25
|
||||
:header-rows: 0
|
||||
|
||||
* - **Log ID: <alarm_id>**
|
||||
- <Description>
|
||||
* - Entity Instance ID
|
||||
- <Entity_Instance_ID>
|
||||
* - Severity
|
||||
- <Severity>
|
||||
|
||||
-----
|
||||
"
|
||||
|
||||
|
||||
foot_format="
|
||||
|
||||
.. raw:: html
|
||||
|
||||
 
|
||||
|
||||
"
|
||||
|
||||
|
10
tox.ini
10
tox.ini
@ -16,10 +16,20 @@ deps =
|
||||
# -c{env:TOX_CONSTRAINTS_FILE:doc/upper-constraints.txt}
|
||||
-r{toxinidir}/doc/requirements.txt
|
||||
commands =
|
||||
git clean -dfx doc/source/fault-mgmt/
|
||||
./get-remote-files.sh -c templates/events.sh -o file -f
|
||||
python3 parser.py -l templates/alarms_template.rst -e tmp/events.yaml -s 100,200,300,400,500,750,800,900 -ts = -type Alarm -outputPath doc/source/fault-mgmt/kubernetes/ -sort Yes -product starlingx -replace "|,OR"
|
||||
python3 parser.py -l templates/logs_template.rst -e tmp/events.yaml -s 100,200,300,400,500,750,800,900 -ts = -type Log -outputPath doc/source/fault-mgmt/kubernetes/ -sort Yes -product starlingx -replace "|,OR"
|
||||
python3 parser.py -l templates/alarms_template.rst -e tmp/events.yaml -s 100,200,300,400,500,750,800,900 -ts = -type Alarm -outputPath doc/source/fault-mgmt/openstack/ -sort Yes -product openstack -replace "|,OR"
|
||||
python3 parser.py -l templates/logs_template.rst -e tmp/events.yaml -s 100,200,300,400,500,750,800,900 -ts = -type Log -outputPath doc/source/fault-mgmt/openstack/ -sort Yes -product openstack -replace "|,OR"
|
||||
sphinx-build -a -E -W --keep-going -d doc/build/doctrees -t starlingx -t openstack -b html doc/source doc/build/html {posargs}
|
||||
git clean -dfx doc/source/fault-mgmt/
|
||||
bash htmlChecks.sh doc/build/html
|
||||
whitelist_externals = bash
|
||||
htmlChecks.sh
|
||||
get-remote-files.sh
|
||||
git
|
||||
|
||||
|
||||
[testenv:api-ref]
|
||||
deps = {[testenv:docs]deps}
|
||||
|
Loading…
Reference in New Issue
Block a user