event | category | FullFormat |
ad.event.ImportCertEvent | info | ad.event.ImportCertEvent|Import certificate succeeded. |
ad.event.ImportCertFailedEvent | error | ad.event.ImportCertFailedEvent|Import certificate failed. |
ad.event.JoinDomainEvent | info | ad.event.JoinDomainEvent|Join domain succeeded. |
ad.event.JoinDomainFailedEvent | error | ad.event.JoinDomainFailedEvent|Join domain failed. |
ad.event.LeaveDomainEvent | info | ad.event.LeaveDomainEvent|Leave domain succeeded. |
ad.event.LeaveDomainFailedEvent | error | ad.event.LeaveDomainFailedEvent|Leave domain failed. |
com.vmware.license.AddLicenseEvent | info | com.vmware.license.AddLicenseEvent|License {licenseKey} added to VirtualCenter |
com.vmware.license.AssignLicenseEvent | info | com.vmware.license.AssignLicenseEvent|License {licenseKey} assigned to asset {entityName} with id {entityId} |
com.vmware.license.DLFDownloadFailedEvent | warning | com.vmware.license.DLFDownloadFailedEvent|Failed to download license information from the host {hostname} due to {errorReason.@enum.com.vmware.license.DLFDownloadFailedEvent.DLFDownloadFailedReason} |
com.vmware.license.LicenseAssignFailedEvent | info | com.vmware.license.LicenseAssignFailedEvent|License assignment on the host fails. Reasons: {errorMessage.@enum.com.vmware.license.LicenseAssignError}. |
com.vmware.license.LicenseCapacityExceededEvent | warning | com.vmware.license.LicenseCapacityExceededEvent|The current license usage ({currentUsage} {costUnitText}) for {edition} exceeds the license capacity ({capacity} {costUnitText}) |
com.vmware.license.LicenseExpiryEvent | warning | com.vmware.license.LicenseExpiryEvent|Your host license expires in {remainingDays} days. The host will disconnect from vCenter Server when its license expires. |
com.vmware.license.LicenseUserThresholdExceededEvent | warning | com.vmware.license.LicenseUserThresholdExceededEvent|The current license usage ({currentUsage} {costUnitText}) for {edition} exceeds the user-defined threshold ({threshold} {costUnitText}) |
com.vmware.license.RemoveLicenseEvent | info | com.vmware.license.RemoveLicenseEvent|License {licenseKey} removed from VirtualCenter |
com.vmware.license.UnassignLicenseEvent | info | com.vmware.license.UnassignLicenseEvent|License unassigned from asset {entityName} with id {entityId} |
com.vmware.vc.HA.AllHostAddrsPingable | info | com.vmware.vc.HA.AllHostAddrsPingable|The vSphere HA agent on the host {host.name} in cluster {computeResource.name} in {datacenter.name} can reach all the cluster management addresses |
com.vmware.vc.HA.AllIsoAddrsPingable | info | com.vmware.vc.HA.AllIsoAddrsPingable|All vSphere HA isolation addresses are reachable by host {host.name} in cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.AnsweredVmLockLostQuestionEvent | warning | com.vmware.vc.HA.AnsweredVmLockLostQuestionEvent|vSphere HA answered the lock-lost question on virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} |
com.vmware.vc.HA.AnsweredVmTerminatePDLEvent | warning | com.vmware.vc.HA.AnsweredVmTerminatePDLEvent|vSphere HA answered a question from host {host.name} in cluster {computeResource.name} about terminating virtual machine {vm.name} |
com.vmware.vc.HA.AutoStartDisabled | info | com.vmware.vc.HA.AutoStartDisabled|vSphere HA disabled the automatic Virtual Machine Startup/Shutdown feature on host {host.name} in cluster {computeResource.name} in {datacenter.name}. Automatic VM restarts will interfere with HA when reacting to a host failure. |
com.vmware.vc.HA.CannotResetVmWithInaccessibleDatastore | warning | com.vmware.vc.HA.CannotResetVmWithInaccessibleDatastore|vSphere HA did not reset VM {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} because the VM had files on inaccessible datastore(s) |
com.vmware.vc.HA.ClusterContainsIncompatibleHosts | warning | com.vmware.vc.HA.ClusterContainsIncompatibleHosts|vSphere HA Cluster {computeResource.name} in {datacenter.name} contains ESX/ESXi 3.5 hosts and more recent host versions, which isn't fully supported. |
com.vmware.vc.HA.ClusterFailoverActionCompletedEvent | info | com.vmware.vc.HA.ClusterFailoverActionCompletedEvent|vSphere HA completed a virtual machine failover action in cluster {computeResource.name} in datacenter {datacenter.name} |
com.vmware.vc.HA.ClusterFailoverActionInitiatedEvent | warning | com.vmware.vc.HA.ClusterFailoverActionInitiatedEvent|vSphere HA initiated a virtual machine failover action in cluster {computeResource.name} in datacenter {datacenter.name} |
com.vmware.vc.HA.DasAgentRunningEvent | info | com.vmware.vc.HA.DasAgentRunningEvent|vSphere HA agent on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} is running |
com.vmware.vc.HA.DasFailoverHostFailedEvent | error | com.vmware.vc.HA.DasFailoverHostFailedEvent|vSphere HA detected a possible failover of failover host {host.name} in cluster {computeResource.name} in {datacenter.name} has failed |
com.vmware.vc.HA.DasFailoverHostIsolatedEvent | error | com.vmware.vc.HA.DasFailoverHostIsolatedEvent|Host {host.name} has been isolated from cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.DasFailoverHostPartitionedEvent | warning | com.vmware.vc.HA.DasFailoverHostPartitionedEvent|Failover Host {host.name} in {computeResource.name} in {datacenter.name} is in a different network partition than the master |
com.vmware.vc.HA.DasFailoverHostUnreachableEvent | error | com.vmware.vc.HA.DasFailoverHostUnreachableEvent|The vSphere HA agent on the failover host {host.name} in cluster {computeResource.name} in {datacenter.name} is not reachable but host responds to ICMP pings |
com.vmware.vc.HA.DasHostCompleteDatastoreFailureEvent | error | com.vmware.vc.HA.DasHostCompleteDatastoreFailureEvent|All shared datastores failed on the host {hostName} in cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.DasHostCompleteNetworkFailureEvent | error | com.vmware.vc.HA.DasHostCompleteNetworkFailureEvent|All VM networks failed on the host {hostName} in cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.DasHostFailedEvent | error | com.vmware.vc.HA.DasHostFailedEvent|vSphere HA detected a possible host failure of host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} |
com.vmware.vc.HA.DasHostIsolatedEvent | error | com.vmware.vc.HA.DasHostIsolatedEvent|vSphere HA detected that host {host.name} is isolated from cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.DasHostMonitoringDisabledEvent | warning | com.vmware.vc.HA.DasHostMonitoringDisabledEvent|vSphere HA host monitoring is disabled. No virtual machine failover will occur until Host Monitoring is re-enabled for cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.FailedRestartAfterIsolationEvent | error | com.vmware.vc.HA.FailedRestartAfterIsolationEvent|vSphere HA was unable to restart virtual machine {vm.name} in cluster {computeResource.name} in datacenter {datacenter.name} after it was powered off in response to a network isolation event. The virtual machine should be manually powered back on. |
com.vmware.vc.HA.HeartbeatDatastoreChanged | info | com.vmware.vc.HA.HeartbeatDatastoreChanged|Datastore {dsName} is {changeType.@enum.com.vmware.vc.HA.HeartbeatDatastoreChange} for storage heartbeating monitored by the vSphere HA agent on host {host.name} in cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.HeartbeatDatastoreNotSufficient | warning | com.vmware.vc.HA.HeartbeatDatastoreNotSufficient|The number of vSphere HA heartbeat datastores for host {host.name} in cluster {computeResource.name} in {datacenter.name} is {selectedNum}, which is less than required: {requiredNum} |
com.vmware.vc.HA.HostAgentErrorEvent | warning | com.vmware.vc.HA.HostAgentErrorEvent|vSphere HA agent for host {host.name} has an error in {computeResource.name} in {datacenter.name}: {reason.@enum.com.vmware.vc.HA.HostAgentErrorReason} |
com.vmware.vc.HA.HostDasAgentHealthyEvent | info | com.vmware.vc.HA.HostDasAgentHealthyEvent|vSphere HA agent on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} is healthy |
com.vmware.vc.HA.HostDasErrorEvent | error | com.vmware.vc.HA.HostDasErrorEvent|vSphere HA agent on {host.name} in cluster {computeResource.name} in {datacenter.name} has an error: {reason.@enum.HostDasErrorEvent.HostDasErrorReason} |
com.vmware.vc.HA.HostDoesNotSupportVsan | error | com.vmware.vc.HA.HostDoesNotSupportVsan|vSphere HA cannot be configured on host {host.name} in cluster {computeResource.name} in {datacenter.name} because vCloud Distributed Storage is enabled but the host does not support that feature |
com.vmware.vc.HA.HostHasNoIsolationAddrsDefined | error | com.vmware.vc.HA.HostHasNoIsolationAddrsDefined|Host {host.name} in cluster {computeResource.name} in {datacenter.name} has no isolation addresses defined as required by vSphere HA. |
com.vmware.vc.HA.HostHasNoMountedDatastores | error | com.vmware.vc.HA.HostHasNoMountedDatastores|vSphere HA cannot be configured on {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} because there are no mounted datastores. |
com.vmware.vc.HA.HostHasNoSslThumbprint | error | com.vmware.vc.HA.HostHasNoSslThumbprint|vSphere HA cannot be configured on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} because its SSL thumbprint has not been verified. Check that vCenter Server is configured to verify SSL thumbprints and that the thumbprint for {host.name} has been verified. |
com.vmware.vc.HA.HostIncompatibleWithHA | error | com.vmware.vc.HA.HostIncompatibleWithHA|The product version of host {host.name} in cluster {computeResource.name} in {datacenter.name} is incompatible with vSphere HA. |
com.vmware.vc.HA.HostPartitionedFromMasterEvent | warning | com.vmware.vc.HA.HostPartitionedFromMasterEvent|vSphere HA detected that host {host.name} is in a different network partition than the master {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.HostStateChangedEvent | info | com.vmware.vc.HA.HostStateChangedEvent|The vSphere HA availability state of the host {host.name} in cluster in {computeResource.name} in {datacenter.name} has changed to {newState.@enum.com.vmware.vc.HA.DasFdmAvailabilityState} |
com.vmware.vc.HA.HostUnconfigureError | warning | com.vmware.vc.HA.HostUnconfigureError|There was an error unconfiguring the vSphere HA agent on host {host.name} in cluster {computeResource.name} in {datacenter.name}. To solve this problem, connect the host to a vCenter Server of version 5.0 or later. |
com.vmware.vc.HA.HostUnconfiguredWithProtectedVms | error | com.vmware.vc.HA.HostUnconfiguredWithProtectedVms|Host {host.name} in cluster {computeResource.name} in {datacenter.name} is disconnected from vCenter Server, but contains {protectedVmCount} protected virtual machine(s) |
com.vmware.vc.HA.InvalidMaster | warning | com.vmware.vc.HA.InvalidMaster|vSphere HA agent on host {remoteHostname} is an invalid master. The host should be examined to determine if it has been compromised. |
com.vmware.vc.HA.NotAllHostAddrsPingable | info | com.vmware.vc.HA.NotAllHostAddrsPingable|The vSphere HA agent on the host {host.name} in cluster {computeResource.name} in {datacenter.name} cannot reach some of the management network addresses of other hosts, and thus HA may not be able to restart VMs if a host failure occurs: {unpingableAddrs} |
com.vmware.vc.HA.StartFTSecondaryFailedEvent | error | com.vmware.vc.HA.StartFTSecondaryFailedEvent|vSphere HA agent failed to start Fault Tolerance secondary VM {secondaryCfgPath} on host {secondaryHost} for primary VM {vm.name} in cluster {computeResource.name} in {datacenter.name}. Reason : {fault.msg}. vSphere HA agent will retry until it times out. |
com.vmware.vc.HA.StartFTSecondarySucceededEvent | info | com.vmware.vc.HA.StartFTSecondarySucceededEvent|vSphere HA agent successfully started Fault Tolerance secondary VM {secondaryCfgPath} on host {secondaryHost} for primary VM {vm.name} in cluster {computeResource.name}. |
com.vmware.vc.HA.UserHeartbeatDatastoreRemoved | warning | com.vmware.vc.HA.UserHeartbeatDatastoreRemoved|vSphere HA removed datastore {dsName} from the set of preferred heartbeat datastores selected for cluster {computeResource.name} in {datacenter.name} because the datastore is removed from inventory |
com.vmware.vc.HA.VMIsHADisabledIsolationEvent | info | com.vmware.vc.HA.VMIsHADisabledIsolationEvent|vSphere HA did not perform an isolation response for {vm.name} in cluster {computeResource.name} in {datacenter.name} because its VM restart priority is Disabled |
com.vmware.vc.HA.VMIsHADisabledRestartEvent | info | com.vmware.vc.HA.VMIsHADisabledRestartEvent|vSphere HA did not attempt to restart {vm.name} in cluster {computeResource.name} in {datacenter.name} because its VM restart priority is Disabled |
com.vmware.vc.HA.VcCannotFindMasterEvent | warning | com.vmware.vc.HA.VcCannotFindMasterEvent|vCenter Server is unable to find a master vSphere HA agent in cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.VcConnectedToMasterEvent | info | com.vmware.vc.HA.VcConnectedToMasterEvent|vCenter Server is connected to a master HA agent running on host {hostname} in {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.VcDisconnectedFromMasterEvent | warning | com.vmware.vc.HA.VcDisconnectedFromMasterEvent|vCenter Server is disconnected from a master HA agent running on host {hostname} in {computeResource.name} in {datacenter.name} |
com.vmware.vc.HA.VmNotProtectedEvent | error | com.vmware.vc.HA.VmNotProtectedEvent|Virtual machine {vm.name} in cluster {computeResource.name} in {datacenter.name} failed to become vSphere HA Protected and HA may not attempt to restart it after a failure. |
com.vmware.vc.HA.VmProtectedEvent | info | com.vmware.vc.HA.VmProtectedEvent|Virtual machine {vm.name} in cluster {computeResource.name} in {datacenter.name} is vSphere HA Protected and HA will attempt to restart it after a failure. |
com.vmware.vc.HA.VmUnprotectedEvent | info | com.vmware.vc.HA.VmUnprotectedEvent|Virtual machine {vm.name} in cluster {computeResource.name} in {datacenter.name} is not vSphere HA Protected. |
com.vmware.vc.HA.VmUnprotectedOnDiskSpaceFull | info | com.vmware.vc.HA.VmUnprotectedOnDiskSpaceFull|vSphere HA has unprotected virtual machine {vm.name} in cluster {computeResource.name} in datacenter {datacenter.name} because it ran out of disk space |
com.vmware.vc.VCHealthStateChangedEvent | info | com.vmware.vc.VCHealthStateChangedEvent|vCenter Service overall health changed from '{oldState}' to '{newState}' |
com.vmware.vc.VmCloneFailedInvalidDestinationEvent | error | com.vmware.vc.VmCloneFailedInvalidDestinationEvent|Cannot clone {vm.name} as {destVmName} to invalid or non-existent destination with ID {invalidMoRef}: {fault} |
com.vmware.vc.VmCloneToResourcePoolFailedEvent | error | com.vmware.vc.VmCloneToResourcePoolFailedEvent|Cannot clone {vm.name} as {destVmName} to resource pool {destResourcePool}: {fault} |
com.vmware.vc.VmDiskConsolidatedEvent | info | com.vmware.vc.VmDiskConsolidatedEvent|Virtual machine {vm.name} disks consolidated successfully on {host.name} in cluster {computeResource.name} in {datacenter.name}. |
com.vmware.vc.VmDiskConsolidationNeeded | warning | com.vmware.vc.VmDiskConsolidationNeeded|Virtual machine {vm.name} disks consolidation is needed on {host.name} in cluster {computeResource.name} in {datacenter.name}. |
com.vmware.vc.VmDiskConsolidationNoLongerNeeded | info | com.vmware.vc.VmDiskConsolidationNoLongerNeeded|Virtual machine {vm.name} disks consolidation is no longer needed on {host.name} in cluster {computeResource.name} in {datacenter.name}. |
com.vmware.vc.VmDiskFailedToConsolidateEvent | warning | com.vmware.vc.VmDiskFailedToConsolidateEvent|Virtual machine {vm.name} disks consolidation failed on {host.name} in cluster {computeResource.name} in {datacenter.name}. |
com.vmware.vc.cim.CIMGroupHealthStateChanged | info | com.vmware.vc.cim.CIMGroupHealthStateChanged|Health of [data.group] changed from [data.oldState] to [data.newState]. [data.cause] |
com.vmware.vc.datastore.UpdateVmFilesFailedEvent | error | com.vmware.vc.datastore.UpdateVmFilesFailedEvent|Failed to update VM files on datastore {ds.name} using host {hostName} |
com.vmware.vc.datastore.UpdatedVmFilesEvent | info | com.vmware.vc.datastore.UpdatedVmFilesEvent|Updated VM files on datastore {ds.name} using host {hostName} |
com.vmware.vc.datastore.UpdatingVmFilesEvent | info | com.vmware.vc.datastore.UpdatingVmFilesEvent|Updating VM files on datastore {ds.name} using host {hostName} |
com.vmware.vc.dvs.LacpConfigInconsistentEvent | info | com.vmware.vc.dvs.LacpConfigInconsistentEvent|Single Link Aggregation Control Group is enabled on Uplink Port Groups while enhanced LACP support is enabled. |
com.vmware.vc.ft.VmAffectedByDasDisabledEvent | warning | com.vmware.vc.ft.VmAffectedByDasDisabledEvent|vSphere HA has been disabled in cluster {computeResource.name} of datacenter {datacenter.name}. vSphere HA will not restart VM {vm.name} or its Secondary VM after a failure. |
com.vmware.vc.guestOperations.GuestOperation | info | com.vmware.vc.guestOperations.GuestOperation|Guest operation {operationName.@enum.com.vmware.vc.guestOp} performed on Virtual machine {vm.name}. |
com.vmware.vc.guestOperations.GuestOperationAuthFailure | warning | com.vmware.vc.guestOperations.GuestOperationAuthFailure|Guest operation authentication failed for operation {operationName.@enum.com.vmware.vc.guestOp} on Virtual machine {vm.name}. |
com.vmware.vc.ha.VmRestartedByHAEvent | warning | com.vmware.vc.ha.VmRestartedByHAEvent|vSphere HA restarted virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} |
com.vmware.vc.host.AutoStartReconfigureFailedEvent | error | com.vmware.vc.host.AutoStartReconfigureFailedEvent|Reconfiguring autostart rules for virtual machines on {host.name} in datacenter {datacenter.name} failed |
com.vmware.vc.host.clear.vFlashResource.inaccessible | info | com.vmware.vc.host.clear.vFlashResource.inaccessible|Host's virtual flash resource is restored to be accessible. |
com.vmware.vc.host.clear.vFlashResource.reachthreshold | info | com.vmware.vc.host.clear.vFlashResource.reachthreshold|Host's virtual flash resource usage dropped below {1}%. |
com.vmware.vc.host.problem.vFlashResource.inaccessible | warning | com.vmware.vc.host.problem.vFlashResource.inaccessible|Host's virtual flash resource is inaccessible. |
com.vmware.vc.host.problem.vFlashResource.reachthreshold | warning | com.vmware.vc.host.problem.vFlashResource.reachthreshold|Host's virtual flash resource usage is more than {1}%. |
com.vmware.vc.host.vFlash.VFlashResourceCapacityExtendedEvent | info | com.vmware.vc.host.vFlash.VFlashResourceCapacityExtendedEvent|Virtual flash resource capacity is extended |
com.vmware.vc.host.vFlash.VFlashResourceConfiguredEvent | info | com.vmware.vc.host.vFlash.VFlashResourceConfiguredEvent|Virtual flash resource is configured on the host |
com.vmware.vc.host.vFlash.VFlashResourceRemovedEvent | info | com.vmware.vc.host.vFlash.VFlashResourceRemovedEvent|Virtual flash resource is removed from the host |
com.vmware.vc.host.vFlash.defaultModuleChangedEvent | info | com.vmware.vc.host.vFlash.defaultModuleChangedEvent|Any new virtual Flash Read Cache configuration request will use {vFlashModule} as default virtual flash module. All existing virtual Flash Read Cache configurations remain unchanged. |
com.vmware.vc.host.vFlash.modulesLoadedEvent | info | com.vmware.vc.host.vFlash.modulesLoadedEvent|Virtual flash modules are loaded or reloaded on the host |
com.vmware.vc.npt.VmAdapterEnteredPassthroughEvent | info | com.vmware.vc.npt.VmAdapterEnteredPassthroughEvent|Network passthrough is active on adapter {deviceLabel} of virtual machine {vm.name} on host {host.name} in {datacenter.name} |
com.vmware.vc.npt.VmAdapterExitedPassthroughEvent | info | com.vmware.vc.npt.VmAdapterExitedPassthroughEvent|Network passthrough is inactive on adapter {deviceLabel} of virtual machine {vm.name} on host {host.name} in {datacenter.name} |
com.vmware.vc.ovfconsumers.CloneOvfConsumerStateErrorEvent | error | com.vmware.vc.ovfconsumers.CloneOvfConsumerStateErrorEvent|Failed to clone state for the entity '{entityName}' on extension {extensionName} |
com.vmware.vc.ovfconsumers.GetOvfEnvironmentSectionsErrorEvent | error | com.vmware.vc.ovfconsumers.GetOvfEnvironmentSectionsErrorEvent|Failed to retrieve OVF environment sections for VM '{vm.name}' from extension {extensionName} |
com.vmware.vc.ovfconsumers.PowerOnAfterCloneErrorEvent | error | com.vmware.vc.ovfconsumers.PowerOnAfterCloneErrorEvent|Powering on VM '{vm.name}' after cloning was blocked by an extension. Message: {description} |
com.vmware.vc.ovfconsumers.RegisterEntityErrorEvent | error | com.vmware.vc.ovfconsumers.RegisterEntityErrorEvent|Failed to register entity '{entityName}' on extension {extensionName} |
com.vmware.vc.ovfconsumers.UnregisterEntitiesErrorEvent | error | com.vmware.vc.ovfconsumers.UnregisterEntitiesErrorEvent|Failed to unregister entities on extension {extensionName} |
com.vmware.vc.ovfconsumers.ValidateOstErrorEvent | error | com.vmware.vc.ovfconsumers.ValidateOstErrorEvent|Failed to validate OVF descriptor on extension {extensionName} |
com.vmware.vc.profile.AnswerFileExportedEvent | info | com.vmware.vc.profile.AnswerFileExportedEvent|Answer file for host {host.name} in datacenter {datacenter.name} has been exported |
com.vmware.vc.profile.AnswerFileUpdatedEvent | info | com.vmware.vc.profile.AnswerFileUpdatedEvent|Host customization settings for host {host.name} in datacenter {datacenter.name} has been updated |
com.vmware.vc.rp.ResourcePoolRenamedEvent | info | com.vmware.vc.rp.ResourcePoolRenamedEvent|Resource pool '{oldName}' has been renamed to '{newName}' |
com.vmware.vc.sdrs.CanceledDatastoreMaintenanceModeEvent | info | com.vmware.vc.sdrs.CanceledDatastoreMaintenanceModeEvent|The datastore maintenance mode operation has been canceled |
com.vmware.vc.sdrs.ConfiguredStorageDrsOnPodEvent | info | com.vmware.vc.sdrs.ConfiguredStorageDrsOnPodEvent|Configured storage DRS on datastore cluster {objectName} |
com.vmware.vc.sdrs.ConsistencyGroupViolationEvent | warning | com.vmware.vc.sdrs.ConsistencyGroupViolationEvent|Datastore cluster {objectName} has datastores that belong to different SRM Consistency Groups |
com.vmware.vc.sdrs.DatastoreEnteredMaintenanceModeEvent | info | com.vmware.vc.sdrs.DatastoreEnteredMaintenanceModeEvent|Datastore {ds.name} has entered maintenance mode |
com.vmware.vc.sdrs.DatastoreEnteringMaintenanceModeEvent | info | com.vmware.vc.sdrs.DatastoreEnteringMaintenanceModeEvent|Datastore {ds.name} is entering maintenance mode |
com.vmware.vc.sdrs.DatastoreExitedMaintenanceModeEvent | info | com.vmware.vc.sdrs.DatastoreExitedMaintenanceModeEvent|Datastore {ds.name} has exited maintenance mode |
com.vmware.vc.sdrs.DatastoreInMultipleDatacentersEvent | warning | com.vmware.vc.sdrs.DatastoreInMultipleDatacentersEvent|Datastore cluster {objectName} has one or more datastores {datastore} shared across multiple datacenters |
com.vmware.vc.sdrs.DatastoreMaintenanceModeErrorsEvent | error | com.vmware.vc.sdrs.DatastoreMaintenanceModeErrorsEvent|Datastore {ds.name} encountered errors while entering maintenance mode |
com.vmware.vc.sdrs.StorageDrsDisabledEvent | info | com.vmware.vc.sdrs.StorageDrsDisabledEvent|Disabled storage DRS on datastore cluster {objectName} |
com.vmware.vc.sdrs.StorageDrsEnabledEvent | info | com.vmware.vc.sdrs.StorageDrsEnabledEvent|Enabled storage DRS on datastore cluster {objectName} with automation level {behavior.@enum.storageDrs.PodConfigInfo.Behavior} |
com.vmware.vc.sdrs.StorageDrsInvocationFailedEvent | error | com.vmware.vc.sdrs.StorageDrsInvocationFailedEvent|Storage DRS invocation failed on datastore cluster {objectName} |
com.vmware.vc.sdrs.StorageDrsNewRecommendationPendingEvent | info | com.vmware.vc.sdrs.StorageDrsNewRecommendationPendingEvent|A new storage DRS recommendation has been generated on datastore cluster {objectName} |
com.vmware.vc.sdrs.StorageDrsNotSupportedHostConnectedToPodEvent | warning | com.vmware.vc.sdrs.StorageDrsNotSupportedHostConnectedToPodEvent|Datastore cluster {objectName} is connected to one or more hosts {host} that do not support storage DRS |
com.vmware.vc.sdrs.StorageDrsRecommendationApplied | info | com.vmware.vc.sdrs.StorageDrsRecommendationApplied|All pending recommendations on datastore cluster {objectName} were applied |
com.vmware.vc.sdrs.StorageDrsStorageMigrationEvent | info | com.vmware.vc.sdrs.StorageDrsStorageMigrationEvent|Storage DRS migrated disks of VM {vm.name} to datastore {ds.name} |
com.vmware.vc.sdrs.StorageDrsStoragePlacementEvent | info | com.vmware.vc.sdrs.StorageDrsStoragePlacementEvent|Storage DRS placed disks of VM {vm.name} on datastore {ds.name} |
com.vmware.vc.sdrs.StoragePodCreatedEvent | info | com.vmware.vc.sdrs.StoragePodCreatedEvent|Created datastore cluster {objectName} |
com.vmware.vc.sdrs.StoragePodDestroyedEvent | info | com.vmware.vc.sdrs.StoragePodDestroyedEvent|Removed datastore cluster {objectName} |
com.vmware.vc.sioc.NotSupportedHostConnectedToDatastoreEvent | warning | com.vmware.vc.sioc.NotSupportedHostConnectedToDatastoreEvent|SIOC has detected that a host: {host} connected to a SIOC-enabled datastore: {objectName} is running an older version of ESX that does not support SIOC. This is an unsupported configuration. |
com.vmware.vc.sms.LunCapabilityInitEvent | warning | com.vmware.vc.sms.LunCapabilityInitEvent|Storage provider [{providerName}] : system capability warning for {eventSubjectId} : {msgTxt} |
com.vmware.vc.sms.LunCapabilityMetEvent | info | com.vmware.vc.sms.LunCapabilityMetEvent|Storage provider [{providerName}] : system capability normal for {eventSubjectId} |
com.vmware.vc.sms.LunCapabilityNotMetEvent | error | com.vmware.vc.sms.LunCapabilityNotMetEvent|Storage provider [{providerName}] : system capability alert for {eventSubjectId} : {msgTxt} |
com.vmware.vc.sms.ThinProvisionedLunThresholdClearedEvent | info | com.vmware.vc.sms.ThinProvisionedLunThresholdClearedEvent|Storage provider [{providerName}] : thin provisioning capacity threshold normal for {eventSubjectId} |
com.vmware.vc.sms.ThinProvisionedLunThresholdCrossedEvent | error | com.vmware.vc.sms.ThinProvisionedLunThresholdCrossedEvent|Storage provider [{providerName}] : thin provisioning capacity threshold alert for {eventSubjectId} |
com.vmware.vc.sms.ThinProvisionedLunThresholdInitEvent | warning | com.vmware.vc.sms.ThinProvisionedLunThresholdInitEvent|Storage provider [{providerName}] : thin provisioning capacity threshold warning for {eventSubjectId} |
com.vmware.vc.sms.datastore.ComplianceStatusCompliantEvent | info | com.vmware.vc.sms.datastore.ComplianceStatusCompliantEvent|Virtual disk {diskKey} on {vmName} connected to datastore {datastore.name} in {datacenter.name} is compliant from storage provider {providerName}. |
com.vmware.vc.sms.datastore.ComplianceStatusNonCompliantEvent | error | com.vmware.vc.sms.datastore.ComplianceStatusNonCompliantEvent|Virtual disk {diskKey} on {vmName} connected to {datastore.name} in {datacenter.name} is not compliant {operationalStatus] from storage provider {providerName}. |
com.vmware.vc.sms.datastore.ComplianceStatusUnknownEvent | warning | com.vmware.vc.sms.datastore.ComplianceStatusUnknownEvent|Virtual disk {diskKey} on {vmName} connected to {datastore.name} in {datacenter.name} compliance status is unknown from storage provider {providerName}. |
com.vmware.vc.sms.provider.health.event | info | com.vmware.vc.sms.provider.health.event|Storage provider [{providerName}] : health event for {eventSubjectId} : {msgTxt} |
com.vmware.vc.sms.provider.system.event | info | com.vmware.vc.sms.provider.system.event|Storage provider [{providerName}] : system event : {msgTxt} |
com.vmware.vc.sms.vm.ComplianceStatusCompliantEvent | info | com.vmware.vc.sms.vm.ComplianceStatusCompliantEvent|Virtual disk {diskKey} on {vm.name} on {host.name} and {computeResource.name} in {datacenter.name} is compliant from storage provider {providerName}. |
com.vmware.vc.sms.vm.ComplianceStatusNonCompliantEvent | error | com.vmware.vc.sms.vm.ComplianceStatusNonCompliantEvent|Virtual disk {diskKey} on {vm.name} on {host.name} and {computeResource.name} in {datacenter.name} is not compliant {operationalStatus] from storage provider {providerName}. |
com.vmware.vc.sms.vm.ComplianceStatusUnknownEvent | warning | com.vmware.vc.sms.vm.ComplianceStatusUnknownEvent|Virtual disk {diskKey} on {vm.name} on {host.name} and {computeResource.name} in {datacenter.name} compliance status is unknown from storage provider {providerName}. |
com.vmware.vc.spbm.ProfileAssociationFailedEvent | error | com.vmware.vc.spbm.ProfileAssociationFailedEvent|Profile association/dissociation failed for {entityName} |
com.vmware.vc.stats.HostQuickStatesNotUpToDateEvent | info | com.vmware.vc.stats.HostQuickStatesNotUpToDateEvent|Quick stats on {host.name} in {computeResource.name} in {datacenter.name} is not up-to-date |
com.vmware.vc.vcp.FtDisabledVmTreatAsNonFtEvent | info | com.vmware.vc.vcp.FtDisabledVmTreatAsNonFtEvent|HA VM Component Protection protects virtual machine {vm.name} on {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} as non-FT virtual machine because the FT state is disabled |
com.vmware.vc.vcp.FtFailoverEvent | info | com.vmware.vc.vcp.FtFailoverEvent|FT Primary VM {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} is going to fail over to Secondary VM due to component failure |
com.vmware.vc.vcp.FtFailoverFailedEvent | error | com.vmware.vc.vcp.FtFailoverFailedEvent|FT virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} failed to failover to secondary |
com.vmware.vc.vcp.FtSecondaryRestartEvent | info | com.vmware.vc.vcp.FtSecondaryRestartEvent|HA VM Component Protection is restarting FT secondary virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} due to component failure |
com.vmware.vc.vcp.FtSecondaryRestartFailedEvent | error | com.vmware.vc.vcp.FtSecondaryRestartFailedEvent|FT Secondary VM {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} failed to restart |
com.vmware.vc.vcp.NeedSecondaryFtVmTreatAsNonFtEvent | info | com.vmware.vc.vcp.NeedSecondaryFtVmTreatAsNonFtEvent|HA VM Component Protection protects virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} as non-FT virtual machine because it has been in the needSecondary state too long |
com.vmware.vc.vcp.TestEndEvent | info | com.vmware.vc.vcp.TestEndEvent|VM Component Protection test ends on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} |
com.vmware.vc.vcp.TestStartEvent | info | com.vmware.vc.vcp.TestStartEvent|VM Component Protection test starts on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} |
com.vmware.vc.vcp.VcpNoActionEvent | info | com.vmware.vc.vcp.VcpNoActionEvent|HA VM Component Protection did not take action on virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} due to the feature configuration setting |
com.vmware.vc.vcp.VmDatastoreFailedEvent | error | com.vmware.vc.vcp.VmDatastoreFailedEvent|Virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} lost access to {datastore} |
com.vmware.vc.vcp.VmNetworkFailedEvent | error | com.vmware.vc.vcp.VmNetworkFailedEvent|Virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} lost access to {network} |
com.vmware.vc.vcp.VmPowerOffHangEvent | error | com.vmware.vc.vcp.VmPowerOffHangEvent|HA VM Component Protection could not power off virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} successfully after trying {numTimes} times and will keep trying |
com.vmware.vc.vcp.VmRestartEvent | info | com.vmware.vc.vcp.VmRestartEvent|HA VM Component Protection is restarting virtual machine {vm.name} due to component failure on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} |
com.vmware.vc.vcp.VmRestartFailedEvent | error | com.vmware.vc.vcp.VmRestartFailedEvent|Virtual machine {vm.name} affected by component failure on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} failed to restart |
com.vmware.vc.vcp.VmWaitForCandidateHostEvent | error | com.vmware.vc.vcp.VmWaitForCandidateHostEvent|HA VM Component Protection could not find a destination host for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} after waiting {numSecWait} seconds and will keep trying |
com.vmware.vc.vflash.SsdConfigurationFailedEvent | error | com.vmware.vc.vflash.SsdConfigurationFailedEvent|Configuration on disk {disk.path} failed. Reason : {fault.msg} |
com.vmware.vc.vm.VmRegisterFailedEvent | error | com.vmware.vc.vm.VmRegisterFailedEvent|Virtual machine {vm.name} registration on {host.name} in datacenter {datacenter.name} failed |
com.vmware.vc.vm.VmStateFailedToRevertToSnapshot | error | com.vmware.vc.vm.VmStateFailedToRevertToSnapshot|Failed to revert the execution state of the virtual machine {vm.name} on host {host.name}, in compute resource {computeResource.name} to snapshot {snapshotName}, with ID {snapshotId} |
com.vmware.vc.vm.VmStateRevertedToSnapshot | info | com.vmware.vc.vm.VmStateRevertedToSnapshot|The execution state of the virtual machine {vm.name} on host {host.name}, in compute resource {computeResource.name} has been reverted to the state of snapshot {snapshotName}, with ID {snapshotId} |
com.vmware.vc.vmam.AppMonitoringNotSupported | warning | com.vmware.vc.vmam.AppMonitoringNotSupported|Application monitoring is not supported on {host.name} in cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.vmam.VmAppHealthMonitoringStateChangedEvent | warning | com.vmware.vc.vmam.VmAppHealthMonitoringStateChangedEvent|vSphere HA detected that the application heartbeat status changed to {status.@enum.VirtualMachine.AppHeartbeatStatusType} for {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.vmam.VmAppHealthStateChangedEvent | warning | com.vmware.vc.vmam.VmAppHealthStateChangedEvent|vSphere HA detected that the application state changed to {state.@enum.vm.GuestInfo.AppStateType} for {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.vmam.VmDasAppHeartbeatFailedEvent | warning | com.vmware.vc.vmam.VmDasAppHeartbeatFailedEvent|vSphere HA detected application heartbeat failure for {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name} |
com.vmware.vc.vsan.DatastoreNoCapacityEvent | error | com.vmware.vc.vsan.DatastoreNoCapacityEvent|VSAN datastore {datastoreName} in cluster {computeResource.name} in datacenter {datacenter.name} does not have capacity |
com.vmware.vc.vsan.HostCommunicationErrorEvent | error | com.vmware.vc.vsan.HostCommunicationErrorEvent|Host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} cannot communicate with all other nodes in the VSAN enabled cluster |
com.vmware.vc.vsan.HostNotInClusterEvent | error | com.vmware.vc.vsan.HostNotInClusterEvent|{host.name} with the VSAN service enabled is not in the vCenter cluster {computeResource.name} in datacenter {datacenter.name} |
com.vmware.vc.vsan.HostNotInVsanClusterEvent | error | com.vmware.vc.vsan.HostNotInVsanClusterEvent|{host.name} is in a VSAN enabled cluster {computeResource.name} in datacenter {datacenter.name} but does not have VSAN service enabled |
com.vmware.vc.vsan.HostVendorProviderDeregistrationFailedEvent | error | com.vmware.vc.vsan.HostVendorProviderDeregistrationFailedEvent|Vendor provider {host.name} deregistration failed |
com.vmware.vc.vsan.HostVendorProviderDeregistrationSuccessEvent | info | com.vmware.vc.vsan.HostVendorProviderDeregistrationSuccessEvent|Vendor provider {host.name} deregistration succeeded |
com.vmware.vc.vsan.HostVendorProviderRegistrationFailedEvent | error | com.vmware.vc.vsan.HostVendorProviderRegistrationFailedEvent|Vendor provider {host.name} registration failed |
com.vmware.vc.vsan.HostVendorProviderRegistrationSuccessEvent | info | com.vmware.vc.vsan.HostVendorProviderRegistrationSuccessEvent|Vendor provider {host.name} registration succeeded |
com.vmware.vc.vsan.NetworkMisConfiguredEvent | error | com.vmware.vc.vsan.NetworkMisConfiguredEvent|VSAN network is not configured on {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} |
com.vmware.vc.vsan.RogueHostFoundEvent | error | com.vmware.vc.vsan.RogueHostFoundEvent|Found another host participating in the VSAN service in cluster {computeResource.name} in datacenter {datacenter.name} which is not a member of this host's vCenter cluster |
com.vmware.vim.eam.agency.create | info | com.vmware.vim.eam.agency.create|{agencyName} created by {ownerName} |
com.vmware.vim.eam.agency.destroyed | info | com.vmware.vim.eam.agency.destroyed|{agencyName} removed from the vSphere ESX Agent Manager |
com.vmware.vim.eam.agency.goalstate | info | com.vmware.vim.eam.agency.goalstate|{agencyName} changed goal state from {oldGoalState} to {newGoalState} |
com.vmware.vim.eam.agency.statusChanged | info | com.vmware.vim.eam.agency.statusChanged|Agency status changed from {oldStatus} to {newStatus} |
com.vmware.vim.eam.agency.updated | info | com.vmware.vim.eam.agency.updated|Configuration updated {agencyName} |
com.vmware.vim.eam.agent.created | info | com.vmware.vim.eam.agent.created|Agent added to host {host.name} ({agencyName}) |
com.vmware.vim.eam.agent.destroyed | info | com.vmware.vim.eam.agent.destroyed|Agent removed from host {host.name} ({agencyName}) |
com.vmware.vim.eam.agent.destroyedNoHost | info | com.vmware.vim.eam.agent.destroyedNoHost|Agent removed from host ({agencyName}) |
com.vmware.vim.eam.agent.markAgentVmAsAvailableAfterPowerOn | info | com.vmware.vim.eam.agent.markAgentVmAsAvailableAfterPowerOn|Agent VM {vm.name} has been powered on. Mark agent as available to proceed agent workflow ({agencyName}) |
com.vmware.vim.eam.agent.markAgentVmAsAvailableAfterProvisioning | info | com.vmware.vim.eam.agent.markAgentVmAsAvailableAfterProvisioning|Agent VM {vm.name} has been provisioned. Mark agent as available to proceed agent workflow ({agencyName}) |
com.vmware.vim.eam.agent.statusChanged | info | com.vmware.vim.eam.agent.statusChanged|Agent status changed from {oldStatus} to {newStatus} |
com.vmware.vim.eam.agent.task.deleteVm | info | com.vmware.vim.eam.agent.task.deleteVm|Agent VM {vmName} is deleted on host {host.name} ({agencyName}) |
com.vmware.vim.eam.agent.task.deployVm | info | com.vmware.vim.eam.agent.task.deployVm|Agent VM {vm.name} is provisioned on host {host.name} ({agencyName}) |
com.vmware.vim.eam.agent.task.powerOffVm | info | com.vmware.vim.eam.agent.task.powerOffVm|Agent VM {vm.name} powered off, on host {host.name} ({agencyName}) |
com.vmware.vim.eam.agent.task.powerOnVm | info | com.vmware.vim.eam.agent.task.powerOnVm|Agent VM {vm.name} powered on, on host {host.name} ({agencyName}) |
com.vmware.vim.eam.agent.task.vibInstalled | info | com.vmware.vim.eam.agent.task.vibInstalled|Agent installed VIB {vib} on host {host.name} ({agencyName}) |
com.vmware.vim.eam.agent.task.vibUninstalled | info | com.vmware.vim.eam.agent.task.vibUninstalled|Agent uninstalled VIB {vib} on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.cannotAccessAgentOVF | error | com.vmware.vim.eam.issue.cannotAccessAgentOVF|Unable to access agent OVF package at {url} ({agencyName}) |
com.vmware.vim.eam.issue.cannotAccessAgentVib | error | com.vmware.vim.eam.issue.cannotAccessAgentVib|Unable to access agent VIB module at {url} ({agencyName}) |
com.vmware.vim.eam.issue.hostInMaintenanceMode | error | com.vmware.vim.eam.issue.hostInMaintenanceMode|Agent cannot complete an operation since the host {host.name} is in maintenance mode ({agencyName}) |
com.vmware.vim.eam.issue.hostInStandbyMode | error | com.vmware.vim.eam.issue.hostInStandbyMode|Agent cannot complete an operation since the host {host.name} is in standby mode ({agencyName}) |
com.vmware.vim.eam.issue.hostPoweredOff | error | com.vmware.vim.eam.issue.hostPoweredOff|Agent cannot complete an operation since the host {host.name} is powered off ({agencyName}) |
com.vmware.vim.eam.issue.incompatibleHostVersion | error | com.vmware.vim.eam.issue.incompatibleHostVersion|Agent is not deployed due to incompatible host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.insufficientIpAddresses | error | com.vmware.vim.eam.issue.insufficientIpAddresses|Insufficient IP addresses in network protocol profile in agent's VM network ({agencyName}) |
com.vmware.vim.eam.issue.insufficientResources | error | com.vmware.vim.eam.issue.insufficientResources|Agent cannot be provisioned due to insufficient resources on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.insufficientSpace | error | com.vmware.vim.eam.issue.insufficientSpace|Agent on {host.name} cannot be provisioned due to insufficient space on datastore ({agencyName}) |
com.vmware.vim.eam.issue.missingAgentIpPool | error | com.vmware.vim.eam.issue.missingAgentIpPool|No network protocol profile associated to agent's VM network ({agencyname}) |
com.vmware.vim.eam.issue.missingDvFilterSwitch | error | com.vmware.vim.eam.issue.missingDvFilterSwitch|dvFilter switch is not configured on host {host.name} ({agencyname}) |
com.vmware.vim.eam.issue.noAgentVmDatastore | error | com.vmware.vim.eam.issue.noAgentVmDatastore|No agent datastore configuration on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.noAgentVmNetwork | error | com.vmware.vim.eam.issue.noAgentVmNetwork|No agent network configuration on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.noCustomAgentVmDatastore | error | com.vmware.vim.eam.issue.noCustomAgentVmDatastore|Agent datastore(s) {customAgentVmDatastoreName} not available on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.noCustomAgentVmNetwork | error | com.vmware.vim.eam.issue.noCustomAgentVmNetwork|Agent network(s) {customAgentVmNetworkName} not available on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.orphandedDvFilterSwitch | error | com.vmware.vim.eam.issue.orphandedDvFilterSwitch|Unused dvFilter switch on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.orphanedAgency | error | com.vmware.vim.eam.issue.orphanedAgency|Orphaned agency found. ({agencyName}) |
com.vmware.vim.eam.issue.ovfInvalidFormat | error | com.vmware.vim.eam.issue.ovfInvalidFormat|OVF used to provision agent on host {host.name} has invalid format ({agencyName}) |
com.vmware.vim.eam.issue.ovfInvalidProperty | error | com.vmware.vim.eam.issue.ovfInvalidProperty|OVF environment used to provision agent on host {host.name} has one or more invalid properties ({agencyName}) |
com.vmware.vim.eam.issue.resolved | info | com.vmware.vim.eam.issue.resolved|Issue {type} resolved (key {key}) |
com.vmware.vim.eam.issue.unknownAgentVm | error | com.vmware.vim.eam.issue.unknownAgentVm|Unknown agent VM {vm.name} |
com.vmware.vim.eam.issue.vibCannotPutHostInMaintenanceMode | error | com.vmware.vim.eam.issue.vibCannotPutHostInMaintenanceMode|Cannot put host into maintenance mode ({agencyName}) |
com.vmware.vim.eam.issue.vibInvalidFormat | error | com.vmware.vim.eam.issue.vibInvalidFormat|Invalid format for VIB module at {url} ({agencyName}) |
com.vmware.vim.eam.issue.vibNotInstalled | error | com.vmware.vim.eam.issue.vibNotInstalled|VIB module for agent is not installed on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.vibRequiresHostInMaintenanceMode | error | com.vmware.vim.eam.issue.vibRequiresHostInMaintenanceMode|Host must be put into maintenance mode to complete agent VIB installation ({agencyName}) |
com.vmware.vim.eam.issue.vibRequiresHostReboot | error | com.vmware.vim.eam.issue.vibRequiresHostReboot|Host {host.name} must be reboot to complete agent VIB installation ({agencyName}) |
com.vmware.vim.eam.issue.vibRequiresManualInstallation | error | com.vmware.vim.eam.issue.vibRequiresManualInstallation|VIB {vib} requires manual installation on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.vibRequiresManualUninstallation | error | com.vmware.vim.eam.issue.vibRequiresManualUninstallation|VIB {vib} requires manual uninstallation on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.vmCorrupted | error | com.vmware.vim.eam.issue.vmCorrupted|Agent VM {vm.name} on host {host.name} is corrupted ({agencyName}) |
com.vmware.vim.eam.issue.vmDeployed | error | com.vmware.vim.eam.issue.vmDeployed|Agent VM {vm.name} is provisioned on host {host.name} when it should be removed ({agencyName}) |
com.vmware.vim.eam.issue.vmMarkedAsTemplate | error | com.vmware.vim.eam.issue.vmMarkedAsTemplate|Agent VM {vm.name} on host {host.name} is marked as template ({agencyName}) |
com.vmware.vim.eam.issue.vmNotDeployed | error | com.vmware.vim.eam.issue.vmNotDeployed|Agent VM is missing on host {host.name} ({agencyName}) |
com.vmware.vim.eam.issue.vmOrphaned | error | com.vmware.vim.eam.issue.vmOrphaned|Orphaned agent VM {vm.name} on host {host.name} detected ({agencyName}) |
com.vmware.vim.eam.issue.vmPoweredOff | error | com.vmware.vim.eam.issue.vmPoweredOff|Agent VM {vm.name} on host {host.name} is expected to be powered on ({agencyName}) |
com.vmware.vim.eam.issue.vmPoweredOn | error | com.vmware.vim.eam.issue.vmPoweredOn|Agent VM {vm.name} on host {host.name} is expected to be powered off ({agencyName}) |
com.vmware.vim.eam.issue.vmSuspended | error | com.vmware.vim.eam.issue.vmSuspended|Agent VM {vm.name} on host {host.name} is expected to be powered on but is suspended ({agencyName}) |
com.vmware.vim.eam.issue.vmWrongFolder | error | com.vmware.vim.eam.issue.vmWrongFolder|Agent VM {vm.name} on host {host.name} is in the wrong VM folder ({agencyName}) |
com.vmware.vim.eam.issue.vmWrongResourcePool | error | com.vmware.vim.eam.issue.vmWrongResourcePool|Agent VM {vm.name} on host {host.name} is in the resource pool ({agencyName}) |
com.vmware.vim.eam.login.invalid | warning | com.vmware.vim.eam.login.invalid|Failed login to vSphere ESX Agent Manager |
com.vmware.vim.eam.login.succeeded | info | com.vmware.vim.eam.login.succeeded|Successful login by {user} into vSphere ESX Agent Manager |
com.vmware.vim.eam.logout | info | com.vmware.vim.eam.logout|User {user} logged out of vSphere ESX Agent Manager by logging out of the vCenter server |
com.vmware.vim.eam.task.scanForUnknownAgentVmsCompleted | info | com.vmware.vim.eam.task.scanForUnknownAgentVmsCompleted|Scan for unknown agent VMs completed |
com.vmware.vim.eam.task.scanForUnknownAgentVmsInitiated | info | com.vmware.vim.eam.task.scanForUnknownAgentVmsInitiated|Scan for unknown agent VMs initiated |
com.vmware.vim.eam.task.setupDvFilter | info | com.vmware.vim.eam.task.setupDvFilter|DvFilter switch '{switchName}' is setup on host {host.name} |
com.vmware.vim.eam.task.tearDownDvFilter | info | com.vmware.vim.eam.task.tearDownDvFilter|DvFilter switch '{switchName}' is teared down on host {host.name} |
com.vmware.vim.eam.unauthorized.access | warning | com.vmware.vim.eam.unauthorized.access|Unauthorized access by {user} in vSphere ESX Agent Manager |
com.vmware.vim.eam.vum.failedtouploadvib | error | com.vmware.vim.eam.vum.failedtouploadvib|Failed to upload {vibUrl} to VMware Update Manager ({agencyName}) |
com.vmware.vim.vsm.dependency.bind.vApp | | com.vmware.vim.vsm.dependency.bind.vApp|event.com.vmware.vim.vsm.dependency.bind.vApp.fullFormat |
com.vmware.vim.vsm.dependency.bind.vm | | com.vmware.vim.vsm.dependency.bind.vm|event.com.vmware.vim.vsm.dependency.bind.vm.fullFormat |
com.vmware.vim.vsm.dependency.create.vApp | | com.vmware.vim.vsm.dependency.create.vApp|event.com.vmware.vim.vsm.dependency.create.vApp.fullFormat |
com.vmware.vim.vsm.dependency.create.vm | | com.vmware.vim.vsm.dependency.create.vm|event.com.vmware.vim.vsm.dependency.create.vm.fullFormat |
com.vmware.vim.vsm.dependency.destroy.vApp | | com.vmware.vim.vsm.dependency.destroy.vApp|event.com.vmware.vim.vsm.dependency.destroy.vApp.fullFormat |
com.vmware.vim.vsm.dependency.destroy.vm | | com.vmware.vim.vsm.dependency.destroy.vm|event.com.vmware.vim.vsm.dependency.destroy.vm.fullFormat |
com.vmware.vim.vsm.dependency.reconfigure.vApp | | com.vmware.vim.vsm.dependency.reconfigure.vApp|event.com.vmware.vim.vsm.dependency.reconfigure.vApp.fullFormat |
com.vmware.vim.vsm.dependency.reconfigure.vm | | com.vmware.vim.vsm.dependency.reconfigure.vm|event.com.vmware.vim.vsm.dependency.reconfigure.vm.fullFormat |
com.vmware.vim.vsm.dependency.unbind.vApp | | com.vmware.vim.vsm.dependency.unbind.vApp|event.com.vmware.vim.vsm.dependency.unbind.vApp.fullFormat |
com.vmware.vim.vsm.dependency.unbind.vm | | com.vmware.vim.vsm.dependency.unbind.vm|event.com.vmware.vim.vsm.dependency.unbind.vm.fullFormat |
com.vmware.vim.vsm.dependency.update.vApp | | com.vmware.vim.vsm.dependency.update.vApp|event.com.vmware.vim.vsm.dependency.update.vApp.fullFormat |
com.vmware.vim.vsm.dependency.update.vm | | com.vmware.vim.vsm.dependency.update.vm|event.com.vmware.vim.vsm.dependency.update.vm.fullFormat |
com.vmware.vim.vsm.provider.register | | com.vmware.vim.vsm.provider.register|event.com.vmware.vim.vsm.provider.register.fullFormat |
com.vmware.vim.vsm.provider.unregister | | com.vmware.vim.vsm.provider.unregister|event.com.vmware.vim.vsm.provider.unregister.fullFormat |
com.vmware.vim.vsm.provider.update | | com.vmware.vim.vsm.provider.update|event.com.vmware.vim.vsm.provider.update.fullFormat |
esx.audit.dcui.defaults.factoryrestore | warning | esx.audit.dcui.defaults.factoryrestore|The host has been restored to default factory settings. Please consult ESXi Embedded and vCenter Server Setup Guide or follow the Ask VMware link for more information. |
esx.audit.dcui.disabled | info | esx.audit.dcui.disabled|The DCUI has been disabled. |
esx.audit.dcui.enabled | info | esx.audit.dcui.enabled|The DCUI has been enabled. |
esx.audit.dcui.host.reboot | warning | esx.audit.dcui.host.reboot|The host is being rebooted through the Direct Console User Interface (DCUI). Please consult ESXi Embedded and vCenter Server Setup Guide or follow the Ask VMware link for more information. |
esx.audit.dcui.host.shutdown | warning | esx.audit.dcui.host.shutdown|The host is being shut down through the Direct Console User Interface (DCUI). Please consult ESXi Embedded and vCenter Server Setup Guide or follow the Ask VMware link for more information. |
esx.audit.dcui.hostagents.restart | info | esx.audit.dcui.hostagents.restart|The management agents on the host are being restarted. Please consult ESXi Embedded and vCenter Server Setup Guide or follow the Ask VMware link for more information. |
esx.audit.dcui.login.failed | error | esx.audit.dcui.login.failed|Authentication of user {1} has failed. Please consult ESXi Embedded and vCenter Server Setup Guide or follow the Ask VMware link for more information. |
esx.audit.dcui.login.passwd.changed | info | esx.audit.dcui.login.passwd.changed|Login password for user {1} has been changed. Please consult ESXi Embedded and vCenter Server Setup Guide or follow the Ask VMware link for more information. |
esx.audit.dcui.network.factoryrestore | warning | esx.audit.dcui.network.factoryrestore|The host has been restored to factory network settings. Please consult ESXi Embedded and vCenter Server Setup Guide or follow the Ask VMware link for more information. |
esx.audit.dcui.network.restart | info | esx.audit.dcui.network.restart|A management interface {1} has been restarted. Please consult ESXi Embedded and vCenter Server Setup Guide or follow the Ask VMware link for more information. |
esx.audit.esxcli.host.poweroff | warning | esx.audit.esxcli.host.poweroff|The host is being powered off through esxcli. Reason for powering off: {1}. Please consult vSphere Documentation Center or follow the Ask VMware link for more information. |
esx.audit.esxcli.host.restart | | esx.audit.esxcli.host.restart|event.esx.audit.esxcli.host.restart.fullFormat |
esx.audit.esximage.hostacceptance.changed | info | esx.audit.esximage.hostacceptance.changed|Host acceptance level changed from {1} to {2} |
esx.audit.esximage.install.novalidation | warning | esx.audit.esximage.install.novalidation|Attempting to install an image profile with validation disabled. This may result in an image with unsatisfied dependencies, file or package conflicts, and potential security violations. |
esx.audit.esximage.install.securityalert | warning | esx.audit.esximage.install.securityalert|SECURITY ALERT: Installing image profile '{1}' with {2}. |
esx.audit.esximage.profile.install.successful | info | esx.audit.esximage.profile.install.successful|Successfully installed image profile '{1}'. Installed {2} VIB(s), removed {3} VIB(s). Please use 'esxcli software profile get' or see log for more detail about the transaction. |
esx.audit.esximage.profile.update.successful | info | esx.audit.esximage.profile.update.successful|Successfully updated host to image profile '{1}'. Installed {2} VIB(s), removed {3} VIB(s). Please use 'esxcli software profile get' or see log for more detail about the transaction. |
esx.audit.esximage.vib.install.successful | info | esx.audit.esximage.vib.install.successful|Successfully installed {1} VIB(s), removed {2} VIB(s). Please use 'esxcli software profile get' or see log for more detail about the transaction. |
esx.audit.esximage.vib.remove.successful | info | esx.audit.esximage.vib.remove.successful|Successfully removed {1} VIB(s). Please use 'esxcli software profile get' or see log for more detail about the transaction. |
esx.audit.host.boot | info | esx.audit.host.boot|Host has booted. |
esx.audit.host.maxRegisteredVMsExceeded | warning | esx.audit.host.maxRegisteredVMsExceeded|The number of virtual machines registered on host {host.name} in cluster {computeResource.name} in {datacenter.name} exceeded limit: {current} registered, {limit} is the maximum supported. |
esx.audit.host.stop.reboot | info | esx.audit.host.stop.reboot|Host is rebooting. |
esx.audit.host.stop.shutdown | info | esx.audit.host.stop.shutdown|Host is shutting down. |
esx.audit.lockdownmode.disabled | info | esx.audit.lockdownmode.disabled|Administrator access to the host has been enabled. |
esx.audit.lockdownmode.enabled | info | esx.audit.lockdownmode.enabled|Administrator access to the host has been disabled. |
esx.audit.maintenancemode.canceled | info | esx.audit.maintenancemode.canceled|The host has canceled entering maintenance mode. |
esx.audit.maintenancemode.entered | info | esx.audit.maintenancemode.entered|The host has entered maintenance mode. |
esx.audit.maintenancemode.entering | info | esx.audit.maintenancemode.entering|The host has begun entering maintenance mode. |
esx.audit.maintenancemode.exited | info | esx.audit.maintenancemode.exited|The host has exited maintenance mode. |
esx.audit.net.firewall.config.changed | info | esx.audit.net.firewall.config.changed|Firewall configuration has changed. Operation '{1}' for rule set {2} succeeded. |
esx.audit.net.firewall.disabled | warning | esx.audit.net.firewall.disabled|Firewall has been disabled. |
esx.audit.net.firewall.enabled | info | esx.audit.net.firewall.enabled|Firewall has been enabled for port {1}. |
esx.audit.net.firewall.port.hooked | info | esx.audit.net.firewall.port.hooked|Port {1} is now protected by Firewall. |
esx.audit.net.firewall.port.removed | warning | esx.audit.net.firewall.port.removed|Port {1} is no longer protected with Firewall. |
esx.audit.net.lacp.disable | info | esx.audit.net.lacp.disable|LACP for VDS {1} is disabled. |
esx.audit.net.lacp.enable | info | esx.audit.net.lacp.enable|LACP for VDS {1} is enabled. |
esx.audit.net.lacp.uplink.connected | info | esx.audit.net.lacp.uplink.connected|LACP info: uplink {1} on VDS {2} got connected. |
esx.audit.shell.disabled | info | esx.audit.shell.disabled|The ESXi command line shell has been disabled. |
esx.audit.shell.enabled | info | esx.audit.shell.enabled|The ESXi command line shell has been enabled. |
esx.audit.ssh.disabled | info | esx.audit.ssh.disabled|SSH access has been disabled. |
esx.audit.ssh.enabled | info | esx.audit.ssh.enabled|SSH access has been enabled. |
esx.audit.usb.config.changed | info | esx.audit.usb.config.changed|USB configuration has changed on host {host.name} in cluster {computeResource.name} in {datacenter.name}. |
esx.audit.uw.secpolicy.alldomains.level.changed | warning | esx.audit.uw.secpolicy.alldomains.level.changed|The enforcement level for all security domains has been changed to {1}. The enforcement level must always be set to enforcing. |
esx.audit.uw.secpolicy.domain.level.changed | warning | esx.audit.uw.secpolicy.domain.level.changed|The enforcement level for security domain {1} has been changed to {2}. The enforcement level must always be set to enforcing. |
esx.audit.vmfs.lvm.device.discovered | info | esx.audit.vmfs.lvm.device.discovered|One or more LVM devices have been discovered on this host. |
esx.audit.vmfs.volume.mounted | info | esx.audit.vmfs.volume.mounted|File system {1} on volume {2} has been mounted in {3} mode on this host. |
esx.audit.vmfs.volume.umounted | info | esx.audit.vmfs.volume.umounted|The volume {1} has been safely un-mounted. The datastore is no longer accessible on this host. |
esx.audit.vsan.clustering.enabled | info | esx.audit.vsan.clustering.enabled|VSAN clustering and directory services have been enabled. |
esx.clear.coredump.configured | info | esx.clear.coredump.configured|A vmkcore disk partition is available and/or a network coredump server has been configured. Host core dumps will be saved. |
esx.clear.net.connectivity.restored | info | esx.clear.net.connectivity.restored|Network connectivity restored on virtual switch {1}, portgroups: {2}. Physical NIC {3} is up. |
esx.clear.net.dvport.connectivity.restored | info | esx.clear.net.dvport.connectivity.restored|Network connectivity restored on DVPorts: {1}. Physical NIC {2} is up. |
esx.clear.net.dvport.redundancy.restored | info | esx.clear.net.dvport.redundancy.restored|Uplink redundancy restored on DVPorts: {1}. Physical NIC {2} is up. |
esx.clear.net.lacp.lag.transition.up | info | esx.clear.net.lacp.lag.transition.up|LACP info: LAG {1} on VDS {2} is up. |
esx.clear.net.lacp.uplink.transition.up | info | esx.clear.net.lacp.uplink.transition.up|LACP info: uplink {1} on VDS {2} is moved into link aggregation group. |
esx.clear.net.lacp.uplink.unblocked | info | esx.clear.net.lacp.uplink.unblocked|LACP info: uplink {1} on VDS {2} is unblocked. |
esx.clear.net.redundancy.restored | info | esx.clear.net.redundancy.restored|Uplink redundancy restored on virtual switch {1}, portgroups: {2}. Physical NIC {3} is up. |
esx.clear.net.vmnic.linkstate.up | info | esx.clear.net.vmnic.linkstate.up|Physical NIC {1} linkstate is up. |
esx.clear.scsi.device.io.latency.improved | info | esx.clear.scsi.device.io.latency.improved|Device {1} performance has improved. I/O latency reduced from {2} microseconds to {3} microseconds. |
esx.clear.scsi.device.state.on | info | esx.clear.scsi.device.state.on|Device {1}, has been turned on administratively. |
esx.clear.scsi.device.state.permanentloss.deviceonline | info | esx.clear.scsi.device.state.permanentloss.deviceonline|Device {1}, that was permanently inaccessible is now online. No data consistency guarantees. |
esx.clear.storage.apd.exit | info | esx.clear.storage.apd.exit|Device or filesystem with identifier {1} has exited the All Paths Down state. |
esx.clear.storage.connectivity.restored | info | esx.clear.storage.connectivity.restored|Connectivity to storage device {1} (Datastores: {2}) restored. Path {3} is active again. |
esx.clear.storage.redundancy.restored | info | esx.clear.storage.redundancy.restored|Path redundancy to storage device {1} (Datastores: {2}) restored. Path {3} is active again. |
esx.clear.vsan.clustering.enabled | info | esx.clear.vsan.clustering.enabled|VSAN clustering and directory services have now been enabled. |
esx.clear.vsan.network.available | | esx.clear.vsan.network.available|event.esx.clear.vsan.network.available.fullFormat |
esx.clear.vsan.vmknic.ready | | esx.clear.vsan.vmknic.ready|event.esx.clear.vsan.vmknic.ready.fullFormat |
esx.problem.3rdParty.error | error | esx.problem.3rdParty.error|A 3rd party component, {1}, running on ESXi has reported an error. Please follow the knowledge base link ({2}) to see the steps to remedy the problem as reported by {3}. The message reported is: {4}. |
esx.problem.3rdParty.info | | esx.problem.3rdParty.info|event.esx.problem.3rdParty.info.fullFormat |
esx.problem.3rdParty.warning | warning | esx.problem.3rdParty.warning|A 3rd party component, {1}, running on ESXi has reported a warning related to a problem. Please follow the knowledge base link ({2}) to see the steps to remedy the problem as reported by {3}. The message reported is: {4}. |
esx.problem.apei.bert.memory.error.corrected | error | esx.problem.apei.bert.memory.error.corrected|A corrected memory error occurred in last boot. The following details were reported. Physical Addr: {1}, Physical Addr Mask: {2}, Node: {3}, Card: {4}, Module: {5}, Bank: {6}, Device: {7}, Row: {8}, Column: {9} Error type: {10} |
esx.problem.apei.bert.memory.error.fatal | error | esx.problem.apei.bert.memory.error.fatal|A fatal memory error occurred in the last boot. The following details were reported. Physical Addr: {1}, Physical Addr Mask: {2}, Node: {3}, Card: {4}, Module: {5}, Bank: {6}, Device: {7}, Row: {8}, Column: {9} Error type: {10} |
esx.problem.apei.bert.memory.error.recoverable | error | esx.problem.apei.bert.memory.error.recoverable|A recoverable memory error occurred in last boot. The following details were reported. Physical Addr: {1}, Physical Addr Mask: {2}, Node: {3}, Card: {4}, Module: {5}, Bank: {6}, Device: {7}, Row: {8}, Column: {9} Error type: {10} |
esx.problem.apei.bert.pcie.error.corrected | error | esx.problem.apei.bert.pcie.error.corrected|A corrected PCIe error occurred in last boot. The following details were reported. Port Type: {1}, Device: {2}, Bus #: {3}, Function: {4}, Slot: {5}, Device Vendor: {6}, Version: {7}, Command Register: {8}, Status Register: {9}. |
esx.problem.apei.bert.pcie.error.fatal | error | esx.problem.apei.bert.pcie.error.fatal|Platform encounterd a fatal PCIe error in last boot. The following details were reported. Port Type: {1}, Device: {2}, Bus #: {3}, Function: {4}, Slot: {5}, Device Vendor: {6}, Version: {7}, Command Register: {8}, Status Register: {9}. |
esx.problem.apei.bert.pcie.error.recoverable | error | esx.problem.apei.bert.pcie.error.recoverable|A recoverable PCIe error occurred in last boot. The following details were reported. Port Type: {1}, Device: {2}, Bus #: {3}, Function: {4}, Slot: {5}, Device Vendor: {6}, Version: {7}, Command Register: {8}, Status Register: {9}. |
esx.problem.application.core.dumped | warning | esx.problem.application.core.dumped|An application ({1}) running on ESXi host has crashed ({2} time(s) so far). A core file might have been created at {3}. |
esx.problem.coredump.unconfigured | warning | esx.problem.coredump.unconfigured|No vmkcore disk partition is available and no network coredump server has been configured. Host core dumps cannot be saved. |
esx.problem.cpu.amd.mce.dram.disabled | error | esx.problem.cpu.amd.mce.dram.disabled|DRAM ECC not enabled. Please enable it in BIOS. |
esx.problem.cpu.intel.ioapic.listing.error | error | esx.problem.cpu.intel.ioapic.listing.error|Not all IO-APICs are listed in the DMAR. Not enabling interrupt remapping on this platform. |
esx.problem.cpu.mce.invalid | error | esx.problem.cpu.mce.invalid|MCE monitoring will be disabled as an unsupported CPU was detected. Please consult the ESX HCL for information on supported hardware. |
esx.problem.cpu.smp.ht.invalid | error | esx.problem.cpu.smp.ht.invalid|Disabling HyperThreading due to invalid configuration: Number of threads: {1}, Number of PCPUs: {2}. |
esx.problem.cpu.smp.ht.numpcpus.max | error | esx.problem.cpu.smp.ht.numpcpus.max|Found {1} PCPUs, but only using {2} of them due to specified limit. |
esx.problem.cpu.smp.ht.partner.missing | error | esx.problem.cpu.smp.ht.partner.missing|Disabling HyperThreading due to invalid configuration: HT partner {1} is missing from PCPU {2}. |
esx.problem.dhclient.lease.none | error | esx.problem.dhclient.lease.none|Unable to obtain a DHCP lease on interface {1}. |
esx.problem.dhclient.lease.offered.error | | esx.problem.dhclient.lease.offered.error|event.esx.problem.dhclient.lease.offered.error.fullFormat |
esx.problem.esximage.install.error | error | esx.problem.esximage.install.error|Could not install image profile: {1} |
esx.problem.esximage.install.invalidhardware | error | esx.problem.esximage.install.invalidhardware|Host doesn't meet image profile '{1}' hardware requirements: {2} |
esx.problem.esximage.install.stage.error | error | esx.problem.esximage.install.stage.error|Could not stage image profile '{1}': {2} |
esx.problem.hardware.acpi.interrupt.routing.device.invalid | error | esx.problem.hardware.acpi.interrupt.routing.device.invalid|Skipping interrupt routing entry with bad device number: {1}. This is a BIOS bug. |
esx.problem.hardware.acpi.interrupt.routing.pin.invalid | error | esx.problem.hardware.acpi.interrupt.routing.pin.invalid|Skipping interrupt routing entry with bad device pin: {1}. This is a BIOS bug. |
esx.problem.hardware.ioapic.missing | error | esx.problem.hardware.ioapic.missing|IOAPIC Num {1} is missing. Please check BIOS settings to enable this IOAPIC. |
esx.problem.host.coredump | warning | esx.problem.host.coredump|An unread host kernel core dump has been found. |
esx.problem.hostd.core.dumped | warning | esx.problem.hostd.core.dumped|{1} crashed ({2} time(s) so far) and a core file might have been created at {3}. This might have caused connections to the host to be dropped. |
esx.problem.iorm.badversion | info | esx.problem.iorm.badversion|Host {1} cannot participate in Storage I/O Control(SIOC) on datastore {2} because the version number {3} of the SIOC agent on this host is incompatible with number {4} of its counterparts on other hosts connected to this datastore. |
esx.problem.iorm.nonviworkload | info | esx.problem.iorm.nonviworkload|An unmanaged I/O workload is detected on a SIOC-enabled datastore: {1}. |
esx.problem.migrate.vmotion.default.heap.create.failed | warning | esx.problem.migrate.vmotion.default.heap.create.failed|Failed to create default migration heap. This might be the result of severe host memory pressure or virtual address space exhaustion. Migration might still be possible, but will be unreliable in cases of extreme host memory pressure. |
esx.problem.migrate.vmotion.server.pending.cnx.listen.socket.shutdown | error | esx.problem.migrate.vmotion.server.pending.cnx.listen.socket.shutdown|The ESXi host's vMotion network server encountered an error while monitoring incoming network connections. Shutting down listener socket. vMotion might not be possible with this host until vMotion is manually re-enabled. Failure status: {1} |
esx.problem.net.connectivity.lost | error | esx.problem.net.connectivity.lost|Lost network connectivity on virtual switch {1}. Physical NIC {2} is down. Affected portgroups:{3}. |
esx.problem.net.dvport.connectivity.lost | error | esx.problem.net.dvport.connectivity.lost|Lost network connectivity on DVPorts: {1}. Physical NIC {2} is down. |
esx.problem.net.dvport.redundancy.degraded | warning | esx.problem.net.dvport.redundancy.degraded|Uplink redundancy degraded on DVPorts: {1}. Physical NIC {2} is down. |
esx.problem.net.dvport.redundancy.lost | warning | esx.problem.net.dvport.redundancy.lost|Lost uplink redundancy on DVPorts: {1}. Physical NIC {2} is down. |
esx.problem.net.e1000.tso6.notsupported | error | esx.problem.net.e1000.tso6.notsupported|Guest-initiated IPv6 TCP Segmentation Offload (TSO) packets ignored. Manually disable TSO inside the guest operating system in virtual machine {1}, or use a different virtual adapter. |
esx.problem.net.fence.port.badfenceid | error | esx.problem.net.fence.port.badfenceid|VMkernel failed to set fenceId {1} on distributed virtual port {2} on switch {3}. Reason: invalid fenceId. |
esx.problem.net.fence.resource.limited | error | esx.problem.net.fence.resource.limited|Vmkernel failed to set fenceId {1} on distributed virtual port {2} on switch {3}. Reason: maximum number of fence networks or ports have been reached. |
esx.problem.net.fence.switch.unavailable | error | esx.problem.net.fence.switch.unavailable|Vmkernel failed to set fenceId {1} on distributed virtual port {2} on switch {3}. Reason: dvSwitch fence property is not set. |
esx.problem.net.firewall.config.failed | error | esx.problem.net.firewall.config.failed|Firewall configuration operation '{1}' failed. The changes were not applied to rule set {2}. |
esx.problem.net.firewall.port.hookfailed | error | esx.problem.net.firewall.port.hookfailed|Adding port {1} to Firewall failed. |
esx.problem.net.gateway.set.failed | error | esx.problem.net.gateway.set.failed|Cannot connect to the specified gateway {1}. Failed to set it. |
esx.problem.net.heap.belowthreshold | warning | esx.problem.net.heap.belowthreshold|{1} free size dropped below {2} percent. |
esx.problem.net.lacp.lag.transition.down | warning | esx.problem.net.lacp.lag.transition.down|LACP warning: LAG {1} on VDS {2} is down. |
esx.problem.net.lacp.peer.noresponse | error | esx.problem.net.lacp.peer.noresponse|LACP error: No peer response on uplink {1} for VDS {2}. |
esx.problem.net.lacp.policy.incompatible | error | esx.problem.net.lacp.policy.incompatible|LACP error: Current teaming policy on VDS {1} is incompatible, supported is IP hash only. |
esx.problem.net.lacp.policy.linkstatus | error | esx.problem.net.lacp.policy.linkstatus|LACP error: Current teaming policy on VDS {1} is incompatible, supported link failover detection is link status only. |
esx.problem.net.lacp.uplink.blocked | warning | esx.problem.net.lacp.uplink.blocked|LACP warning: uplink {1} on VDS {2} is blocked. |
esx.problem.net.lacp.uplink.disconnected | warning | esx.problem.net.lacp.uplink.disconnected|LACP warning: uplink {1} on VDS {2} got disconnected. |
esx.problem.net.lacp.uplink.fail.duplex | error | esx.problem.net.lacp.uplink.fail.duplex|LACP error: Duplex mode across all uplink ports must be full, VDS {1} uplink {2} has different mode. |
esx.problem.net.lacp.uplink.fail.speed | error | esx.problem.net.lacp.uplink.fail.speed|LACP error: Speed across all uplink ports must be same, VDS {1} uplink {2} has different speed. |
esx.problem.net.lacp.uplink.inactive | error | esx.problem.net.lacp.uplink.inactive|LACP error: All uplinks on VDS {1} must be active. |
esx.problem.net.lacp.uplink.transition.down | warning | esx.problem.net.lacp.uplink.transition.down|LACP warning: uplink {1} on VDS {2} is moved out of link aggregation group. |
esx.problem.net.migrate.bindtovmk | warning | esx.problem.net.migrate.bindtovmk|The ESX advanced configuration option /Migrate/Vmknic is set to an invalid vmknic: {1}. /Migrate/Vmknic specifies a vmknic that vMotion binds to for improved performance. Update the configuration option with a valid vmknic. Alternatively, if you do not want vMotion to bind to a specific vmknic, remove the invalid vmknic and leave the option blank. |
esx.problem.net.migrate.unsupported.latency | warning | esx.problem.net.migrate.unsupported.latency|ESXi has detected {1}ms round-trip vMotion network latency between host {2} and {3}. High latency vMotion networks are supported only if both ESXi hosts have been configured for vMotion latency tolerance. |
esx.problem.net.portset.port.full | error | esx.problem.net.portset.port.full|Portset {1} has reached the maximum number of ports ({2}). Cannot apply for any more free ports. |
esx.problem.net.portset.port.vlan.invalidid | error | esx.problem.net.portset.port.vlan.invalidid|{1} VLANID {2} is invalid. VLAN ID must be between 0 and 4095. |
esx.problem.net.proxyswitch.port.unavailable | warning | esx.problem.net.proxyswitch.port.unavailable|Virtual NIC with hardware address {1} failed to connect to distributed virtual port {2} on switch {3}. There are no more ports available on the host proxy switch. |
esx.problem.net.redundancy.degraded | warning | esx.problem.net.redundancy.degraded|Uplink redundancy degraded on virtual switch {1}. Physical NIC {2} is down. Affected portgroups:{3}. |
esx.problem.net.redundancy.lost | warning | esx.problem.net.redundancy.lost|Lost uplink redundancy on virtual switch {1}. Physical NIC {2} is down. Affected portgroups:{3}. |
esx.problem.net.uplink.mtu.failed | warning | esx.problem.net.uplink.mtu.failed|VMkernel failed to set the MTU value {1} on the uplink {2}. |
esx.problem.net.vmknic.ip.duplicate | warning | esx.problem.net.vmknic.ip.duplicate|A duplicate IP address was detected for {1} on the interface {2}. The current owner is {3}. |
esx.problem.net.vmnic.linkstate.down | warning | esx.problem.net.vmnic.linkstate.down|Physical NIC {1} linkstate is down. |
esx.problem.net.vmnic.linkstate.flapping | warning | esx.problem.net.vmnic.linkstate.flapping|Taking down physical NIC {1} because the link is unstable. |
esx.problem.net.vmnic.watchdog.reset | warning | esx.problem.net.vmnic.watchdog.reset|Uplink {1} has recovered from a transient failure due to watchdog timeout |
esx.problem.ntpd.clock.correction.error | error | esx.problem.ntpd.clock.correction.error|NTP daemon stopped. Time correction {1} > {2} seconds. Manually set the time and restart ntpd. |
esx.problem.pageretire.platform.retire.request | info | esx.problem.pageretire.platform.retire.request|Memory page retirement requested by platform firmware. FRU ID: {1}. Refer to System Hardware Log: {2} |
esx.problem.pageretire.selectedmpnthreshold.host.exceeded | warning | esx.problem.pageretire.selectedmpnthreshold.host.exceeded|Number of host physical memory pages that have been selected for retirement ({1}) exceeds threshold ({2}). |
esx.problem.scsi.apd.event.descriptor.alloc.failed | warning | esx.problem.scsi.apd.event.descriptor.alloc.failed|No memory to allocate APD (All Paths Down) event subsystem. |
esx.problem.scsi.device.close.failed | warning | esx.problem.scsi.device.close.failed|"Failed to close the device {1} properly, plugin {2}. |
esx.problem.scsi.device.detach.failed | warning | esx.problem.scsi.device.detach.failed|Detach failed for device :{1}. Exceeded the number of devices that can be detached, please cleanup stale detach entries. |
esx.problem.scsi.device.filter.attach.failed | warning | esx.problem.scsi.device.filter.attach.failed|Failed to attach filters to device '%s' during registration. Plugin load failed or the filter rules are incorrect. |
esx.problem.scsi.device.io.bad.plugin.type | warning | esx.problem.scsi.device.io.bad.plugin.type|Bad plugin type for device {1}, plugin {2} |
esx.problem.scsi.device.io.inquiry.failed | warning | esx.problem.scsi.device.io.inquiry.failed|Failed to get standard inquiry for device {1} from Plugin {2}. |
esx.problem.scsi.device.io.invalid.disk.qfull.value | warning | esx.problem.scsi.device.io.invalid.disk.qfull.value|QFullSampleSize should be bigger than QFullThreshold. LUN queue depth throttling algorithm will not function as expected. Please set the QFullSampleSize and QFullThreshold disk configuration values in ESX correctly. |
esx.problem.scsi.device.io.latency.high | warning | esx.problem.scsi.device.io.latency.high|Device {1} performance has deteriorated. I/O latency increased from average value of {2} microseconds to {3} microseconds. |
esx.problem.scsi.device.io.qerr.change.config | warning | esx.problem.scsi.device.io.qerr.change.config|QErr set to 0x{1} for device {2}. This may cause unexpected behavior. The system is not configured to change the QErr setting of device. The QErr value supported by system is 0x{3}. Please check the SCSI ChangeQErrSetting configuration value for ESX. |
esx.problem.scsi.device.io.qerr.changed | warning | esx.problem.scsi.device.io.qerr.changed|QErr set to 0x{1} for device {2}. This may cause unexpected behavior. The device was originally configured to the supported QErr setting of 0x{3}, but this has been changed and could not be changed back. |
esx.problem.scsi.device.is.local.failed | warning | esx.problem.scsi.device.is.local.failed|Failed to verify if the device {1} from plugin {2} is a local - not shared - device |
esx.problem.scsi.device.is.pseudo.failed | warning | esx.problem.scsi.device.is.pseudo.failed|Failed to verify if the device {1} from plugin {2} is a pseudo device |
esx.problem.scsi.device.is.ssd.failed | warning | esx.problem.scsi.device.is.ssd.failed|Failed to verify if the device {1} from plugin {2} is a Solid State Disk device |
esx.problem.scsi.device.limitreached | error | esx.problem.scsi.device.limitreached|The maximum number of supported devices of {1} has been reached. A device from plugin {2} could not be created. |
esx.problem.scsi.device.state.off | info | esx.problem.scsi.device.state.off|Device {1}, has been turned off administratively. |
esx.problem.scsi.device.state.permanentloss | error | esx.problem.scsi.device.state.permanentloss|Device {1} has been removed or is permanently inaccessible. Affected datastores (if any): {2}. |
esx.problem.scsi.device.state.permanentloss.noopens | info | esx.problem.scsi.device.state.permanentloss.noopens|Permanently inaccessible device {1} has no more opens. It is now safe to unmount datastores (if any) {2} and delete the device. |
esx.problem.scsi.device.state.permanentloss.pluggedback | error | esx.problem.scsi.device.state.permanentloss.pluggedback|Device {1} has been plugged back in after being marked permanently inaccessible. No data consistency guarantees. |
esx.problem.scsi.device.state.permanentloss.withreservationheld | error | esx.problem.scsi.device.state.permanentloss.withreservationheld|Device {1} has been removed or is permanently inaccessible, while holding a reservation. Affected datastores (if any): {2}. |
esx.problem.scsi.device.thinprov.atquota | warning | esx.problem.scsi.device.thinprov.atquota|Space utilization on thin-provisioned device {1} exceeded configured threshold. Affected datastores (if any): {2}. |
esx.problem.scsi.scsipath.limitreached | error | esx.problem.scsi.scsipath.limitreached|The maximum number of supported paths of {1} has been reached. Path {2} could not be added. |
esx.problem.scsi.unsupported.plugin.type | warning | esx.problem.scsi.unsupported.plugin.type|Scsi Device Allocation not supported for plugin type {1} |
esx.problem.storage.apd.start | warning | esx.problem.storage.apd.start|Device or filesystem with identifier {1} has entered the All Paths Down state. |
esx.problem.storage.apd.timeout | warning | esx.problem.storage.apd.timeout|Device or filesystem with identifier {1} has entered the All Paths Down Timeout state after being in the All Paths Down state for {2} seconds. I/Os will now be fast failed. |
esx.problem.storage.connectivity.devicepor | warning | esx.problem.storage.connectivity.devicepor|Frequent PowerOn Reset Unit Attentions are occurring on device {1}. This might indicate a storage problem. Affected datastores: {2} |
esx.problem.storage.connectivity.lost | error | esx.problem.storage.connectivity.lost|Lost connectivity to storage device {1}. Path {2} is down. Affected datastores: {3}. |
esx.problem.storage.connectivity.pathpor | warning | esx.problem.storage.connectivity.pathpor|Frequent PowerOn Reset Unit Attentions are occurring on path {1}. This might indicate a storage problem. Affected device: {2}. Affected datastores: {3} |
esx.problem.storage.connectivity.pathstatechanges | info | esx.problem.storage.connectivity.pathstatechanges|Frequent path state changes are occurring for path {1}. This might indicate a storage problem. Affected device: {2}. Affected datastores: {3} |
esx.problem.storage.iscsi.discovery.connect.error | error | esx.problem.storage.iscsi.discovery.connect.error|iSCSI discovery to {1} on {2} failed. The iSCSI Initiator could not establish a network connection to the discovery address. |
esx.problem.storage.iscsi.discovery.login.error | error | esx.problem.storage.iscsi.discovery.login.error|iSCSI discovery to {1} on {2} failed. The Discovery target returned a login error of: {3}. |
esx.problem.storage.iscsi.target.connect.error | error | esx.problem.storage.iscsi.target.connect.error|Login to iSCSI target {1} on {2} failed. The iSCSI initiator could not establish a network connection to the target. |
esx.problem.storage.iscsi.target.login.error | error | esx.problem.storage.iscsi.target.login.error|Login to iSCSI target {1} on {2} failed. Target returned login error of: {3}. |
esx.problem.storage.iscsi.target.permanently.lost | error | esx.problem.storage.iscsi.target.permanently.lost|The iSCSI target {2} was permanently removed from {1}. |
esx.problem.storage.redundancy.degraded | warning | esx.problem.storage.redundancy.degraded|Path redundancy to storage device {1} degraded. Path {2} is down. Affected datastores: {3}. |
esx.problem.storage.redundancy.lost | warning | esx.problem.storage.redundancy.lost|Lost path redundancy to storage device {1}. Path {2} is down. Affected datastores: {3}. |
esx.problem.syslog.config | warning | esx.problem.syslog.config|System logging is not configured on host {host.name}. Please check Syslog options for the host under Configuration -> Software -> Advanced Settings in vSphere client. |
esx.problem.syslog.nonpersistent | warning | esx.problem.syslog.nonpersistent|System logs on host {host.name} are stored on non-persistent storage. Consult product documentation to configure a syslog server or a scratch partition. |
esx.problem.vfat.filesystem.full.other | error | esx.problem.vfat.filesystem.full.other|The VFAT filesystem {1} (UUID {2}) is full. |
esx.problem.vfat.filesystem.full.scratch | error | esx.problem.vfat.filesystem.full.scratch|The host's scratch partition, which is the VFAT filesystem {1} (UUID {2}), is full. |
esx.problem.visorfs.failure | error | esx.problem.visorfs.failure|An operation on the root filesystem has failed. |
esx.problem.visorfs.inodetable.full | error | esx.problem.visorfs.inodetable.full|The root filesystem's file table is full. As a result, the file {1} could not be created by the application '{2}'. |
esx.problem.visorfs.ramdisk.full | error | esx.problem.visorfs.ramdisk.full|The ramdisk '{1}' is full. As a result, the file {2} could not be written. |
esx.problem.visorfs.ramdisk.inodetable.full | error | esx.problem.visorfs.ramdisk.inodetable.full|The file table of the ramdisk '{1}' is full. As a result, the file {2} could not be created by the application '{3}'. |
esx.problem.vm.kill.unexpected.fault.failure | error | esx.problem.vm.kill.unexpected.fault.failure|The VM using the config file {1} could not fault in a guest physical page from the hypervisor level swap file at {2}. The VM is terminated as further progress is impossible. |
esx.problem.vm.kill.unexpected.forcefulPageRetire | error | esx.problem.vm.kill.unexpected.forcefulPageRetire|The VM using the config file {1} contains the host physical page {2} which was scheduled for immediate retirement. To avoid system instability the VM is forcefully powered off. |
esx.problem.vm.kill.unexpected.noSwapResponse | error | esx.problem.vm.kill.unexpected.noSwapResponse|The VM using the config file {1} did not respond to {2} swap actions in {3} seconds and is forcefully powered off to prevent system instability. |
esx.problem.vm.kill.unexpected.vmtrack | error | esx.problem.vm.kill.unexpected.vmtrack|The VM using the config file {1} is allocating too many pages while system is critically low in free memory. It is forcefully terminated to prevent system instability. |
esx.problem.vmfs.ats.support.lost | | esx.problem.vmfs.ats.support.lost|event.esx.problem.vmfs.ats.support.lost.fullFormat |
esx.problem.vmfs.error.volume.is.locked | error | esx.problem.vmfs.error.volume.is.locked|Volume on device {1} is locked, possibly because some remote host encountered an error during a volume operation and could not recover. |
esx.problem.vmfs.extent.offline | error | esx.problem.vmfs.extent.offline|An attached device {1} may be offline. The file system {2} is now in a degraded state. While the datastore is still available, parts of data that reside on the extent that went offline might be inaccessible. |
esx.problem.vmfs.extent.online | info | esx.problem.vmfs.extent.online|Device {1} backing file system {2} came online. This extent was previously offline. All resources on this device are now available. |
esx.problem.vmfs.heartbeat.recovered | info | esx.problem.vmfs.heartbeat.recovered|Successfully restored access to volume {1} ({2}) following connectivity issues. |
esx.problem.vmfs.heartbeat.timedout | info | esx.problem.vmfs.heartbeat.timedout|Lost access to volume {1} ({2}) due to connectivity issues. Recovery attempt is in progress and outcome will be reported shortly. |
esx.problem.vmfs.heartbeat.unrecoverable | error | esx.problem.vmfs.heartbeat.unrecoverable|Lost connectivity to volume {1} ({2}) and subsequent recovery attempts have failed. |
esx.problem.vmfs.journal.createfailed | error | esx.problem.vmfs.journal.createfailed|No space for journal on volume {1} ({2}). Opening volume in read-only metadata mode with limited write support. |
esx.problem.vmfs.lock.corruptondisk | error | esx.problem.vmfs.lock.corruptondisk|At least one corrupt on-disk lock was detected on volume {1} ({2}). Other regions of the volume might be damaged too. |
esx.problem.vmfs.nfs.server.disconnect | error | esx.problem.vmfs.nfs.server.disconnect|Lost connection to server {1} mount point {2} mounted as {3} ({4}). |
esx.problem.vmfs.nfs.server.restored | info | esx.problem.vmfs.nfs.server.restored|Restored connection to server {1} mount point {2} mounted as {3} ({4}). |
esx.problem.vmfs.resource.corruptondisk | error | esx.problem.vmfs.resource.corruptondisk|At least one corrupt resource metadata region was detected on volume {1} ({2}). Other regions of the volume might be damaged too. |
esx.problem.vmsyslogd.remote.failure | error | esx.problem.vmsyslogd.remote.failure|The host "{1}" has become unreachable. Remote logging to this host has stopped. |
esx.problem.vmsyslogd.storage.failure | error | esx.problem.vmsyslogd.storage.failure|Logging to storage has failed. Logs are no longer being stored locally on this host. |
esx.problem.vmsyslogd.storage.logdir.invalid | error | esx.problem.vmsyslogd.storage.logdir.invalid|The configured log directory {1} cannot be used. The default directory {2} will be used instead. |
esx.problem.vmsyslogd.unexpected | error | esx.problem.vmsyslogd.unexpected|Log daemon has failed for an unexpected reason: {1} |
esx.problem.vpxa.core.dumped | warning | esx.problem.vpxa.core.dumped|{1} crashed ({2} time(s) so far) and a core file might have been created at {3}. This might have caused connections to the host to be dropped. |
esx.problem.vsan.clustering.disabled | warning | esx.problem.vsan.clustering.disabled|VSAN clustering and directory services have been disabled thus will be no longer available. |
esx.problem.vsan.net.not.ready | error | esx.problem.vsan.net.not.ready|vmknic {1} that is currently configured to be used with VSAN doesn't have an IP address yet. There are no other active network configuration and therefore the VSAN node doesn't have network connectivity. |
esx.problem.vsan.net.redundancy.lost | warning | esx.problem.vsan.net.redundancy.lost|VSAN network configuration doesn't have any redundancy. This might be a problem if further network configuration is removed. |
esx.problem.vsan.net.redundancy.reduced | warning | esx.problem.vsan.net.redundancy.reduced|VSAN network configuration redundancy has been reduced. This might be a problem if further network configuration is removed. |
esx.problem.vsan.no.network.connectivity | error | esx.problem.vsan.no.network.connectivity|VSAN doesn't have any network configuration. This can severely impact several objects in the VSAN datastore. |
esx.problem.vsan.vmknic.not.ready | warning | esx.problem.vsan.vmknic.not.ready|vmknic {1} that is currently configured to be used with VSAN doesn't have an IP address yet. However, there are other network configuration which are active. If those configurations are removed that may cause problems. |
hbr.primary.AppQuiescedDeltaCompletedEvent | info | hbr.primary.AppQuiescedDeltaCompletedEvent|Application consistent delta completed for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} ({bytes} bytes transferred) |
hbr.primary.ConnectionRestoredToHbrServerEvent | info | hbr.primary.ConnectionRestoredToHbrServerEvent|Connection to VR Server restored for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}. |
hbr.primary.DeltaAbortedEvent | warning | hbr.primary.DeltaAbortedEvent|Delta aborted for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}: {reason.@enum.hbr.primary.ReasonForDeltaAbort} |
hbr.primary.DeltaCompletedEvent | info | hbr.primary.DeltaCompletedEvent|Delta completed for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} ({bytes} bytes transferred). |
hbr.primary.DeltaStartedEvent | info | hbr.primary.DeltaStartedEvent|Delta started by {userName} for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}. |
hbr.primary.FSQuiescedDeltaCompletedEvent | warning | hbr.primary.FSQuiescedDeltaCompletedEvent|File system consistent delta completed for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} ({bytes} bytes transferred) |
hbr.primary.FailedToStartDeltaEvent | error | hbr.primary.FailedToStartDeltaEvent|Failed to start delta for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}: {reason.@enum.fault.ReplicationVmFault.ReasonForFault} |
hbr.primary.FailedToStartSyncEvent | error | hbr.primary.FailedToStartSyncEvent|Failed to start full sync for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}: {reason.@enum.fault.ReplicationVmFault.ReasonForFault} |
hbr.primary.InvalidDiskReplicationConfigurationEvent | error | hbr.primary.InvalidDiskReplicationConfigurationEvent|Replication configuration is invalid for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}, disk {diskKey}: {reasonForFault.@enum.fault.ReplicationDiskConfigFault.ReasonForFault} |
hbr.primary.InvalidVmReplicationConfigurationEvent | error | hbr.primary.InvalidVmReplicationConfigurationEvent|Replication configuration is invalid for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}: {reasonForFault.@enum.fault.ReplicationVmConfigFault.ReasonForFault} |
hbr.primary.NoConnectionToHbrServerEvent | warning | hbr.primary.NoConnectionToHbrServerEvent|No connection to VR Server for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}: {reason.@enum.hbr.primary.ReasonForNoServerConnection} |
hbr.primary.NoProgressWithHbrServerEvent | error | hbr.primary.NoProgressWithHbrServerEvent|VR Server error for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}: {reason.@enum.hbr.primary.ReasonForNoServerProgress} |
hbr.primary.QuiesceNotSupported | warning | hbr.primary.QuiesceNotSupported|Quiescing is not supported for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}. |
hbr.primary.SyncCompletedEvent | info | hbr.primary.SyncCompletedEvent|Full sync completed for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} ({bytes} bytes transferred). |
hbr.primary.SyncStartedEvent | info | hbr.primary.SyncStartedEvent|Full sync started by {userName} for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}. |
hbr.primary.UnquiescedDeltaCompletedEvent | warning | hbr.primary.UnquiescedDeltaCompletedEvent|Delta completed for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} ({bytes} bytes transferred). |
hbr.primary.VmReplicationConfigurationChangedEvent | info | hbr.primary.VmReplicationConfigurationChangedEvent|Replication configuration changed for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} ({numDisks} disks, {rpo} minutes RPO, VR Server is {vrServerAddress}:{vrServerPort}). |
AccountCreatedEvent | info | An account was created on host {host.name} |
AccountRemovedEvent | info | Account {account} was removed on host {host.name} |
AccountUpdatedEvent | info | An account was updated on host {host.name} |
AdminPasswordNotChangedEvent | info | The default password for the root user on the host {host.name} has not been changed |
AlarmAcknowledgedEvent | info | Acknowledged alarm '{alarm.name}' on {entity.name} |
AlarmActionTriggeredEvent | info | Alarm '{alarm.name}' on {entity.name} triggered an action |
AlarmClearedEvent | info | Manually cleared alarm '{alarm.name}' on {entity.name} from {from.@enum.ManagedEntity.Status} |
AlarmCreatedEvent | info | Created alarm '{alarm.name}' on {entity.name} |
AlarmEmailCompletedEvent | info | Alarm '{alarm.name}' on {entity.name} sent email to {to} |
AlarmEmailFailedEvent | error | Alarm '{alarm.name}' on {entity.name} cannot send email to {to} |
AlarmEvent | info | <internal> |
AlarmReconfiguredEvent | info | Reconfigured alarm '{alarm.name}' on {entity.name} |
AlarmRemovedEvent | info | Removed alarm '{alarm.name}' on {entity.name} |
AlarmScriptCompleteEvent | info | Alarm '{alarm.name}' on {entity.name} ran script {script} |
AlarmScriptFailedEvent | error | Alarm '{alarm.name}' on {entity.name} did not complete script: {reason.msg} |
AlarmSnmpCompletedEvent | info | Alarm '{alarm.name}': an SNMP trap for entity {entity.name} was sent |
AlarmSnmpFailedEvent | error | Alarm '{alarm.name}' on entity {entity.name} did not send SNMP trap: {reason.msg} |
AlarmStatusChangedEvent | info | Alarm '{alarm.name}' on {entity.name} changed from {from.@enum.ManagedEntity.Status} to {to.@enum.ManagedEntity.Status} |
AllVirtualMachinesLicensedEvent | info | All running virtual machines are licensed |
AlreadyAuthenticatedSessionEvent | info | User cannot logon since the user is already logged on |
AuthorizationEvent | info | <internal> |
BadUsernameSessionEvent | error | Cannot login {userName}@{ipAddress} |
CanceledHostOperationEvent | info | The operation performed on host {host.name} in {datacenter.name} was canceled |
ChangeOwnerOfFileEvent | info | Changed ownership of file name {filename} from {oldOwner} to {newOwner} on {host.name} in {datacenter.name}. |
ChangeOwnerOfFileFailedEvent | error | Cannot change ownership of file name {filename} from {owner} to {attemptedOwner} on {host.name} in {datacenter.name}. |
ClusterComplianceCheckedEvent | info | Checked cluster for compliance |
ClusterCreatedEvent | info | Created cluster {computeResource.name} in {datacenter.name} |
ClusterDestroyedEvent | info | Removed cluster {computeResource.name} in datacenter {datacenter.name} |
ClusterEvent | info | <internal> |
ClusterOvercommittedEvent | error | Insufficient capacity in cluster {computeResource.name} to satisfy resource configuration in {datacenter.name} |
ClusterReconfiguredEvent | info | Reconfigured cluster {computeResource.name} in datacenter {datacenter.name} |
ClusterStatusChangedEvent | info | Configuration status on cluster {computeResource.name} changed from {oldStatus.@enum.ManagedEntity.Status} to {newStatus.@enum.ManagedEntity.Status} in {datacenter.name} |
CustomFieldDefAddedEvent | info | Created new custom field definition {name} |
CustomFieldDefEvent | info | <internal> |
CustomFieldDefRemovedEvent | info | Removed field definition {name} |
CustomFieldDefRenamedEvent | info | Renamed field definition from {name} to {newName} |
CustomFieldEvent | info | <internal> |
CustomFieldValueChangedEvent | info | Changed custom field {name} on {entity.name} in {datacenter.name} to {value} |
CustomizationEvent | info | <internal> |
CustomizationFailed | info | Cannot complete customization of VM {vm.name}. See customization log at {logLocation} on the guest OS for details. |
CustomizationLinuxIdentityFailed | error | An error occurred while setting up Linux identity. See log file '{logLocation}' on guest OS for details. |
CustomizationNetworkSetupFailed | error | An error occurred while setting up network properties of the guest OS. See the log file {logLocation} in the guest OS for details. |
CustomizationStartedEvent | info | Started customization of VM {vm.name}. Customization log located at {logLocation} in the guest OS. |
CustomizationSucceeded | info | Customization of VM {vm.name} succeeded. Customization log located at {logLocation} in the guest OS. |
CustomizationSysprepFailed | error | The version of Sysprep {sysprepVersion} provided for customizing VM {vm.name} does not match the version of guest OS {systemVersion}. See the log file {logLocation} in the guest OS for more information. |
CustomizationUnknownFailure | error | An error occurred while customizing VM {vm.name}. For details reference the log file {logLocation} in the guest OS. |
DVPortgroupCreatedEvent | info | dvPort group {net.name} in {datacenter.name} was added to switch {dvs.name}. |
DVPortgroupDestroyedEvent | info | dvPort group {net.name} in {datacenter.name} was deleted. |
DVPortgroupEvent | info | |
DVPortgroupReconfiguredEvent | info | dvPort group {net.name} in {datacenter.name} was reconfigured. |
DVPortgroupRenamedEvent | info | dvPort group {oldName} in {datacenter.name} was renamed to {newName} |
DasAdmissionControlDisabledEvent | info | vSphere HA admission control disabled for cluster {computeResource.name} in {datacenter.name} |
DasAdmissionControlEnabledEvent | info | vSphere HA admission control enabled for cluster {computeResource.name} in {datacenter.name} |
DasAgentFoundEvent | info | Re-established contact with a primary host in this vSphere HA cluster |
DasAgentUnavailableEvent | error | Unable to contact a primary vSphere HA agent in cluster {computeResource.name} in {datacenter.name} |
DasClusterIsolatedEvent | error | All hosts in the vSphere HA cluster {computeResource.name} in {datacenter.name} were isolated from the network. Check the network configuration for proper network redundancy in the management network. |
DasDisabledEvent | info | vSphere HA disabled for cluster {computeResource.name} in {datacenter.name} |
DasEnabledEvent | info | vSphere HA enabled for cluster {computeResource.name} in {datacenter.name} |
DasHostFailedEvent | error | A possible host failure has been detected by vSphere HA on {failedHost.name} in cluster {computeResource.name} in {datacenter.name} |
DasHostIsolatedEvent | warning | Host {isolatedHost.name} has been isolated from cluster {computeResource.name} in {datacenter.name} |
DatacenterCreatedEvent | info | Created datacenter {datacenter.name} in folder {parent.name} |
DatacenterEvent | info | <internal> |
DatacenterRenamedEvent | info | Renamed datacenter from {oldName} to {newName} |
DatastoreCapacityIncreasedEvent | info | Datastore {datastore.name} increased in capacity from {oldCapacity} bytes to {newCapacity} bytes in {datacenter.name} |
DatastoreDestroyedEvent | info | Removed unconfigured datastore {datastore.name} |
DatastoreDiscoveredEvent | info | Discovered datastore {datastore.name} on {host.name} in {datacenter.name} |
DatastoreDuplicatedEvent | error | Multiple datastores named {datastore} detected on host {host.name} in {datacenter.name} |
DatastoreEvent | info | <internal> |
DatastoreFileCopiedEvent | info | File or directory {sourceFile} copied from {sourceDatastore.name} to {datastore.name} as {targetFile} |
DatastoreFileDeletedEvent | info | File or directory {targetFile} deleted from {datastore.name} |
DatastoreFileEvent | info | <internal> |
DatastoreFileMovedEvent | info | File or directory {sourceFile} moved from {sourceDatastore.name} to {datastore.name} as {targetFile} |
DatastoreIORMReconfiguredEvent | info | Reconfigured Storage I/O Control on datastore {datastore.name} |
DatastorePrincipalConfigured | info | Configured datastore principal {datastorePrincipal} on host {host.name} in {datacenter.name} |
DatastoreRemovedOnHostEvent | info | Removed datastore {datastore.name} from {host.name} in {datacenter.name} |
DatastoreRenamedEvent | info | Renamed datastore from {oldName} to {newName} in {datacenter.name} |
DatastoreRenamedOnHostEvent | info | Renamed datastore from {oldName} to {newName} in {datacenter.name} |
DrsDisabledEvent | info | Disabled DRS on cluster {computeResource.name} in datacenter {datacenter.name} |
DrsEnabledEvent | info | Enabled DRS on {computeResource.name} with automation level {behavior} in {datacenter.name} |
DrsEnteredStandbyModeEvent | info | DRS put {host.name} into standby mode |
DrsEnteringStandbyModeEvent | info | DRS is putting {host.name} into standby mode |
DrsExitStandbyModeFailedEvent | error | DRS cannot move {host.name} out of standby mode |
DrsExitedStandbyModeEvent | info | DRS moved {host.name} out of standby mode |
DrsExitingStandbyModeEvent | info | DRS is moving {host.name} out of standby mode |
DrsInvocationFailedEvent | error | DRS invocation not completed |
DrsRecoveredFromFailureEvent | info | DRS has recovered from the failure |
DrsResourceConfigureFailedEvent | error | Unable to apply DRS resource settings on host {host.name} in {datacenter.name}. {reason.msg}. This can significantly reduce the effectiveness of DRS. |
DrsResourceConfigureSyncedEvent | info | Resource configuration specification returns to synchronization from previous failure on host '{host.name}' in {datacenter.name} |
DrsRuleComplianceEvent | info | {vm.name} on {host.name} in {datacenter.name} is now compliant with DRS VM-Host affinity rules |
DrsRuleViolationEvent | info | {vm.name} on {host.name} in {datacenter.name} is violating a DRS VM-Host affinity rule |
DrsVmMigratedEvent | info | DRS migrated {vm.name} from {sourceHost.name} to {host.name} in cluster {computeResource.name} in {datacenter.name} |
DrsVmPoweredOnEvent | info | DRS powered On {vm.name} on {host.name} in {datacenter.name} |
DuplicateIpDetectedEvent | info | Virtual machine {macAddress} on host {host.name} has a duplicate IP {duplicateIP} |
DvpgImportEvent | info | Import operation with type {importType} was performed on {net.name} |
DvpgRestoreEvent | info | Restore operation was performed on {net.name} |
DvsCreatedEvent | info | A vSphere Distributed Switch {dvs.name} was created in {datacenter.name}. |
DvsDestroyedEvent | info | vSphere Distributed Switch {dvs.name} in {datacenter.name} was deleted. |
DvsEvent | info | vSphere Distributed Switch event |
DvsHealthStatusChangeEvent | info | Health check status was changed in vSphere Distributed Switch {dvs.name} on host {host.name} in {datacenter.name} |
DvsHostBackInSyncEvent | info | The vSphere Distributed Switch {dvs.name} configuration on the host was synchronized with that of the vCenter Server. |
DvsHostJoinedEvent | info | The host {hostJoined.name} joined the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsHostLeftEvent | info | The host {hostLeft.name} left the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsHostStatusUpdated | info | The host {hostMember.name} changed status on the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsHostWentOutOfSyncEvent | warning | The vSphere Distributed Switch {dvs.name} configuration on the host differed from that of the vCenter Server. |
DvsImportEvent | info | Import operation with type {importType} was performed on {dvs.name} |
DvsMergedEvent | info | vSphere Distributed Switch {srcDvs.name} was merged into {dstDvs.name} in {datacenter.name}. |
DvsPortBlockedEvent | info | The dvPort {portKey} was blocked in the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsPortConnectedEvent | info | The dvPort {portKey} was connected in the vSphere Distributed Switch {dvs.name} in {datacenter.name} |
DvsPortCreatedEvent | info | New ports were created in the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsPortDeletedEvent | info | Deleted ports in the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsPortDisconnectedEvent | info | The dvPort {portKey} was disconnected in the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsPortEnteredPassthruEvent | info | The dvPort {portKey} was in passthrough mode in the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsPortExitedPassthruEvent | info | The dvPort {portKey} was not in passthrough mode in the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsPortJoinPortgroupEvent | info | The dvPort {portKey} was moved into the dvPort group {portgroupName} in {datacenter.name}. |
DvsPortLeavePortgroupEvent | info | The dvPort {portKey} was moved out of the dvPort group {portgroupName} in {datacenter.name}. |
DvsPortLinkDownEvent | info | The dvPort {portKey} link was down in the vSphere Distributed Switch {dvs.name} in {datacenter.name} |
DvsPortLinkUpEvent | info | The dvPort {portKey} link was up in the vSphere Distributed Switch {dvs.name} in {datacenter.name} |
DvsPortReconfiguredEvent | info | Reconfigured ports in the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsPortRuntimeChangeEvent | info | The dvPort {portKey} runtime information changed in the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsPortUnblockedEvent | info | The dvPort {portKey} was unblocked in the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsPortVendorSpecificStateChangeEvent | info | The dvPort {portKey} vendor specific state changed in the vSphere Distributed Switch {dvs.name} in {datacenter.name}. |
DvsReconfiguredEvent | info | The vSphere Distributed Switch {dvs.name} in {datacenter.name} was reconfigured. |
DvsRenamedEvent | info | The vSphere Distributed Switch {oldName} in {datacenter.name} was renamed to {newName}. |
DvsRestoreEvent | info | Restore operation was performed on {dvs.name} |
DvsUpgradeAvailableEvent | info | An upgrade for the vSphere Distributed Switch {dvs.name} in datacenter {datacenter.name} is available. |
DvsUpgradeInProgressEvent | info | An upgrade for the vSphere Distributed Switch {dvs.name} in datacenter {datacenter.name} is in progress. |
DvsUpgradeRejectedEvent | info | Cannot complete an upgrade for the vSphere Distributed Switch {dvs.name} in datacenter {datacenter.name} |
DvsUpgradedEvent | info | vSphere Distributed Switch {dvs.name} in datacenter {datacenter.name} was upgraded. |
EnteredMaintenanceModeEvent | info | Host {host.name} in {datacenter.name} has entered maintenance mode |
EnteredStandbyModeEvent | info | The host {host.name} is in standby mode |
EnteringMaintenanceModeEvent | info | Host {host.name} in {datacenter.name} has started to enter maintenance mode |
EnteringStandbyModeEvent | info | The host {host.name} is entering standby mode |
ErrorUpgradeEvent | error | {message} |
Event | info | <internal> |
ExitMaintenanceModeEvent | info | Host {host.name} in {datacenter.name} has exited maintenance mode |
ExitStandbyModeFailedEvent | error | The host {host.name} could not exit standby mode |
ExitedStandbyModeEvent | info | The host {host.name} is no longer in standby mode |
ExitingStandbyModeEvent | info | The host {host.name} is exiting standby mode |
FailoverLevelRestored | info | Sufficient resources are available to satisfy vSphere HA failover level in cluster {computeResource.name} in {datacenter.name} |
GeneralEvent | info | General event: {message} |
GeneralHostErrorEvent | error | Error detected on {host.name} in {datacenter.name}: {message} |
GeneralHostInfoEvent | info | Issue detected on {host.name} in {datacenter.name}: {message} |
GeneralHostWarningEvent | warning | Issue detected on {host.name} in {datacenter.name}: {message} |
GeneralUserEvent | | User logged event: {message} |
GeneralVmErrorEvent | error | Error detected for {vm.name} on {host.name} in {datacenter.name}: {message} |
GeneralVmInfoEvent | info | Issue detected for {vm.name} on {host.name} in {datacenter.name}: {message} |
GeneralVmWarningEvent | warning | Issue detected for {vm.name} on {host.name} in {datacenter.name}: {message} |
GhostDvsProxySwitchDetectedEvent | info | The vSphere Distributed Switch corresponding to the proxy switches {switchUuid} on the host {host.name} does not exist in vCenter Server or does not contain this host. |
GhostDvsProxySwitchRemovedEvent | info | A ghost proxy switch {switchUuid} on the host {host.name} was resolved. |
GlobalMessageChangedEvent | info | The message changed: {message} |
HealthStatusChangedEvent | info | {componentName} status changed from {oldStatus} to {newStatus} |
HostAddFailedEvent | error | Cannot add host {hostname} to datacenter {datacenter.name} |
HostAddedEvent | info | Added host {host.name} to datacenter {datacenter.name} |
HostAdminDisableEvent | warning | Administrator access to the host {host.name} is disabled |
HostAdminEnableEvent | warning | Administrator access to the host {host.name} has been restored |
HostCnxFailedAccountFailedEvent | error | Cannot connect {host.name} in {datacenter.name}: cannot configure management account |
HostCnxFailedAlreadyManagedEvent | error | Cannot connect {host.name} in {datacenter.name}: already managed by {serverName} |
HostCnxFailedBadCcagentEvent | error | Cannot connect host {host.name} in {datacenter.name} : server agent is not responding |
HostCnxFailedBadUsernameEvent | error | Cannot connect {host.name} in {datacenter.name}: incorrect user name or password |
HostCnxFailedBadVersionEvent | error | Cannot connect {host.name} in {datacenter.name}: incompatible version |
HostCnxFailedCcagentUpgradeEvent | error | Cannot connect host {host.name} in {datacenter.name}. Did not install or upgrade vCenter agent service. |
HostCnxFailedEvent | error | Cannot connect {host.name} in {datacenter.name}: error connecting to host |
HostCnxFailedNetworkErrorEvent | error | Cannot connect {host.name} in {datacenter.name}: network error |
HostCnxFailedNoAccessEvent | error | Cannot connect host {host.name} in {datacenter.name}: account has insufficient privileges |
HostCnxFailedNoConnectionEvent | error | Cannot connect host {host.name} in {datacenter.name} |
HostCnxFailedNoLicenseEvent | error | Cannot connect {host.name} in {datacenter.name}: not enough CPU licenses |
HostCnxFailedNotFoundEvent | error | Cannot connect {host.name} in {datacenter.name}: incorrect host name |
HostCnxFailedTimeoutEvent | error | Cannot connect {host.name} in {datacenter.name}: time-out waiting for host response |
HostComplianceCheckedEvent | info | Host {host.name} checked for compliance. |
HostCompliantEvent | info | Host {host.name} is in compliance with the attached profile |
HostConfigAppliedEvent | info | Host configuration changes applied. |
HostConnectedEvent | info | Connected to {host.name} in {datacenter.name} |
HostConnectionLostEvent | error | Host {host.name} in {datacenter.name} is not responding |
HostDVPortEvent | info | dvPort connected to host {host.name} in {datacenter.name} changed status |
HostDasDisabledEvent | info | vSphere HA agent on {host.name} in cluster {computeResource.name} in {datacenter.name} is disabled |
HostDasDisablingEvent | info | vSphere HA is being disabled on {host.name} in cluster {computeResource.name} in datacenter {datacenter.name} |
HostDasEnabledEvent | info | vSphere HA agent on {host.name} in cluster {computeResource.name} in {datacenter.name} is enabled |
HostDasEnablingEvent | warning | Enabling vSphere HA agent on {host.name} in cluster {computeResource.name} in {datacenter.name} |
HostDasErrorEvent | error | vSphere HA agent on {host.name} in cluster {computeResource.name} in {datacenter.name} has an error {message}: {reason.@enum.HostDasErrorEvent.HostDasErrorReason} |
HostDasEvent | info | <internal> |
HostDasOkEvent | info | vSphere HA agent on host {host.name} in cluster {computeResource.name} in {datacenter.name} is configured correctly |
HostDisconnectedEvent | info | Disconnected from {host.name} in {datacenter.name}. Reason: {reason.@enum.HostDisconnectedEvent.ReasonCode} |
HostEnableAdminFailedEvent | error | Cannot restore some administrator permissions to the host {host.name} |
HostEvent | info | <internal> |
HostExtraNetworksEvent | error | Host {host.name} has the following extra networks not used by other hosts for vSphere HA communication:{ips}. Consider using vSphere HA advanced option das.allowNetwork to control network usage |
HostGetShortNameFailedEvent | error | Cannot complete command 'hostname -s' on host {host.name} or returned incorrect name format |
HostInAuditModeEvent | info | Host {host.name} is running in audit mode. The host's configuration will not be persistent across reboots. |
HostInventoryFullEvent | error | Maximum ({capacity}) number of hosts allowed for this edition of vCenter Server has been reached |
HostInventoryUnreadableEvent | info | The virtual machine inventory file on host {host.name} is damaged or unreadable. |
HostIpChangedEvent | info | IP address of the host {host.name} changed from {oldIP} to {newIP} |
HostIpInconsistentEvent | error | Configuration of host IP address is inconsistent on host {host.name}: address resolved to {ipAddress} and {ipAddress2} |
HostIpToShortNameFailedEvent | error | Cannot resolve IP address to short name on host {host.name} |
HostIsolationIpPingFailedEvent | error | vSphere HA agent on host {host.name} in cluster {computeResource.name} in {datacenter.name} could not reach isolation address: {isolationIp} |
HostLicenseExpiredEvent | error | A host license for {host.name} has expired |
HostLocalPortCreatedEvent | info | A host local port {hostLocalPort.portKey} is created on vSphere Distributed Switch {hostLocalPort.switchUuid} to recover from management network connectivity loss on virtual NIC device {hostLocalPort.vnic} on the host {host.name}. |
HostMissingNetworksEvent | error | Host {host.name} does not have the following networks used by other hosts for vSphere HA communication:{ips}. Consider using vSphere HA advanced option das.allowNetwork to control network usage |
HostMonitoringStateChangedEvent | info | vSphere HA host monitoring state in {computeResource.name} in {datacenter.name} changed to {state.@enum.DasConfigInfo.ServiceState} |
HostNoAvailableNetworksEvent | error | Host {host.name} in cluster {computeResource.name} in {datacenter.name} currently has no available networks for vSphere HA Communication. The following networks are currently used by HA: {ips} |
HostNoHAEnabledPortGroupsEvent | error | Host {host.name} in cluster {computeResource.name} in {datacenter.name} has no port groups enabled for vSphere HA communication. |
HostNoRedundantManagementNetworkEvent | warning | Host {host.name} in cluster {computeResource.name} in {datacenter.name} currently has no management network redundancy |
HostNonCompliantEvent | error | Host {host.name} is not in compliance with the attached profile |
HostNotInClusterEvent | error | Host {host.name} is not a cluster member in {datacenter.name} |
HostOvercommittedEvent | error | Insufficient capacity in host {computeResource.name} to satisfy resource configuration in {datacenter.name} |
HostPrimaryAgentNotShortNameEvent | error | Primary agent {primaryAgent} was not specified as a short name to host {host.name} |
HostProfileAppliedEvent | info | Profile is applied on the host {host.name} |
HostReconnectionFailedEvent | error | Cannot reconnect to {host.name} in {datacenter.name} |
HostRemovedEvent | info | Removed host {host.name} in {datacenter.name} |
HostShortNameInconsistentEvent | error | Host names {shortName} and {shortName2} both resolved to the same IP address. Check the host's network configuration and DNS entries |
HostShortNameToIpFailedEvent | error | Cannot resolve short name {shortName} to IP address on host {host.name} |
HostShutdownEvent | info | Shut down of {host.name} in {datacenter.name}: {reason} |
HostStatusChangedEvent | info | Configuration status on host {computeResource.name} changed from {oldStatus.@enum.ManagedEntity.Status} to {newStatus.@enum.ManagedEntity.Status} in {datacenter.name} |
HostSyncFailedEvent | error | Cannot synchronize host {host.name}. {reason.msg} |
HostUpgradeFailedEvent | error | Cannot install or upgrade vCenter agent service on {host.name} in {datacenter.name} |
HostUserWorldSwapNotEnabledEvent | warning | The userworld swap is not enabled on the host {host.name} |
HostVnicConnectedToCustomizedDVPortEvent | info | Host {host.name} vNIC {vnic.vnic} was reconfigured to use dvPort {vnic.port.portKey} with port level configuration, which might be different from the dvPort group. |
HostWwnChangedEvent | warning | WWNs are changed for {host.name} |
HostWwnConflictEvent | error | The WWN ({wwn}) of {host.name} conflicts with the currently registered WWN |
IncorrectHostInformationEvent | error | Host {host.name} did not provide the information needed to acquire the correct set of licenses |
InfoUpgradeEvent | info | {message} |
InsufficientFailoverResourcesEvent | error | Insufficient resources to satisfy vSphere HA failover level on cluster {computeResource.name} in {datacenter.name} |
InvalidEditionEvent | error | The license edition '{feature}' is invalid |
vim.event.LicenseDowngradedEvent | warning | vim.event.LicenseDowngradedEvent|License downgrade: {licenseKey} removes the following features: {lostFeatures} |
LicenseEvent | info | <internal> |
LicenseExpiredEvent | error | License {feature.featureName} has expired |
LicenseNonComplianceEvent | error | License inventory is not compliant. Licenses are overused |
LicenseRestrictedEvent | error | Unable to acquire licenses due to a restriction in the option file on the license server. |
LicenseServerAvailableEvent | info | License server {licenseServer} is available |
LicenseServerUnavailableEvent | error | License server {licenseServer} is unavailable |
LocalDatastoreCreatedEvent | info | Created local datastore {datastore.name} on {host.name} in {datacenter.name} |
LocalTSMEnabledEvent | info | ESXi Shell for the host {host.name} has been enabled |
LockerMisconfiguredEvent | warning | Datastore {datastore} which is configured to back the locker does not exist |
LockerReconfiguredEvent | info | Locker was reconfigured from {oldDatastore} to {newDatastore} datastore |
MigrationErrorEvent | error | Unable to migrate {vm.name} from {host.name} in {datacenter.name}: {fault.msg} |
MigrationEvent | info | <internal> |
MigrationHostErrorEvent | error | Unable to migrate {vm.name} from {host.name} to {dstHost.name} in {datacenter.name}: {fault.msg} |
MigrationHostWarningEvent | warning | Migration of {vm.name} from {host.name} to {dstHost.name} in {datacenter.name}: {fault.msg} |
MigrationResourceErrorEvent | error | Cannot migrate {vm.name} from {host.name} to {dstHost.name} and resource pool {dstPool.name} in {datacenter.name}: {fault.msg} |
MigrationResourceWarningEvent | warning | Migration of {vm.name} from {host.name} to {dstHost.name} and resource pool {dstPool.name} in {datacenter.name}: {fault.msg} |
MigrationWarningEvent | warning | Migration of {vm.name} from {host.name} in {datacenter.name}: {fault.msg} |
MtuMatchEvent | info | The MTU configured in the vSphere Distributed Switch matches the physical switch connected to uplink port {healthResult.uplinkPortKey} in vSphere Distributed Switch {dvs.name} on host {host.name} in {datacenter.name} |
MtuMismatchEvent | error | The MTU configured in the vSphere Distributed Switch does not match the physical switch connected to uplink port {healthResult.uplinkPortKey} in vSphere Distributed Switch {dvs.name} on host {host.name} in {datacenter.name} |
NASDatastoreCreatedEvent | info | Created NAS datastore {datastore.name} on {host.name} in {datacenter.name} |
NetworkRollbackEvent | error | Network configuration on the host {host.name} is rolled back as it disconnects the host from vCenter server. |
NoAccessUserEvent | error | Cannot login user {userName}@{ipAddress}: no permission |
NoDatastoresConfiguredEvent | info | No datastores have been configured on the host {host.name} |
NoLicenseEvent | error | A required license {feature.featureName} is not reserved |
NoMaintenanceModeDrsRecommendationForVM | info | Unable to automatically migrate {vm.name} from {host.name} |
NonVIWorkloadDetectedOnDatastoreEvent | info | An unmanaged I/O workload is detected on a SIOC-enabled datastore: {datastore.name}. |
NotEnoughResourcesToStartVmEvent | warning | Not enough resources to failover {vm.name} in {computeResource.name} in {datacenter.name}. vSphere HA will retry when resources become available. |
OutOfSyncDvsHost | warning | The vSphere Distributed Switch configuration on some hosts differed from that of the vCenter Server. |
PermissionAddedEvent | info | Permission created for {principal} on {entity.name}, role is {role.name}, propagation is {propagate.@enum.auth.Permission.propagate} |
PermissionEvent | info | <internal> |
PermissionRemovedEvent | info | Permission rule removed for {principal} on {entity.name} |
PermissionUpdatedEvent | info | Permission changed for {principal} on {entity.name}, role is {role.name}, propagation is {propagate.@enum.auth.Permission.propagate} |
ProfileAssociatedEvent | info | Profile {profile.name} attached. |
ProfileChangedEvent | info | Profile {profile.name} was changed. |
ProfileCreatedEvent | info | Profile is created. |
ProfileDissociatedEvent | info | Profile {profile.name} detached. |
ProfileEvent | info | <internal> |
ProfileReferenceHostChangedEvent | info | Profile {profile.name} reference host changed. |
ProfileRemovedEvent | info | Profile was removed. |
RecoveryEvent | info | The host {hostName} network connectivity was recovered on the management virtual NIC {vnic} by connecting to a new port {portKey} on the vSphere Distributed Switch {dvsUuid}. |
RemoteTSMEnabledEvent | info | SSH for the host {host.name} has been enabled |
ResourcePoolCreatedEvent | info | Created resource pool {resourcePool.name} in compute-resource {computeResource.name} in {datacenter.name} |
ResourcePoolDestroyedEvent | info | Removed resource pool {resourcePool.name} on {computeResource.name} in {datacenter.name} |
ResourcePoolEvent | info | <internal> |
ResourcePoolMovedEvent | info | Moved resource pool {resourcePool.name} from {oldParent.name} to {newParent.name} on {computeResource.name} in {datacenter.name} |
ResourcePoolReconfiguredEvent | info | Updated configuration for {resourcePool.name} in compute-resource {computeResource.name} in {datacenter.name} |
ResourceViolatedEvent | error | Resource usage exceeds configuration for resource pool {resourcePool.name} in compute-resource {computeResource.name} in {datacenter.name} |
RoleAddedEvent | info | New role {role.name} created |
RoleEvent | info | <internal> |
RoleRemovedEvent | info | Role {role.name} removed |
RoleUpdatedEvent | info | Modified role {role.name} |
RollbackEvent | info | The Network API {methodName} on this entity caused the host {hostName} to be disconnected from the vCenter Server. The configuration change was rolled back on the host. |
ScheduledTaskCompletedEvent | info | Task {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} completed successfully |
ScheduledTaskCreatedEvent | info | Created task {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} |
ScheduledTaskEmailCompletedEvent | info | Task {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} sent email to {to} |
ScheduledTaskEmailFailedEvent | error | Task {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} cannot send email to {to}: {reason.msg} |
ScheduledTaskEvent | info | <internal> |
ScheduledTaskFailedEvent | error | Task {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} cannot be completed: {reason.msg} |
ScheduledTaskReconfiguredEvent | info | Reconfigured task {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} |
ScheduledTaskRemovedEvent | info | Removed task {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} |
ScheduledTaskStartedEvent | info | Running task {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} |
ServerLicenseExpiredEvent | error | A vCenter Server license has expired |
ServerStartedSessionEvent | info | vCenter started |
SessionEvent | info | <internal> |
SessionTerminatedEvent | info | A session for user '{terminatedUsername}' has stopped |
TaskEvent | info | Task: {info.descriptionId} |
TaskTimeoutEvent | info | Task: {info.descriptionId} time-out |
TeamingMatchEvent | info | Teaming configuration in the vSphere Distributed Switch {dvs.name} on host {host.name} matches the physical switch configuration in {datacenter.name}. Detail: {healthResult.summary.@enum.dvs.VmwareDistributedVirtualSwitch.TeamingMatchStatus} |
TeamingMisMatchEvent | error | Teaming configuration in the vSphere Distributed Switch {dvs.name} on host {host.name} does not match the physical switch configuration in {datacenter.name}. Detail: {healthResult.summary.@enum.dvs.VmwareDistributedVirtualSwitch.TeamingMatchStatus} |
TemplateBeingUpgradedEvent | info | Upgrading template {legacyTemplate} |
TemplateUpgradeEvent | info | <internal> |
TemplateUpgradeFailedEvent | info | Cannot upgrade template {legacyTemplate} due to: {reason.msg} |
TemplateUpgradedEvent | info | Template {legacyTemplate} upgrade completed |
TimedOutHostOperationEvent | warning | The operation performed on {host.name} in {datacenter.name} timed out |
UnlicensedVirtualMachinesEvent | info | There are {unlicensed} unlicensed virtual machines on host {host} - there are only {available} licenses available |
UnlicensedVirtualMachinesFoundEvent | info | {unlicensed} unlicensed virtual machines found on host {host} |
UpdatedAgentBeingRestartedEvent | info | The agent on host {host.name} is updated and will soon restart |
UpgradeEvent | info | <internal> |
UplinkPortMtuNotSupportEvent | error | Not all VLAN MTU settings on the external physical switch allow the vSphere Distributed Switch maximum MTU size packets to pass on the uplink port {healthResult.uplinkPortKey} in vSphere Distributed Switch {dvs.name} on host {host.name} in {datacenter.name}. |
UplinkPortMtuSupportEvent | info | All VLAN MTU settings on the external physical switch allow the vSphere Distributed Switch maximum MTU size packets to pass on the uplink port {healthResult.uplinkPortKey} in vSphere Distributed Switch {dvs.name} on host {host.name} in {datacenter.name}. |
UplinkPortVlanTrunkedEvent | info | The configured VLAN in the vSphere Distributed Switch was trunked by the physical switch connected to uplink port {healthResult.uplinkPortKey} in vSphere Distributed Switch {dvs.name} on host {host.name} in {datacenter.name}. |
UplinkPortVlanUntrunkedEvent | error | Not all the configured VLANs in the vSphere Distributed Switch were trunked by the physical switch connected to uplink port {healthResult.uplinkPortKey} in vSphere Distributed Switch {dvs.name} on host {host.name} in {datacenter.name}. |
UserAssignedToGroup | info | User {userLogin} was added to group {group} |
UserLoginSessionEvent | info | User {userName}@{ipAddress} logged in as {userAgent} |
UserLogoutSessionEvent | info | User {userName}@{ipAddress} logged out (login time: {loginTime}, number of API invocations: {callCount}, user agent: {userAgent}) |
UserPasswordChanged | info | Password was changed for account {userLogin} on host {host.name} |
UserUnassignedFromGroup | info | User {userLogin} removed from group {group} |
UserUpgradeEvent | | {message} |
VMFSDatastoreCreatedEvent | info | Created VMFS datastore {datastore.name} on {host.name} in {datacenter.name} |
VMFSDatastoreExpandedEvent | info | Expanded VMFS datastore {datastore.name} on {host.name} in {datacenter.name} |
VMFSDatastoreExtendedEvent | info | Extended VMFS datastore {datastore.name} on {host.name} in {datacenter.name} |
VMotionLicenseExpiredEvent | error | A vMotion license for {host.name} has expired |
VcAgentUninstallFailedEvent | error | Cannot uninstall vCenter agent from {host.name} in {datacenter.name}. {reason.@enum.fault.AgentInstallFailed.Reason} |
VcAgentUninstalledEvent | info | vCenter agent has been uninstalled from {host.name} in {datacenter.name} |
VcAgentUpgradeFailedEvent | error | Cannot upgrade vCenter agent on {host.name} in {datacenter.name}. {reason.@enum.fault.AgentInstallFailed.Reason} |
VcAgentUpgradedEvent | info | vCenter agent has been upgraded on {host.name} in {datacenter.name} |
VimAccountPasswordChangedEvent | info | VIM account password was changed on host {host.name} |
VmAcquiredMksTicketEvent | info | Remote console to {vm.name} on {host.name} in {datacenter.name} has been opened |
VmAcquiredTicketEvent | info | A ticket for {vm.name} of type {ticketType} on {host.name} in {datacenter.name} has been acquired |
VmAutoRenameEvent | info | Invalid name for {vm.name} on {host.name} in {datacenter.name}. Renamed from {oldName} to {newName} |
VmBeingClonedEvent | info | Cloning {vm.name} on host {host.name} in {datacenter.name} to {destName} on host {destHost.name} |
VmBeingClonedNoFolderEvent | info | Cloning {vm.name} on host {host.name} in {datacenter.name} to {destName} on host {destHost.name} to a vApp |
VmBeingCreatedEvent | info | Creating {vm.name} on host {host.name} in {datacenter.name} |
VmBeingDeployedEvent | info | Deploying {vm.name} on host {host.name} in {datacenter.name} from template {srcTemplate.name} |
VmBeingHotMigratedEvent | info | Migrating {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name} in {datacenter.name} |
VmBeingMigratedEvent | info | Relocating {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name} in {datacenter.name} |
VmBeingRelocatedEvent | info | Relocating {vm.name} in {datacenter.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name} |
VmCloneEvent | info | <internal> |
VmCloneFailedEvent | error | Cannot clone {vm.name}: {reason.msg} |
VmClonedEvent | info | Clone of {sourceVm.name} completed |
VmConfigMissingEvent | info | Configuration file for {vm.name} on {host.name} in {datacenter.name} cannot be found |
VmConnectedEvent | info | Virtual machine {vm.name} is connected |
VmCreatedEvent | info | Created virtual machine {vm.name} on {host.name} in {datacenter.name} |
VmDVPortEvent | info | dvPort connected to VM {vm.name} on {host.name} in {datacenter.name} changed status |
VmDasBeingResetEvent | info | {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name} reset by vSphere HA. Reason: {reason.@enum.VmDasBeingResetEvent.ReasonCode} |
VmDasBeingResetWithScreenshotEvent | info | {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name} reset by vSphere HA. Reason: {reason.@enum.VmDasBeingResetEvent.ReasonCode}. A screenshot is saved at {screenshotFilePath}. |
VmDasResetFailedEvent | warning | vSphere HA cannot reset {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name} |
VmDasUpdateErrorEvent | error | Unable to update vSphere HA agents given the state of {vm.name} |
VmDasUpdateOkEvent | info | vSphere HA agents have been updated with the current state of the virtual machine |
VmDateRolledBackEvent | error | Disconnecting all hosts as the date of virtual machine {vm.name} has been rolled back |
VmDeployFailedEvent | error | Cannot deploy template: {reason.msg} |
VmDeployedEvent | info | Template {srcTemplate.name} deployed on host {host.name} |
VmDisconnectedEvent | info | {vm.name} on host {host.name} in {datacenter.name} is disconnected |
VmDiscoveredEvent | info | Discovered {vm.name} on {host.name} in {datacenter.name} |
VmDiskFailedEvent | error | Cannot create virtual disk {disk} |
VmEmigratingEvent | info | Migrating {vm.name} off host {host.name} in {datacenter.name} |
VmEndRecordingEvent | info | End a recording session on {vm.name} |
VmEndReplayingEvent | info | End a replay session on {vm.name} |
VmEvent | info | <internal> |
VmFailedMigrateEvent | error | Cannot migrate {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name} in {datacenter.name} |
VmFailedRelayoutEvent | error | Cannot complete relayout {vm.name} on {host.name} in {datacenter.name}: {reason.msg} |
VmFailedRelayoutOnVmfs2DatastoreEvent | error | Cannot complete relayout for virtual machine {vm.name} which has disks on a VMFS2 volume. |
VmFailedStartingSecondaryEvent | error | vCenter cannot start the Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name}. Reason: {reason.@enum.VmFailedStartingSecondaryEvent.FailureReason} |
VmFailedToPowerOffEvent | error | Cannot power off {vm.name} on {host.name} in {datacenter.name}: {reason.msg} |
VmFailedToPowerOnEvent | error | Cannot power on {vm.name} on {host.name} in {datacenter.name}. {reason.msg} |
VmFailedToRebootGuestEvent | error | Cannot reboot the guest OS for {vm.name} on {host.name} in {datacenter.name}. {reason.msg} |
VmFailedToResetEvent | error | Cannot suspend {vm.name} on {host.name} in {datacenter.name}: {reason.msg} |
VmFailedToShutdownGuestEvent | error | {vm.name} cannot shut down the guest OS on {host.name} in {datacenter.name}: {reason.msg} |
VmFailedToStandbyGuestEvent | error | {vm.name} cannot standby the guest OS on {host.name} in {datacenter.name}: {reason.msg} |
VmFailedToSuspendEvent | error | Cannot suspend {vm.name} on {host.name} in {datacenter.name}: {reason.msg} |
VmFailedUpdatingSecondaryConfig | error | vCenter cannot update the Fault Tolerance secondary VM configuration for {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} |
VmFailoverFailed | warning | vSphere HA unsuccessfully failed over {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name}. vSphere HA will retry if the maximum number of attempts has not been exceeded. Reason: {reason.msg} |
VmFaultToleranceStateChangedEvent | info | Fault Tolerance state of {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} changed from {oldState.@enum.VirtualMachine.FaultToleranceState} to {newState.@enum.VirtualMachine.FaultToleranceState} |
VmFaultToleranceTurnedOffEvent | info | Fault Tolerance protection has been turned off for {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} |
VmFaultToleranceVmTerminatedEvent | info | The Fault Tolerance VM {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} has been terminated. {reason.@enum.VmFaultToleranceVmTerminatedEvent.TerminateReason} |
VmGuestRebootEvent | info | Guest OS reboot for {vm.name} on {host.name} in {datacenter.name} |
VmGuestShutdownEvent | info | Guest OS shut down for {vm.name} on {host.name} in {datacenter.name} |
VmGuestStandbyEvent | info | Guest OS standby for {vm.name} on {host.name} in {datacenter.name} |
VmHealthMonitoringStateChangedEvent | info | vSphere HA VM monitoring state in {computeResource.name} in {datacenter.name} changed to {state.@enum.DasConfigInfo.VmMonitoringState} |
VmInstanceUuidAssignedEvent | info | Assign a new instance UUID ({instanceUuid}) to {vm.name} |
VmInstanceUuidChangedEvent | info | The instance UUID of {vm.name} has been changed from ({oldInstanceUuid}) to ({newInstanceUuid}) |
VmInstanceUuidConflictEvent | error | The instance UUID ({instanceUuid}) of {vm.name} conflicts with the instance UUID assigned to {conflictedVm.name} |
VmMacAssignedEvent | info | New MAC address ({mac}) assigned to adapter {adapter} for {vm.name} |
VmMacChangedEvent | warning | Changed MAC address from {oldMac} to {newMac} for adapter {adapter} for {vm.name} |
VmMacConflictEvent | error | The MAC address ({mac}) of {vm.name} conflicts with MAC assigned to {conflictedVm.name} |
VmMaxFTRestartCountReached | warning | vSphere HA stopped trying to restart Secondary VM {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name} because the maximum VM restart count was reached |
VmMaxRestartCountReached | warning | vSphere HA stopped trying to restart {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name} because the maximum VM restart count was reached |
VmMessageErrorEvent | error | Error message on {vm.name} on {host.name} in {datacenter.name}: {message} |
VmMessageEvent | info | Message on {vm.name} on {host.name} in {datacenter.name}: {message} |
VmMessageWarningEvent | warning | Warning message on {vm.name} on {host.name} in {datacenter.name}: {message} |
VmMigratedEvent | info | Migration of virtual machine {vm.name} from {sourceHost.name}, {sourceDatastore.name} to {host.name}, {ds.name} completed |
VmNoCompatibleHostForSecondaryEvent | error | No compatible host for the Fault Tolerance secondary VM {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} |
VmNoNetworkAccessEvent | warning | Not all networks for {vm.name} are accessible by {destHost.name} |
VmOrphanedEvent | warning | {vm.name} does not exist on {host.name} in {datacenter.name} |
VmPowerOffOnIsolationEvent | info | vSphere HA powered off {vm.name} on the isolated host {isolatedHost.name} in cluster {computeResource.name} in {datacenter.name} |
VmPoweredOffEvent | info | {vm.name} on {host.name} in {datacenter.name} is powered off |
VmPoweredOnEvent | info | {vm.name} on {host.name} in {datacenter.name} is powered on |
VmPoweringOnWithCustomizedDVPortEvent | info | Virtual machine {vm.name} powered On with vNICs connected to dvPorts that have a port level configuration, which might be different from the dvPort group configuration. |
VmPrimaryFailoverEvent | error | Fault Tolerance VM ({vm.name}) failed over to {host.name} in cluster {computeResource.name} in {datacenter.name}. {reason.@enum.VirtualMachine.NeedSecondaryReason} |
VmReconfiguredEvent | info | Reconfigured {vm.name} on {host.name} in {datacenter.name} |
VmRegisteredEvent | info | Registered {vm.name} on {host.name} in {datacenter.name} |
VmRelayoutSuccessfulEvent | info | Relayout of {vm.name} on {host.name} in {datacenter.name} completed |
VmRelayoutUpToDateEvent | info | {vm.name} on {host.name} in {datacenter.name} is in the correct format and relayout is not necessary |
VmReloadFromPathEvent | info | {vm.name} on {host.name} reloaded from new configuration {configPath}. |
VmReloadFromPathFailedEvent | error | {vm.name} on {host.name} could not be reloaded from {configPath}. |
VmRelocateFailedEvent | error | Cannot relocate virtual machine '{vm.name}' in {datacenter.name} |
VmRelocateSpecEvent | info | <internal> |
VmRelocatedEvent | info | Completed the relocation of the virtual machine |
VmRemoteConsoleConnectedEvent | info | Remote console connected to {vm.name} on host {host.name} |
VmRemoteConsoleDisconnectedEvent | info | Remote console disconnected from {vm.name} on host {host.name} |
VmRemovedEvent | info | Removed {vm.name} on {host.name} from {datacenter.name} |
VmRenamedEvent | warning | Renamed {vm.name} from {oldName} to {newName} in {datacenter.name} |
VmRequirementsExceedCurrentEVCModeEvent | warning | Feature requirements of {vm.name} exceed capabilities of {host.name}'s current EVC mode. |
VmResettingEvent | info | {vm.name} on {host.name} in {datacenter.name} is reset |
VmResourcePoolMovedEvent | info | Moved {vm.name} from resource pool {oldParent.name} to {newParent.name} in {datacenter.name} |
VmResourceReallocatedEvent | info | Changed resource allocation for {vm.name} |
VmRestartedOnAlternateHostEvent | info | Virtual machine {vm.name} was restarted on {host.name} since {sourceHost.name} failed |
VmResumingEvent | info | {vm.name} on {host.name} in {datacenter.name} is resumed |
VmSecondaryAddedEvent | info | A Fault Tolerance secondary VM has been added for {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} |
VmSecondaryDisabledBySystemEvent | error | vCenter disabled Fault Tolerance on VM {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} because the Secondary VM could not be powered On. |
VmSecondaryDisabledEvent | info | Disabled Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} |
VmSecondaryEnabledEvent | info | Enabled Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} |
VmSecondaryStartedEvent | info | Started Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} |
VmShutdownOnIsolationEvent | info | vSphere HA shut down {vm.name} was shut down on the isolated host {isolatedHost.name} in cluster {computeResource.name} in {datacenter.name}: {shutdownResult.@enum.VmShutdownOnIsolationEvent.Operation} |
VmStartRecordingEvent | info | Start a recording session on {vm.name} |
VmStartReplayingEvent | info | Start a replay session on {vm.name} |
VmStartingEvent | info | {vm.name} on host {host.name} in {datacenter.name} is starting |
VmStartingSecondaryEvent | info | Starting Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} |
VmStaticMacConflictEvent | error | The static MAC address ({mac}) of {vm.name} conflicts with MAC assigned to {conflictedVm.name} |
VmStoppingEvent | info | {vm.name} on {host.name} in {datacenter.name} is stopping |
VmSuspendedEvent | info | {vm.name} on {host.name} in {datacenter.name} is suspended |
VmSuspendingEvent | info | {vm.name} on {host.name} in {datacenter.name} is being suspended |
VmTimedoutStartingSecondaryEvent | error | Starting the Fault Tolerance secondary VM {vm.name} on host {host.name} in cluster {computeResource.name} in {datacenter.name} timed out within {timeout} ms |
VmUnsupportedStartingEvent | warning | Unsupported guest OS {guestId} for {vm.name} on {host.name} in {datacenter.name} |
VmUpgradeCompleteEvent | info | Virtual machine compatibility upgraded to {version.@enum.vm.hwVersion} |
VmUpgradeFailedEvent | error | Cannot upgrade virtual machine compatibility. |
VmUpgradingEvent | info | Upgrading virtual machine compatibility of {vm.name} in {datacenter.name} to {version.@enum.vm.hwVersion} |
VmUuidAssignedEvent | info | Assigned new BIOS UUID ({uuid}) to {vm.name} on {host.name} in {datacenter.name} |
VmUuidChangedEvent | warning | Changed BIOS UUID from {oldUuid} to {newUuid} for {vm.name} on {host.name} in {datacenter.name} |
VmUuidConflictEvent | error | BIOS ID ({uuid}) of {vm.name} conflicts with that of {conflictedVm.name} |
VmVnicPoolReservationViolationClearEvent | info | The reservation violation on the virtual NIC network resource pool {vmVnicResourcePoolName} with key {vmVnicResourcePoolKey} on {dvs.name} is cleared |
VmVnicPoolReservationViolationRaiseEvent | info | The reservation allocated to the virtual NIC network resource pool {vmVnicResourcePoolName} with key {vmVnicResourcePoolKey} on {dvs.name} is violated |
VmWwnAssignedEvent | info | New WWNs assigned to {vm.name} |
VmWwnChangedEvent | warning | WWNs are changed for {vm.name} |
VmWwnConflictEvent | error | The WWN ({wwn}) of {vm.name} conflicts with the currently registered WWN |
WarningUpgradeEvent | warning | {message} |
IScsiBootFailureEvent | warning | Booting from iSCSI failed with an error. See the VMware Knowledge Base for information on configuring iBFT networking. |
vprob.net.connectivity.lost | error | vprob.net.connectivity.lost|Lost network connectivity on virtual switch {1}. Physical NIC {2} is down. Affected portgroups:{3}. |
vprob.net.e1000.tso6.notsupported | error | vprob.net.e1000.tso6.notsupported|Guest-initiated IPv6 TCP Segmentation Offload (TSO) packets ignored. Manually disable TSO inside the guest operating system in virtual machine {1}, or use a different virtual adapter. |
vprob.net.migrate.bindtovmk | warning | vprob.net.migrate.bindtovmk|The ESX advanced config option /Migrate/Vmknic is set to an invalid vmknic: {1}. /Migrate/Vmknic specifies a vmknic that vMotion binds to for improved performance. Please update the config option with a valid vmknic or, if you do not want vMotion to bind to a specific vmknic, remove the invalid vmknic and leave the option blank. |
vprob.net.proxyswitch.port.unavailable | warning | vprob.net.proxyswitch.port.unavailable|Virtual NIC with hardware address {1} failed to connect to distributed virtual port {2} on switch {3}. There are no more ports available on the host proxy switch. |
vprob.net.redundancy.degraded | warning | vprob.net.redundancy.degraded|Uplink redundancy degraded on virtual switch {1}. Physical NIC {2} is down. {3} uplinks still up. Affected portgroups:{4}. |
vprob.net.redundancy.lost | warning | vprob.net.redundancy.lost|Lost uplink redundancy on virtual switch {1}. Physical NIC {2} is down. Affected portgroups:{3}. |
vprob.scsi.device.thinprov.atquota | warning | vprob.scsi.device.thinprov.atquota|Space utilization on thin-provisioned device {1} exceeded configured threshold. |
vprob.storage.connectivity.lost | error | vprob.storage.connectivity.lost|Lost connectivity to storage device {1}. Path {2} is down. Affected datastores: {3}. |
vprob.storage.redundancy.degraded | warning | vprob.storage.redundancy.degraded|Path redundancy to storage device {1} degraded. Path {2} is down. {3} remaining active paths. Affected datastores: {4}. |
vprob.storage.redundancy.lost | warning | vprob.storage.redundancy.lost|Lost path redundancy to storage device {1}. Path {2} is down. Affected datastores: {3}. |
vprob.vmfs.error.volume.is.locked | error | vprob.vmfs.error.volume.is.locked|Volume on device {1} is locked, possibly because some remote host encountered an error during a volume operation and could not recover. |
vprob.vmfs.extent.offline | error | vprob.vmfs.extent.offline|An attached device {1} might be offline. The file system {2} is now in a degraded state. While the datastore is still available, parts of data that reside on the extent that went offline might be inaccessible. |
vprob.vmfs.extent.online | info | vprob.vmfs.extent.online|Device {1} backing file system {2} came online. This extent was previously offline. All resources on this device are now available. |
vprob.vmfs.heartbeat.recovered | info | vprob.vmfs.heartbeat.recovered|Successfully restored access to volume {1} ({2}) following connectivity issues. |
vprob.vmfs.heartbeat.timedout | info | vprob.vmfs.heartbeat.timedout|Lost access to volume {1} ({2}) due to connectivity issues. Recovery attempt is in progress and outcome will be reported shortly. |
vprob.vmfs.heartbeat.unrecoverable | error | vprob.vmfs.heartbeat.unrecoverable|Lost connectivity to volume {1} ({2}) and subsequent recovery attempts have failed. |
vprob.vmfs.journal.createfailed | error | vprob.vmfs.journal.createfailed|No space for journal on volume {1} ({2}). Opening volume in read-only metadata mode with limited write support. |
vprob.vmfs.lock.corruptondisk | error | vprob.vmfs.lock.corruptondisk|At least one corrupt on-disk lock was detected on volume {1} ({2}). Other regions of the volume may be damaged too. |
vprob.vmfs.nfs.server.disconnect | error | vprob.vmfs.nfs.server.disconnect|Lost connection to server {1} mount point {2} mounted as {3} ({4}). |
vprob.vmfs.nfs.server.restored | info | vprob.vmfs.nfs.server.restored|Restored connection to server {1} mount point {2} mounted as {3} ({4}). |
vprob.vmfs.resource.corruptondisk | error | vprob.vmfs.resource.corruptondisk|At least one corrupt resource metadata region was detected on volume {1} ({2}). Other regions of the volume might be damaged too. |
com.vmware.vcIntegrity.CancelTask | info | com.vmware.vcIntegrity.CancelTask|Canceling task on [data.name]. |
com.vmware.vcIntegrity.CheckNotification | info | com.vmware.vcIntegrity.CheckNotification|Successfully downloaded notifications. New notifications: [data.Notifications] |
com.vmware.vcIntegrity.CheckNotificationFailed | error | com.vmware.vcIntegrity.CheckNotificationFailed|Could not download notifications. |
com.vmware.vcIntegrity.CheckPXEBootHostFailure | error | com.vmware.vcIntegrity.CheckPXEBootHostFailure|Cannot determine whether host {host.name} is PXE booted. The host will be excluded for the current operation. |
com.vmware.vcIntegrity.CriticallyLowDiskSpace | warning | com.vmware.vcIntegrity.CriticallyLowDiskSpace|VMware vSphere Update Manager is critically low on storage space! Location: [data.Volume]. Available space: [data.FreeSpace]MB. |
com.vmware.vcIntegrity.DisableToolsRemediateOnReboot | info | com.vmware.vcIntegrity.DisableToolsRemediateOnReboot|Successfully disabled the option for VMware Tools upgrade on VM power cycle for [data.name]. |
com.vmware.vcIntegrity.DisableToolsRemediateOnRebootFailed | error | com.vmware.vcIntegrity.DisableToolsRemediateOnRebootFailed|Could not disable the option for VMware Tools upgrade on VM power cycle for [data.name]. |
com.vmware.vcIntegrity.DownloadAlert | info | com.vmware.vcIntegrity.DownloadAlert|VMware vSphere Update Manager download alert (critical/total): ESX [data.esxCritical]/[data.esxTotal] |
com.vmware.vcIntegrity.DownloadFailedPatchBinary | error | com.vmware.vcIntegrity.DownloadFailedPatchBinary|Could not download patch packages for following patches: [data.message]. |
com.vmware.vcIntegrity.DownloadGuestPackage | info | com.vmware.vcIntegrity.DownloadGuestPackage|Successfully downloaded guest patch packages. New packages: [data.Updates] |
com.vmware.vcIntegrity.DownloadGuestPackageFailed | error | com.vmware.vcIntegrity.DownloadGuestPackageFailed|Could not download guest patch packages. |
com.vmware.vcIntegrity.DownloadGuestUnixPackage | info | com.vmware.vcIntegrity.DownloadGuestUnixPackage|Successfully downloaded guest patch packages for UNIX. New packages: [data.Updates] |
com.vmware.vcIntegrity.DownloadGuestUnixPackageFailed | error | com.vmware.vcIntegrity.DownloadGuestUnixPackageFailed|Could not download guest patch packages for UNIX. |
com.vmware.vcIntegrity.DownloadGuestUnixUpdate | info | com.vmware.vcIntegrity.DownloadGuestUnixUpdate|Successfully downloaded guest patch definitions for UNIX. New patches: [data.Updates] |
com.vmware.vcIntegrity.DownloadGuestUnixUpdateFailed | error | com.vmware.vcIntegrity.DownloadGuestUnixUpdateFailed|Could not download guest patch definitions for UNIX. |
com.vmware.vcIntegrity.DownloadGuestUpdate | info | com.vmware.vcIntegrity.DownloadGuestUpdate|Successfully downloaded guest patch definitions. New patches: [data.Updates] |
com.vmware.vcIntegrity.DownloadGuestUpdateFailed | error | com.vmware.vcIntegrity.DownloadGuestUpdateFailed|Could not download guest patch definitions. |
com.vmware.vcIntegrity.DownloadHostPackage | info | com.vmware.vcIntegrity.DownloadHostPackage|Successfully downloaded host patch packages. New packages: [data.Updates] |
com.vmware.vcIntegrity.DownloadHostPackageFailed | error | com.vmware.vcIntegrity.DownloadHostPackageFailed|Could not download host patch packages. |
com.vmware.vcIntegrity.DownloadHostUpdate | info | com.vmware.vcIntegrity.DownloadHostUpdate|Successfully downloaded host patch definitions. New patches: [data.Updates] |
com.vmware.vcIntegrity.DownloadHostUpdateFailed | error | com.vmware.vcIntegrity.DownloadHostUpdateFailed|Could not download host patch definitions. |
com.vmware.vcIntegrity.EnableToolsRemediateOnReboot | info | com.vmware.vcIntegrity.EnableToolsRemediateOnReboot|Successfully enabled the option for VMware Tools upgrade on VM power cycle for [data.name]. |
com.vmware.vcIntegrity.EnableToolsRemediateOnRebootFailed | error | com.vmware.vcIntegrity.EnableToolsRemediateOnRebootFailed|Could not enable the option for VMware Tools upgrade on VM power cycle for [data.name]. |
com.vmware.vcIntegrity.FailToLock | warning | com.vmware.vcIntegrity.FailToLock|There are running tasks for the entity [data.name] that cannot finish within a specific time. The operation will stop. |
com.vmware.vcIntegrity.FtFailedEvent | | com.vmware.vcIntegrity.FtFailedEvent|event.com.vmware.vcIntegrity.FtFailedEvent.fullFormat |
com.vmware.vcIntegrity.GADvdMountError | error | com.vmware.vcIntegrity.GADvdMountError|VMware vSphere Update Manager Guest Agent could not access the DVD drive on {vm.name}. Verify that a DVD drive is available and retry the operation. |
com.vmware.vcIntegrity.GAError | error | com.vmware.vcIntegrity.GAError|An internal error occurred in communication with VMware vSphere Update Manager Guest Agent on {vm.name}. Verify that the VM is powered on and retry the operation. |
com.vmware.vcIntegrity.GAInstallFailed | error | com.vmware.vcIntegrity.GAInstallFailed|Could not install VMware vSphere Update Manager Guest Agent on {vm.name}. Make sure that the VM is powered on. |
com.vmware.vcIntegrity.GAInstalled | info | com.vmware.vcIntegrity.GAInstalled|VMware vSphere Update Manager Guest Agent successfully installed on {vm.name}. |
com.vmware.vcIntegrity.GARuntimeError | error | com.vmware.vcIntegrity.GARuntimeError|An unknown internal error occurred during the required operation on {vm.name}. Check the logs for more details and retry the operation. |
com.vmware.vcIntegrity.GATimeout | error | com.vmware.vcIntegrity.GATimeout|VMware vSphere Update Manager Guest Agent could not respond in time on {vm.name}. Verify that the VM is powered on and that the Guest Agent is running. |
com.vmware.vcIntegrity.HostFirewallClose | info | com.vmware.vcIntegrity.HostFirewallClose|Close [data.name] firewall ports. |
com.vmware.vcIntegrity.HostFirewallOpen | info | com.vmware.vcIntegrity.HostFirewallOpen|Open [data.name] firewall ports. |
com.vmware.vcIntegrity.HostPatchBundleImportCancelledEvent | | com.vmware.vcIntegrity.HostPatchBundleImportCancelledEvent|event.com.vmware.vcIntegrity.HostPatchBundleImportCancelledEvent.fullFormat |
com.vmware.vcIntegrity.HostPatchBundleImportSuccessfulEvent | | com.vmware.vcIntegrity.HostPatchBundleImportSuccessfulEvent|event.com.vmware.vcIntegrity.HostPatchBundleImportSuccessfulEvent.fullFormat |
com.vmware.vcIntegrity.HostPatchBundleImportUnknownErrorEvent | | com.vmware.vcIntegrity.HostPatchBundleImportUnknownErrorEvent|event.com.vmware.vcIntegrity.HostPatchBundleImportUnknownErrorEvent.fullFormat |
com.vmware.vcIntegrity.HostPatchInputRecalledFailure | | com.vmware.vcIntegrity.HostPatchInputRecalledFailure|event.com.vmware.vcIntegrity.HostPatchInputRecalledFailure.fullFormat |
com.vmware.vcIntegrity.HostPatchPrerequisiteRecalledFailure | | com.vmware.vcIntegrity.HostPatchPrerequisiteRecalledFailure|event.com.vmware.vcIntegrity.HostPatchPrerequisiteRecalledFailure.fullFormat |
com.vmware.vcIntegrity.HostPatchRemediateHostConflict | error | com.vmware.vcIntegrity.HostPatchRemediateHostConflict|Host patch [data.patch] conflicts with the package [data.conflictPackage] installed on the host and cannot be remediated. Remove the patch from the baseline or include any suggested additional patches in the baseline and retry remediation operation. |
com.vmware.vcIntegrity.HostPatchRemediateInputConflict | error | com.vmware.vcIntegrity.HostPatchRemediateInputConflict|Host patch [data.patch] conflicts with patch [data.conflictPatch] included in the baseline and cannot be remediated. Remove either of the patch from the baseline and retry the remediation. |
com.vmware.vcIntegrity.HostPatchStageHostConflict | error | com.vmware.vcIntegrity.HostPatchStageHostConflict|Host patch [data.patch] conflicts with the package [data.conflictPackage] installed on the host and cannot be staged. Remove the patch from the baseline or include any suggested additional patches in the baseline and retry stage operation. |
com.vmware.vcIntegrity.HostPatchStageInputConflict | error | com.vmware.vcIntegrity.HostPatchStageInputConflict|Host patch [data.patch] conflicts with patch [data.conflictPatch] included in the baseline and cannot be staged. Remove either of the patch from the baseline and retry the stage operation. |
com.vmware.vcIntegrity.HostSkippedDpmEvent | warning | com.vmware.vcIntegrity.HostSkippedDpmEvent|Cannot remediate host {host.name} because it is a part of a VMware DPM enabled cluster. |
com.vmware.vcIntegrity.HostSkippedDpmFtEvent | warning | com.vmware.vcIntegrity.HostSkippedDpmFtEvent|Cannot remediate host {host.name} because it is a part of a VMware DPM enabled cluster and contains one or more Primary or Secondary VMs on which FT is enabled. |
com.vmware.vcIntegrity.HostSkippedDpmScanEvent | warning | com.vmware.vcIntegrity.HostSkippedDpmScanEvent|Cannot scan host {host.name} because it is a part of a VMware DPM enabled cluster. |
com.vmware.vcIntegrity.HostSkippedDpmStageEvent | warning | com.vmware.vcIntegrity.HostSkippedDpmStageEvent|Cannot stage host {host.name} because it is a part of a VMware DPM enabled cluster. |
com.vmware.vcIntegrity.HostSkippedFtDiffPatchesEvent | warning | com.vmware.vcIntegrity.HostSkippedFtDiffPatchesEvent|Host {host.name} has FT enabled VMs. If you apply different patches to hosts in a cluster, FT cannot be re-enabled. |
com.vmware.vcIntegrity.HostSkippedFtEvent | warning | com.vmware.vcIntegrity.HostSkippedFtEvent|Cannot remediate host {host.name} because it contains one or more Primary or Secondary VMs on which FT is enabled. |
com.vmware.vcIntegrity.HostSkippedFtPairEvent | warning | com.vmware.vcIntegrity.HostSkippedFtPairEvent|Host {host.name} has FT enabled VMs. The host on which the Secondary VMs reside is not selected for remediation. As a result FT cannot be re-enabled. |
com.vmware.vcIntegrity.HostSkippedHacEvent | warning | com.vmware.vcIntegrity.HostSkippedHacEvent|Cannot remediate host {host.name} because it is a part of a HA admission control enabled cluster. |
com.vmware.vcIntegrity.HostSkippedPxeUpgradeUnsupported | warning | com.vmware.vcIntegrity.HostSkippedPxeUpgradeUnsupported|Upgrade operations are not supported on host {host.name} because it is PXE booted. |
com.vmware.vcIntegrity.HostSkippedRemovableDeviceEvent | warning | com.vmware.vcIntegrity.HostSkippedRemovableDeviceEvent|Cannot remediate host {host.name} because it has VMs with a connected removable device. Disconnect all removable devices before remediation. |
com.vmware.vcIntegrity.HostUpdateErrorEsxFileDownload | error | com.vmware.vcIntegrity.HostUpdateErrorEsxFileDownload|Host [data.name] cannot download files from the VMware vSphere Update Manager patch store. Check the network connectivity and firewall setup, and verify that the host can access the configured patch store. |
com.vmware.vcIntegrity.HostUpdateErrorNotInstallable | error | com.vmware.vcIntegrity.HostUpdateErrorNotInstallable|The selected patches [data.arg1] cannot be installed on the host. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateConflictDependencies | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateConflictDependencies|The patches selected for remediation on the host [data.name] depend on other patches that have conflicts. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateDefault | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateDefault|Remediation did not succeed for [data.name]. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateDeviceAttached | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateDeviceAttached|Remediation did not succeed for [data.name]. The host has virtual machines [data.arg1] with connected removable media devices. This prevents the host from entering maintenance mode. Disconnect the removable devices and try again. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateEnterMmode | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateEnterMmode|Remediation did not succeed for [data.name]. The host could not enter maintenance mode. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateExitMmode | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateExitMmode|Remediation did not succeed for [data.name]. The host could not exit maintenance mode. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateHostReboot | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateHostReboot|Remediation did not succeed for [data.name]. The host did not reboot after remediation. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateHostRebootReconnect | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateHostRebootReconnect|Remediation did not succeed for [data.name]. VMware vSphere Update Manager timed out waiting for the host to reconnect after a reboot. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateHostReconnect | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateHostReconnect|Remediation did not succeed for [data.name]. VMware vSphere Update Manager timed out waiting for the host to reconnect. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateHostRestoreVm | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateHostRestoreVm|Remediation did not succeed for [data.name]. Restoring the power state or device connection state for one or more virtual machines on the host did not succeed. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateMetadataCorrupt | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateMetadataCorrupt|Remediation did not succeed for [data.name]. The patch metadata is corrupted. This might be caused by an invalid format of metadata content. You can try to re-download the patches. |
com.vmware.vcIntegrity.HostUpdateErrorRemediateVibDownload | error | com.vmware.vcIntegrity.HostUpdateErrorRemediateVibDownload|Remediation did not succeed for [data.name]. There were errors while downloading one or more software packages. Check the VMware vSphere Update Manager network connectivity settings. |
com.vmware.vcIntegrity.HostUpgradeAgentDeployFailure | error | com.vmware.vcIntegrity.HostUpgradeAgentDeployFailure|Cannot deploy upgrade agent on host. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailConflictingVibs | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailConflictingVibs|The upgrade contains conflicting VIBs. Remove the conflicting VIBs or use Image Builder to create a custom upgrade ISO image that contains the newer versions of the conflicting VIBs, and try to upgrade again. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailDvsBreakageUnsure | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailDvsBreakageUnsure|Cannot determine whether the upgrade breaks Cisco Nexus 1000V virtual network switch feature on the host. If the host does not have the feature, you can ignore this warning. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailDvsBreaks | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailDvsBreaks|Cisco Nexus 1000V virtual network switch feature installed on the host will be removed during upgrade. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailEESXInsufficientSpaceForImage | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailEESXInsufficientSpaceForImage|Cannot create a ramdisk of size [data.expected]MB to store the upgrade image. Check if the host has sufficient memory. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailESXInsufficientSpaceForImage | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailESXInsufficientSpaceForImage|Upgrade requires at least [data.expected]MB free space on boot partition to store the upgrade image, only [data.found]MB found. Retry after freeing up sufficient space or perform a CD-based installation. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailHostHardwareMismatch | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailHostHardwareMismatch|The upgrade is not supported on the host hardware. The upgrade ISO image contains VIBs that failed the host hardware compatibility check. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailIncompatibleDPInImage | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailIncompatibleDPInImage|Cisco Nexus 1000V virtual network switch software package [data.found] in the upgrade image is incompatible with the Cisco Nexus 1000V software package [data.expected] installed on the host. Upgrading the host will remove the feature from the host. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailIncompatiblePartitionLayout | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailIncompatiblePartitionLayout|The host cannot be upgraded due to incompatible partition layout. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailIncompatiblePasswords | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailIncompatiblePasswords|The passwords cannot be migrated because the password encryption scheme is incompatible. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailIncompatibleWithDvsCP | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailIncompatibleWithDvsCP|Cisco Nexus 1000V virtual network switch software package [data.found] in the upgrade image is incompatible with the Cisco Nexus 1000V VSM. Upgrading the host will remove the feature from the host. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailInsufficientMemory | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailInsufficientMemory|Insufficient memory found on the host: [data.expected]MB required, [data.found]MB found. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailInsufficientSpaceForConfig | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailInsufficientSpaceForConfig|Upgrade requires at least [data.expected]MB free space on a local VMFS datastore, only [data.found]MB found. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailMissingDPBreaksDvsCP | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailMissingDPBreaksDvsCP|There is no Cisco Nexus 1000V virtual network switch software package in the upgrade image that is compatible with the Cisco Nexus 1000V VSM. Upgrading the host will remove the feature from the host. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailMissingDPInImage | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailMissingDPInImage|There is no Cisco Nexus 1000V virtual network switch software package in the upgrade image [data.found]. Upgrading the host will remove the feature from the host. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailNoMD5RootPassword | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailNoMD5RootPassword|The root password is not using MD5 hashing, causing it to be authenticated up to only 8 characters. For instructions on how to correct this, see VMware KB 1024500 at http://kb.vmware.com/kb/1024500. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailNoMinCpuCores | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailNoMinCpuCores|New ESXi version requires a minimum of [data.expected] processor cores. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailNoVt | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailNoVt|Processor does not support hardware virtualization or it is disabled in BIOS. Virtual machine performance may be slow. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailNonVmwareSoftware | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailNonVmwareSoftware|The software modules [data.found] found on the host are not part of the upgrade image. These modules will be removed during upgrade. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailNxbitEnabled | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailNxbitEnabled|No eXecute (NX) bit is not enabled on the host. New ESXi version requires a CPU with NX/XD bit supported and enabled. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailPendingReboot | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailPendingReboot|Host software configuration requires a reboot. Reboot the host and try upgrade again. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailPowerPathBreaks | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailPowerPathBreaks|EMC PowerPath module [data.found] installed on the host will be removed during upgrade. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailScriptInitFailed | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailScriptInitFailed|Host upgrade validity checks are not successful. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailTbootRequired | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailTbootRequired|Trusted boot is enabled on the host but the upgrade does not contain the software package esx-tboot. Upgrading the host will remove the trusted boot feature. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailUnknown | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailUnknown|The upgrade precheck script returned unknown error. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailUnsupportedConfig | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailUnsupportedConfig|Error in ESX configuration file (esx.conf). |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailUnsupportedDevices | warning | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailUnsupportedDevices|Unsupported devices [data.found] found on the host. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailUnsupportedHostVersion | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailUnsupportedHostVersion|Host version [data.found] is not supported for upgrade. |
com.vmware.vcIntegrity.HostUpgradePrecheckTestFailUnsupportedLongMode | error | com.vmware.vcIntegrity.HostUpgradePrecheckTestFailUnsupportedLongMode|Host CPU is unsupported. New ESXi version requires a 64-bit CPU with support for LAHF/SAHF instructions in long mode. |
com.vmware.vcIntegrity.HostUpgradeProgressAborted | error | com.vmware.vcIntegrity.HostUpgradeProgressAborted|Host upgrade installer stopped. |
com.vmware.vcIntegrity.HostUpgradeProgressAuth | info | com.vmware.vcIntegrity.HostUpgradeProgressAuth|Host upgrade in progress: Configuring authentication. |
com.vmware.vcIntegrity.HostUpgradeProgressBootloader | info | com.vmware.vcIntegrity.HostUpgradeProgressBootloader|Host upgrade in progress: Boot setup. |
com.vmware.vcIntegrity.HostUpgradeProgressClearpart | info | com.vmware.vcIntegrity.HostUpgradeProgressClearpart|Host upgrade in progress: Clearing partitions. |
com.vmware.vcIntegrity.HostUpgradeProgressComplete | info | com.vmware.vcIntegrity.HostUpgradeProgressComplete|Host upgrade installer completed. |
com.vmware.vcIntegrity.HostUpgradeProgressKeyboard | info | com.vmware.vcIntegrity.HostUpgradeProgressKeyboard|Host upgrade in progress: Setting keyboard. |
com.vmware.vcIntegrity.HostUpgradeProgressLanguage | info | com.vmware.vcIntegrity.HostUpgradeProgressLanguage|Host upgrade in progress: Setting language. |
com.vmware.vcIntegrity.HostUpgradeProgressMigrating | info | com.vmware.vcIntegrity.HostUpgradeProgressMigrating|Host upgrade in progress: Migrating ESX v3 configuration to ESX v4. |
com.vmware.vcIntegrity.HostUpgradeProgressMount | info | com.vmware.vcIntegrity.HostUpgradeProgressMount|Host upgrade in progress: Mounting file systems. |
com.vmware.vcIntegrity.HostUpgradeProgressNetworking | info | com.vmware.vcIntegrity.HostUpgradeProgressNetworking|Host upgrade in progress: Installing network configuration. |
com.vmware.vcIntegrity.HostUpgradeProgressPackages | info | com.vmware.vcIntegrity.HostUpgradeProgressPackages|Host upgrade in progress: Installing packages. |
com.vmware.vcIntegrity.HostUpgradeProgressPartphys | info | com.vmware.vcIntegrity.HostUpgradeProgressPartphys|Host upgrade in progress: Partitioning physical hard drives. |
com.vmware.vcIntegrity.HostUpgradeProgressPartvirt | info | com.vmware.vcIntegrity.HostUpgradeProgressPartvirt|Host upgrade in progress: Partitioning virtual hard drives. |
com.vmware.vcIntegrity.HostUpgradeProgressPostscript | info | com.vmware.vcIntegrity.HostUpgradeProgressPostscript|Host upgrade in progress: Running postinstallation script. |
com.vmware.vcIntegrity.HostUpgradeProgressRootpass | info | com.vmware.vcIntegrity.HostUpgradeProgressRootpass|Host upgrade in progress: Setting root password |
com.vmware.vcIntegrity.HostUpgradeProgressTimezone | info | com.vmware.vcIntegrity.HostUpgradeProgressTimezone|Host upgrade in progress: Setting timezone. |
com.vmware.vcIntegrity.HostUpgradeProgressUnknown | info | com.vmware.vcIntegrity.HostUpgradeProgressUnknown|Host upgrade in progress. |
com.vmware.vcIntegrity.HostUpgradeRunScriptFailure | error | com.vmware.vcIntegrity.HostUpgradeRunScriptFailure|Cannot run upgrade script on host. |
com.vmware.vcIntegrity.IncompatibleTools | error | com.vmware.vcIntegrity.IncompatibleTools|Could not install VMware vSphere Update Manager Guest Agent on {vm.name} because VMware Tools is not installed or is of an incompatible version. The required version is [data.requiredVersion] and the installed version is [data.installedVersion]. |
com.vmware.vcIntegrity.InstallAddOnUpdate | info | com.vmware.vcIntegrity.InstallAddOnUpdate|The following additional patches are included to resolve a conflict for installation on [data.entityName]: [data.message]. |
com.vmware.vcIntegrity.InstallSuggestion | info | com.vmware.vcIntegrity.InstallSuggestion|To resolve a conflict for installation on [data.entityName], the following additional patches might need to be included in the baseline: [data.message]. |
com.vmware.vcIntegrity.InstallSuggestionNotFound | info | com.vmware.vcIntegrity.InstallSuggestionNotFound|VMware vSphere Update Manager could not find patches to resolve the conflict for installation on [data.entityName]. |
com.vmware.vcIntegrity.InstallUpdate | info | com.vmware.vcIntegrity.InstallUpdate|Installation of patches [data.updateId] started on host [data.entityName]. |
com.vmware.vcIntegrity.InstallUpdateComplete | info | com.vmware.vcIntegrity.InstallUpdateComplete|Installation of patches succeeded on [data.entityName]. |
com.vmware.vcIntegrity.InstallUpdateError | error | com.vmware.vcIntegrity.InstallUpdateError|Could not install patches on [data.entityName]. |
com.vmware.vcIntegrity.LinuxOffLineScanNotSupported | error | com.vmware.vcIntegrity.LinuxOffLineScanNotSupported|Cannot scan [data.name] for patches. Scan of powered off or suspended Linux VMs is not supported. |
com.vmware.vcIntegrity.LowDiskSpace | warning | com.vmware.vcIntegrity.LowDiskSpace|VMware vSphere Update Manager is running out of storage space. Location: [data.Volume]. Available space: [data.FreeSpace]MB. |
com.vmware.vcIntegrity.MetadataCorrupted | info | com.vmware.vcIntegrity.MetadataCorrupted|Patch definition for [data.name] is corrupt. Check the logs for more details. Re-downloading patch definitions might resolve this problem. |
com.vmware.vcIntegrity.MetadataNotFound | info | com.vmware.vcIntegrity.MetadataNotFound|Patch definitions for [data.name] are missing. Download patch definitions first. |
com.vmware.vcIntegrity.NoRequiredLicense | error | com.vmware.vcIntegrity.NoRequiredLicense|There is no VMware vSphere Update Manager license for [data.name] for the required operation. |
com.vmware.vcIntegrity.NotificationCriticalInfoAlertEvent | | com.vmware.vcIntegrity.NotificationCriticalInfoAlertEvent|event.com.vmware.vcIntegrity.NotificationCriticalInfoAlertEvent.fullFormat |
com.vmware.vcIntegrity.NotificationDownloadAlertEvent | | com.vmware.vcIntegrity.NotificationDownloadAlertEvent|event.com.vmware.vcIntegrity.NotificationDownloadAlertEvent.fullFormat |
com.vmware.vcIntegrity.NotificationImportantInfoAlertEvent | | com.vmware.vcIntegrity.NotificationImportantInfoAlertEvent|event.com.vmware.vcIntegrity.NotificationImportantInfoAlertEvent.fullFormat |
com.vmware.vcIntegrity.NotificationModerateInfoAlertEvent | | com.vmware.vcIntegrity.NotificationModerateInfoAlertEvent|event.com.vmware.vcIntegrity.NotificationModerateInfoAlertEvent.fullFormat |
com.vmware.vcIntegrity.NotificationRecallAlertEvent | | com.vmware.vcIntegrity.NotificationRecallAlertEvent|event.com.vmware.vcIntegrity.NotificationRecallAlertEvent.fullFormat |
com.vmware.vcIntegrity.NotificationRecallFixAlertEvent | | com.vmware.vcIntegrity.NotificationRecallFixAlertEvent|event.com.vmware.vcIntegrity.NotificationRecallFixAlertEvent.fullFormat |
com.vmware.vcIntegrity.PXEBootedHostEvent | | com.vmware.vcIntegrity.PXEBootedHostEvent|event.com.vmware.vcIntegrity.PXEBootedHostEvent.fullFormat |
com.vmware.vcIntegrity.PackageImportEvent | | com.vmware.vcIntegrity.PackageImportEvent|event.com.vmware.vcIntegrity.PackageImportEvent.fullFormat |
com.vmware.vcIntegrity.PackageImportFailureEvent | | com.vmware.vcIntegrity.PackageImportFailureEvent|event.com.vmware.vcIntegrity.PackageImportFailureEvent.fullFormat |
com.vmware.vcIntegrity.RebootHostComplete | info | com.vmware.vcIntegrity.RebootHostComplete|Host [data.entityName] is successfully rebooted. |
com.vmware.vcIntegrity.RebootHostError | error | com.vmware.vcIntegrity.RebootHostError|Cannot reboot host [data.entityName]. |
com.vmware.vcIntegrity.RebootHostStart | info | com.vmware.vcIntegrity.RebootHostStart|Start rebooting host [data.entityName]. |
com.vmware.vcIntegrity.RebootHostWait | info | com.vmware.vcIntegrity.RebootHostWait|Waiting for host [data.entityName] to reboot. |
com.vmware.vcIntegrity.ReconfigureClusterFailedEvent | error | com.vmware.vcIntegrity.ReconfigureClusterFailedEvent|VMware vSphere Update Manager could not restore HA admission control/DPM settings for cluster {computeResource.name} to their original values. These settings have been changed for patch installation. Check the cluster settings and restore them manually. |
com.vmware.vcIntegrity.Remediate | info | com.vmware.vcIntegrity.Remediate|Remediation succeeded for [data.name]. |
com.vmware.vcIntegrity.RemediateDisconnectedHost | error | com.vmware.vcIntegrity.RemediateDisconnectedHost|Could not remediate {host.name} because the host has an invalid connection state: [data.state]. |
com.vmware.vcIntegrity.RemediateDisconnectedVm | error | com.vmware.vcIntegrity.RemediateDisconnectedVm|Could not remediate {vm.name} because the virtual machine has an invalid connection state: [data.state]. |
com.vmware.vcIntegrity.RemediateDpmDisableHost | error | com.vmware.vcIntegrity.RemediateDpmDisableHost|Could not remediate host {host.name} because its power state is invalid. The host is in standby mode and the individual VMware DPM settings of the host are set to Disabled or Manual. |
com.vmware.vcIntegrity.RemediateFailed | error | com.vmware.vcIntegrity.RemediateFailed|Remediation did not succeed for [data.name]. |
com.vmware.vcIntegrity.RemediateHostInvalidPowerState | error | com.vmware.vcIntegrity.RemediateHostInvalidPowerState|Cannot remediate the host {host.name} because its power state is [data.state]. |
com.vmware.vcIntegrity.RemediateHostOnUnsupportedHost | error | com.vmware.vcIntegrity.RemediateHostOnUnsupportedHost|Could not remediate {host.name} because it is of unsupported version [data.version]. |
com.vmware.vcIntegrity.RemediateOrphanedVm | info | com.vmware.vcIntegrity.RemediateOrphanedVm|Could not remediate orphaned VM {vm.name}. |
com.vmware.vcIntegrity.RemediateStart | info | com.vmware.vcIntegrity.RemediateStart|Remediating object [data.name]. |
com.vmware.vcIntegrity.RemediateVmOnUnsupportedHost | error | com.vmware.vcIntegrity.RemediateVmOnUnsupportedHost|Could not remediate {vm.name} because host {host.name} is of unsupported version [data.version]. |
com.vmware.vcIntegrity.RemediationStatusEvent | info | com.vmware.vcIntegrity.RemediationStatusEvent|Current progress of remediation: [data.noOfSucceededHosts] hosts completed successfully, [data.noOfFailedHosts] hosts completed with errors, [data.noOfHostsBeingRemediatedCurrently] hosts are being remediated, [data.noOfWaitingHosts] hosts are waiting to start remediation, and [data.noOfRetryHosts] hosts could not enter maintenance mode and are waiting to retry. |
com.vmware.vcIntegrity.Scan | info | com.vmware.vcIntegrity.Scan|Successfully scanned [data.name]. |
com.vmware.vcIntegrity.ScanCancelled | error | com.vmware.vcIntegrity.ScanCancelled|Scanning of [data.name] is canceled by user. |
com.vmware.vcIntegrity.ScanDisconnectedHost | error | com.vmware.vcIntegrity.ScanDisconnectedHost|Could not scan {host.name} because the host has an invalid connection state: [data.state]. |
com.vmware.vcIntegrity.ScanDisconnectedVm | error | com.vmware.vcIntegrity.ScanDisconnectedVm|Could not scan {vm.name} because the virtual machine has an invalid connection state: [data.state]. |
com.vmware.vcIntegrity.ScanDpmDisableHost | error | com.vmware.vcIntegrity.ScanDpmDisableHost|Could not scan host {host.name} because its power state is invalid. The host is in standby mode and the individual VMware DPM settings of the host are set to Disabled or Manual. |
com.vmware.vcIntegrity.ScanFailed | error | com.vmware.vcIntegrity.ScanFailed|Could not scan [data.name]. |
com.vmware.vcIntegrity.ScanHostInvalidPowerState | error | com.vmware.vcIntegrity.ScanHostInvalidPowerState|Cannot scan the host {host.name} because its power state is [data.state]. |
com.vmware.vcIntegrity.ScanHostOnUnsupportedHost | error | com.vmware.vcIntegrity.ScanHostOnUnsupportedHost|Could not scan {host.name} for patches because it is of unsupported version [data.version]. |
com.vmware.vcIntegrity.ScanMissingUpdate | warning | com.vmware.vcIntegrity.ScanMissingUpdate|Found a missing patch: [data.message] when scanning [data.name]. Re-downloading patch definitions might resolve this problem. |
com.vmware.vcIntegrity.ScanOrphanedVm | info | com.vmware.vcIntegrity.ScanOrphanedVm|Could not scan orphaned VM {vm.name}. |
com.vmware.vcIntegrity.ScanStart | info | com.vmware.vcIntegrity.ScanStart|Scanning object [data.name]. |
com.vmware.vcIntegrity.ScanUnsupportedVolume | warning | com.vmware.vcIntegrity.ScanUnsupportedVolume|{vm.name} contains an unsupported volume [data.volumeLabel]. Scan results for this VM might be incomplete. |
com.vmware.vcIntegrity.ScanVmOnUnsupportedHost | error | com.vmware.vcIntegrity.ScanVmOnUnsupportedHost|Could not scan {vm.name} because host {host.name} is of unsupported version [data.version]. |
com.vmware.vcIntegrity.SequentialRemediateFailedEvent | error | com.vmware.vcIntegrity.SequentialRemediateFailedEvent|An error occured during the sequential remediation of hosts in cluster {computeResource.name}. Check the related events for more details. |
com.vmware.vcIntegrity.SkipSuspendedVm | info | com.vmware.vcIntegrity.SkipSuspendedVm|Suspended VM {vm.name} has been skipped. |
com.vmware.vcIntegrity.Stage | info | com.vmware.vcIntegrity.Stage|Staging succeeded for [data.name]. |
com.vmware.vcIntegrity.StageDisconnectedHost | error | com.vmware.vcIntegrity.StageDisconnectedHost|Could not stage patches to {host.name} because the host has an invalid connection state: [data.state]. |
com.vmware.vcIntegrity.StageDpmDisableHost | error | com.vmware.vcIntegrity.StageDpmDisableHost|Could not stage patches to host {host.name} because its power state is invalid. The host is in standby mode and the individual VMware DPM settings of the host are set to Disabled or Manual. |
com.vmware.vcIntegrity.StageFailed | error | com.vmware.vcIntegrity.StageFailed|Staging did not succeed for [data.name][data.message]. |
com.vmware.vcIntegrity.StageHostInvalidPowerState | error | com.vmware.vcIntegrity.StageHostInvalidPowerState|Cannot stage patches to the host {host.name} because its power state is [data.state]. |
com.vmware.vcIntegrity.StageHostOnUnsupportedHost | error | com.vmware.vcIntegrity.StageHostOnUnsupportedHost|Could not stage patches to {host.name} because it is of unsupported version [data.version]. |
com.vmware.vcIntegrity.StageStart | info | com.vmware.vcIntegrity.StageStart|Staging patches to host [data.name]. |
com.vmware.vcIntegrity.StageUpdate | info | com.vmware.vcIntegrity.StageUpdate|Started staging of patches [data.updateId] on [data.entityName]. |
com.vmware.vcIntegrity.StageUpdateComplete | info | com.vmware.vcIntegrity.StageUpdateComplete|Staging of patch to [data.entityName] succeeded. |
com.vmware.vcIntegrity.StageUpdateError | error | com.vmware.vcIntegrity.StageUpdateError|Cannot stage patch [data.updateId] to [data.entityName]. |
com.vmware.vcIntegrity.SysprepDisabled | info | com.vmware.vcIntegrity.SysprepDisabled|Sysprep is disabled during the remediation. |
com.vmware.vcIntegrity.SysprepEnabled | info | com.vmware.vcIntegrity.SysprepEnabled|Sysprep settings are restored. |
com.vmware.vcIntegrity.SysprepHandleFailure | error | com.vmware.vcIntegrity.SysprepHandleFailure|Cannot access the sysprep settings for VM {vm.name}. Retry the operation after disabling sysprep for the VM. |
com.vmware.vcIntegrity.SysprepNotFound | error | com.vmware.vcIntegrity.SysprepNotFound|Cannot locate the sysprep settings for VM {vm.name}. For Windows 7 and Windows 2008 R2, offline VM remediation is supported only if the system volume is present in the primary disk partition. Retry the operation after disabling sysprep for the VM. |
com.vmware.vcIntegrity.ToolsRemediate | info | com.vmware.vcIntegrity.ToolsRemediate|VMware Tools upgrade succeeded for [data.name]. |
com.vmware.vcIntegrity.ToolsRemediateFailed | error | com.vmware.vcIntegrity.ToolsRemediateFailed|VMware Tools upgrade did not succeed for [data.name]. |
com.vmware.vcIntegrity.ToolsScan | info | com.vmware.vcIntegrity.ToolsScan|Successfully scanned [data.name] for VMware Tools upgrades. |
com.vmware.vcIntegrity.ToolsScanFailed | error | com.vmware.vcIntegrity.ToolsScanFailed|Could not scan [data.name] for VMware Tools upgrades. |
com.vmware.vcIntegrity.ToolsScanInstallNotSupported | warning | com.vmware.vcIntegrity.ToolsScanInstallNotSupported|VMware Tools is not installed on [data.name]. VMware vSphere Update Manager supports upgrading only an existing VMware Tools installation. |
com.vmware.vcIntegrity.ToolsUpgradeRemediateSkippedOnHost | warning | com.vmware.vcIntegrity.ToolsUpgradeRemediateSkippedOnHost|VMware Tools upgrade was not performed on {vm.name}. VMware Tools upgrade is supported only for VMs that run on ESX/ESXi 4.0 and higher. VMware Tools upgrade is not supported for virtual appliances. |
com.vmware.vcIntegrity.ToolsUpgradeScanSkippedOnHost | warning | com.vmware.vcIntegrity.ToolsUpgradeScanSkippedOnHost|VMware Tools upgrade scan was not performed on {vm.name}. VMware Tools upgrade scan is supported only for VMs that run on ESX/ESXi 4.0 and higher. VMware Tools upgrade scan is not supported for virtual appliances. |
com.vmware.vcIntegrity.UnsupportedHostRemediateSpecialVMEvent | error | com.vmware.vcIntegrity.UnsupportedHostRemediateSpecialVMEvent|The host [data.name] has a VM [data.vm] with VMware vSphere Update Manager or VMware vCenter Server installed. The VM must be moved to another host for the remediation to proceed. |
com.vmware.vcIntegrity.UnsupportedLinuxAction | warning | com.vmware.vcIntegrity.UnsupportedLinuxAction|Action is not supported for Linux VM/VA {vm.name}. VMware Tools is not installed or the machine cannot start. |
com.vmware.vcIntegrity.UnsupportedOs | warning | com.vmware.vcIntegrity.UnsupportedOs|Scan or remediation is not supported on [data.name] because of unsupported OS [data.os]. |
com.vmware.vcIntegrity.UnsupportedPXEBootHost | info | com.vmware.vcIntegrity.UnsupportedPXEBootHost|Scanning, remediation, and staging are not supported on PXE booted ESXi hosts. |
com.vmware.vcIntegrity.UnsupportedSpecialVMEvent | error | com.vmware.vcIntegrity.UnsupportedSpecialVMEvent|VM [data.name] has either VMware vSphere Update Manager or VMware vCenter Server installed. This VM will be ignored for scan and remediation. |
com.vmware.vcIntegrity.UnsupportedVaAction | warning | com.vmware.vcIntegrity.UnsupportedVaAction|Action is not supported for offline or suspended virtual appliance {vm.name}. |
com.vmware.vcIntegrity.VAAutoUpdateOn | error | com.vmware.vcIntegrity.VAAutoUpdateOn|Auto update is set to ON for virtual appliance [data.name]. |
com.vmware.vcIntegrity.VADiscovery | info | com.vmware.vcIntegrity.VADiscovery|Successfully discovered virtual appliance [data.name]. |
com.vmware.vcIntegrity.VADiscoveryFailed | error | com.vmware.vcIntegrity.VADiscoveryFailed|Could not discover virtual appliance [data.name]. |
com.vmware.vcIntegrity.VADownloadGenericFailure | error | com.vmware.vcIntegrity.VADownloadGenericFailure|Could not download virtual appliance upgrade metadata. |
com.vmware.vcIntegrity.VADownloadInvalidUrl | error | com.vmware.vcIntegrity.VADownloadInvalidUrl|[data.name] is not a valid virtual appliance download URL. |
com.vmware.vcIntegrity.VADownloadMetadataFailure | error | com.vmware.vcIntegrity.VADownloadMetadataFailure|Could not download virtual appliance upgrade metadata for [data.name]. |
com.vmware.vcIntegrity.VADownloadSuccess | info | com.vmware.vcIntegrity.VADownloadSuccess|Successfully downloaded virtual appliance upgrade metadata. |
com.vmware.vcIntegrity.VARepositoryAddressNotSet | error | com.vmware.vcIntegrity.VARepositoryAddressNotSet|No repository address is set for virtual appliance [data.name]. The appliance does not support updates by vCenter Server. |
com.vmware.vcIntegrity.VAScan | info | com.vmware.vcIntegrity.VAScan|Successfully scanned [data.name] for VA upgrades. |
com.vmware.vcIntegrity.VAScanFailed | error | com.vmware.vcIntegrity.VAScanFailed|Could not scan [data.name] for VA upgrades. |
com.vmware.vcIntegrity.VMHardwareUpgradeRemediate | info | com.vmware.vcIntegrity.VMHardwareUpgradeRemediate|Virtual Hardware upgrade succeeded for [data.name]. |
com.vmware.vcIntegrity.VMHardwareUpgradeRemediateFailed | error | com.vmware.vcIntegrity.VMHardwareUpgradeRemediateFailed|Could not perform Virtual Hardware upgrade on [data.name]. |
com.vmware.vcIntegrity.VMHardwareUpgradeRemediateSkippedOnHost | warning | com.vmware.vcIntegrity.VMHardwareUpgradeRemediateSkippedOnHost|Virtual Hardware upgrade was not performed for {vm.name}. Virtual Hardware upgrade is supported only for VMs that run on ESX/ESXi 4.0 and higher. Virtual Hardware upgrade is not supported for virtual appliances. |
com.vmware.vcIntegrity.VMHardwareUpgradeScan | info | com.vmware.vcIntegrity.VMHardwareUpgradeScan|Successfully scanned [data.name] for Virtual Hardware upgrades. |
com.vmware.vcIntegrity.VMHardwareUpgradeScanFailed | error | com.vmware.vcIntegrity.VMHardwareUpgradeScanFailed|Could not scan [data.name] for Virtual Hardware upgrades. |
com.vmware.vcIntegrity.VMHardwareUpgradeScanSkippedOnHost | warning | com.vmware.vcIntegrity.VMHardwareUpgradeScanSkippedOnHost|Virtual Hardware upgrade scan was not performed for {vm.name}. Virtual Hardware upgrade scan is supported only for VMs that run on ESX/ESXi 4.0 and higher. Virtual Hardware upgrade scan is not supported for virtual appliances. |
com.vmware.vcIntegrity.VMHardwareUpgradeToolsNotInstalled | error | com.vmware.vcIntegrity.VMHardwareUpgradeToolsNotInstalled|Virtual Hardware upgrade did not succeed for {vm.name}, because VMware Tools is not installed. To upgrade virtual hardware, VMware Tools must be the latest version. |
com.vmware.vcIntegrity.VMHardwareUpgradeToolsNotLatest | error | com.vmware.vcIntegrity.VMHardwareUpgradeToolsNotLatest|Virtual Hardware upgrade did not succeed for {vm.name}, because VMware Tools is not the latest version supported by the host. To upgrade virtual hardware, VMware Tools must be the latest version. |
com.vmware.vcIntegrity.VMHardwareUpgradeToolsUnknown | error | com.vmware.vcIntegrity.VMHardwareUpgradeToolsUnknown|Virtual Hardware upgrade did not succeed for {vm.name}, because VMware Tools state is unknown. To upgrade virtual hardware, VMware Tools must be the latest version. |
com.vmware.vcIntegrity.VMHardwareUpgradeToolsUnmanaged | error | com.vmware.vcIntegrity.VMHardwareUpgradeToolsUnmanaged|Virtual Hardware upgrade did not succeed for {vm.name}, because VMware Tools state is not managed by VMware vSphere. To upgrade virtual hardware, VMware Tools must be the latest version. |
com.vmware.vcIntegrity.VMToolsAutoUpgradeUnsupported | error | com.vmware.vcIntegrity.VMToolsAutoUpgradeUnsupported|The version of VMware Tools installed in {vm.name} does not support automatic upgrade. Upgrade VMware Tools manually. |
com.vmware.vcIntegrity.VMToolsNotRunning | error | com.vmware.vcIntegrity.VMToolsNotRunning|Error while waiting for VMware Tools to respond. Verify that VMware Tools is running in VM {vm.name}. |
com.vmware.vcIntegrity.VibPrerequisitesMissingForInstall | warning | com.vmware.vcIntegrity.VibPrerequisitesMissingForInstall|Patch [data.inputBulletin] was excluded from the remediation because its prerequisite [data.missingPrereq] is neither installed on the host nor included in the baseline. Include the prerequisites in a Patch or Extension baseline and retry the remediation. You can also add the baselines to a baseline group for convenience and perform the remediation. |
com.vmware.vcIntegrity.VibPrerequisitesMissingForStage | warning | com.vmware.vcIntegrity.VibPrerequisitesMissingForStage|Patch [data.inputBulletin] was excluded from the stage operation because its prerequisite [data.missingPrereq] is neither installed on the host nor included in the baseline. Include the prerequisites in a Patch or Extension baseline and retry the stage operation. You can also add the baselines to a baseline group for convenience and perform the stage operation. |
com.vmware.vcIntegrity.VmDevicesRestoreFailedEvent | error | com.vmware.vcIntegrity.VmDevicesRestoreFailedEvent|VMware vSphere Update Manager could not restore the original removable device connection settings for all VMs in cluster {computeResource.name}. These settings have been changed for patch installation. You can manually restore the settings for the VMs. |
com.vmware.vcIntegrity.VmMigrationFailedEvent | error | com.vmware.vcIntegrity.VmMigrationFailedEvent|Cannot migrate VM {vm.name} from [data.srcHost] to [data.destHost]. |
com.vmware.vcIntegrity.VmPowerRestoreFailedEvent | error | com.vmware.vcIntegrity.VmPowerRestoreFailedEvent|VMware vSphere Update Manager could not restore the original power state for all VMs in cluster {computeResource.name}. These settings have been changed for patch installation. You can manually restore the original power state of the VMs. |
com.vmware.vcIntegrity.VmotionCompatibilityCheckFailedEvent | error | com.vmware.vcIntegrity.VmotionCompatibilityCheckFailedEvent|Cannot check compatibility of the VM {vm.name} for migration with vMotion to host [data.hostName]. |
com.vmware.vshield.dlp.VmScanEnd | info | com.vmware.vshield.dlp.VmScanEnd|vShield Data Security scan ended on the VirtualMachine. |
com.vmware.vshield.dlp.VmScanStart | info | com.vmware.vshield.dlp.VmScanStart|vShield Data Security scan started on the VirtualMachine. |
com.vmware.vshield.edge.BadState | error | com.vmware.vshield.edge.BadState|vShield Manager found vShield Edge in bad state. Needs a force sync. |
com.vmware.vshield.edge.VixagentNotConnectedToVc | error | com.vmware.vshield.edge.VixagentNotConnectedToVc|vShield Edge Communication Agent not connected to vCenter Server. |
com.vmware.vshield.edge.appliance.BadState | error | com.vmware.vshield.edge.appliance.BadState|vShield Edge is in Bad State. Needs a force sync. |
com.vmware.vshield.edge.appliance.CommunicationFailed | warning | com.vmware.vshield.edge.appliance.CommunicationFailed|Failed to communicate with the Vshield Edge VM. |
com.vmware.vshield.edge.appliance.ForceSynced | info | com.vmware.vshield.edge.appliance.ForceSynced|vShield Edge was force synced |
com.vmware.vshield.edge.appliance.MovedInVcInventory | info | com.vmware.vshield.edge.appliance.MovedInVcInventory|vShield Edge appliance moved in the vCenter inventory. This might have been because of activities like vMotion, DRS. |
com.vmware.vshield.edge.gateway.Created | info | com.vmware.vshield.edge.gateway.Created|Created vShield Edge Gateway with name : ''{0}'' and id : ''{1}'' |
com.vmware.vshield.edge.gateway.Deleted | info | com.vmware.vshield.edge.gateway.Deleted|Deleted vShield Edge Gateway with id : {0} |
com.vmware.vshield.edge.gateway.HealthcheckNoPulse | error | com.vmware.vshield.edge.gateway.HealthcheckNoPulse|None of the VShield Edge VMs found in serving state. There is a possibility of network disruption. |
com.vmware.vshield.edge.gateway.HealthcheckRecovery | info | com.vmware.vshield.edge.gateway.HealthcheckRecovery|vShield Edge Gateway has recovered and now responding to health check. |
com.vmware.vshield.edge.gateway.NoActiveOrSelfStateAppliance | error | com.vmware.vshield.edge.gateway.NoActiveOrSelfStateAppliance|None of the Vshield Edge VMs found in serving state. There is a possibility of network disruption. |
com.vmware.vshield.edge.gateway.Redeployed | info | com.vmware.vshield.edge.gateway.Redeployed|Redeployed vShield Edge Gateway with id : {0} |
com.vmware.vshield.edge.gateway.SizeChanged | info | com.vmware.vshield.edge.gateway.SizeChanged|Appliance size for vShield Edge Gateway with id : {0} changed to {1} |
com.vmware.vshield.edge.highavailability.ChannelDisconnected | warning | com.vmware.vshield.edge.highavailability.ChannelDisconnected|vShield Edge HighAvailability communication channel is disconnected from peer node |
com.vmware.vshield.edge.highavailability.StateSelf | info | com.vmware.vshield.edge.highavailability.StateSelf|vShield Edge HighAvailability is disabled. VM has moved to SELF state. |
com.vmware.vshield.edge.ipsec.ChannelStateDown | info | com.vmware.vshield.edge.ipsec.ChannelStateDown|IPsec Channel from localIp : {0} to peerIp : {1} changed the state to down |
com.vmware.vshield.edge.ipsec.ChannelStateUp | info | com.vmware.vshield.edge.ipsec.ChannelStateUp|IPsec Channel from localIp : {0} to peerIp : {1} changed the state to up |
com.vmware.vshield.edge.ipsec.ChannelWrongState | info | com.vmware.vshield.edge.ipsec.ChannelWrongState|IPsec Channel from localIp : {0} to peerIp : {1} changed to a wrong state |
com.vmware.vshield.edge.ipsec.TunnelStateDown | info | com.vmware.vshield.edge.ipsec.TunnelStateDown|IPsec Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to down |
com.vmware.vshield.edge.ipsec.TunnelStateUp | info | com.vmware.vshield.edge.ipsec.TunnelStateUp|IPsec Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to up |
com.vmware.vshield.edge.ipsec.TunnelWrongState | info | com.vmware.vshield.edge.ipsec.TunnelWrongState|IPsec Tunnel from localSubnet : {0} to peerSubnet : {1} changed to a wrong state |
com.vmware.vshield.edge.loadbalancer.BackendServerDown | info | com.vmware.vshield.edge.loadbalancer.BackendServerDown|LoadBalancer virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to down |
com.vmware.vshield.edge.loadbalancer.BackendServerUp | info | com.vmware.vshield.edge.loadbalancer.BackendServerUp|LoadBalancer virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to up |
com.vmware.vshield.edge.loadbalancer.BackendServerWrongState | info | com.vmware.vshield.edge.loadbalancer.BackendServerWrongState|LoadBalancer virtualServer : {0} Protocol : {1} serverIp : {2} changed to a wrong state |
com.vmware.vshield.edge.upgrade.ToSpockGateway | info | com.vmware.vshield.edge.upgrade.ToSpockGateway|vShield Edge 5.0 successfully upgraded to vShield Edge 5.1.2 gateway with id : {0} and name : {1} |
com.vmware.vshield.vsm.usermgmt.vcUserRoleAssigned | info | com.vmware.vshield.vsm.usermgmt.vcUserRoleAssigned|User Role assigned for vCenter User on vSM. |
com.vmware.vshield.vsm.usermgmt.vcUserRoleUnassigned | info | com.vmware.vshield.vsm.usermgmt.vcUserRoleUnassigned|User Role unassigned for vCenter User on vSM. |
com.vmware.vxlan.instance.initfail | | com.vmware.vxlan.instance.initfail|event.com.vmware.vxlan.instance.initfail.fullFormat |
com.vmware.vxlan.instance.notexist | | com.vmware.vxlan.instance.notexist|event.com.vmware.vxlan.instance.notexist.fullFormat |
com.vmware.vxlan.mcastgroup.joinfail | | com.vmware.vxlan.mcastgroup.joinfail|event.com.vmware.vxlan.mcastgroup.joinfail.fullFormat |
com.vmware.vxlan.port.initfail | | com.vmware.vxlan.port.initfail|event.com.vmware.vxlan.port.initfail.fullFormat |
com.vmware.vxlan.vmknic.initfail | | com.vmware.vxlan.vmknic.initfail|event.com.vmware.vxlan.vmknic.initfail.fullFormat |
com.vmware.vxlan.vmknic.ipchange | | com.vmware.vxlan.vmknic.ipchange|event.com.vmware.vxlan.vmknic.ipchange.fullFormat |
com.vmware.vxlan.vmknic.noip | | com.vmware.vxlan.vmknic.noip|event.com.vmware.vxlan.vmknic.noip.fullFormat |