Original Publication Date: May 8, 2012
Server Automation release notes contain information for users who are familiar with the installation and maintenance of Server Automation, Storage Visibility and Automation, SE Connector, Application Deployment Manager, DMA integration, and SA-OO integration. The notes contain information that is not included in books or online Help.
Note: The latest information for 9.13 can be found in the latest version of these release notes, located at: http://support.openview.hp.com/selfsolve/manuals.
This document contains the following information:
What's New in This Release |
Installation |
Known Issues |
Fixed Issues |
Deprecated Features |
Documentation Information |
HP Software Support |
Legal Notices |
This section describes new functionality and other relevant release-specific information.
For information about what was new in previous releases, please see the corresponding release notes at:
http://support.openview.hp.com/selfsolve/manuals.
Make sure you follow these steps before you install SA 9.13:
The SA 9.13 release provides qualified managed support for the following:
Platforms:
Core Platforms:
Features:
Integrations:
Network Automation (9.20)
For more support and compatibility information, download the HP Server Automation Support and Compatibility Matrix for this release from the HP Software Support Online web site at: http://support.openview.hp.com/selfsolve/manuals.
For a list of supported operating systems and platforms for Storage Visibility and Automation Managed Servers, SE Connector, SAN Arrays, Fibre Channel Adapters, SAN Switches, File System Software, Database Support, and Storage Essentials Compatibility, see the Storage Visibility and Automation Support and Compatibility Matrix.
The Remediation Performance and Scalability component is missing from 9.13. There will be a hotfix that resolves the issue.
Since 9.11, SA Agents have used Python 2.4.4.
After you upgrade to SA 9.13, you should also upgrade to the 9.13 SA Agents on each Managed Server in the facility.
You can continue to use non-upgraded agents, but you will not benefit from the following new features in this release:
For more information on the Agent Upgrade tool, on installing the agent, and on bringing servers under SA management, see the SA User Guide: Server Automation.
This capability enables you to create separate “sandboxes” for Configuration File component deployment. This is useful if you want files defined in Configuration File components to only be installed in controlled file system locations.
You can now copy and paste application components into compatible tiers in the same application or a different application.
You can now run workflows that are designed to execute different steps on different targets. An example of a bridged execution workflow is a Database Refresh workflow that extracts the contents of a database on one target (the Source) and creates a new database with the same contents on another target (the Destination). This type of workflow is useful when you want to clone a database—for example, to move it from a traditional IT infrastructure location into a private cloud, or to populate a test database with real production data.
For more information, see "Bridged Execution Workflows" in the User Guide: Database and Middleware Automation.
The Database and Middleware web interface now displays the deployment name on the History page after a workflow is executed. This makes your automation audit trail more robust by providing a record of the specific deployment used for each workflow execution. The deployment name is listed immediately below the workflow name.
NOTE: The deployment name is only displayed for deployments that are executed after SA 9.13 is installed.
Updates pertaining specifically to the OO-SA integration (Server Automation operations performed within Operations Orchestration) are delivered via the HP Live Network at https://hpln.hp.com/.
The SA-OO Integration feature is not available until you import the required OO SDK Client Certificate into SA. SA 9.13 is bundled with two OO SDK certificates that apply to different OO versions, as shown in the following table. Before you import the certificate into SA, make sure it is compatible with your OO version.
OO SDK Certificate |
OO Version |
---|---|
oocert2007 |
OO 7.5x, 7.6x, 9.0 |
oocert2011 | OO 9.02.0002 or later |
To integrate the new certificate, you must perform the steps in the procedure detailed below.
Note: If your architecture includes a master core and one or more secondary cores, follow the steps in this section for the master core and for each of the secondary cores. Similarly, if your SA computer has a sliced-core installation with one or more slices, repeat the steps for each slice.
/etc/init.d/opsware-sas stop twist
In previous SA versions, if you created OGFS scripts and added them to your build plans, the scripts were executed in the file system in the following directory:
/opsw/.Server.ID/<mid>/files/<user>
Now, the scripts are executed in this new location:
/opsw/.Server.ID/<mid>
As a result of this change, you must add the following line to your build-plan scripts to avoid compatibility issues with user scripts:
cd files/root (Linux)
cd files/Administrator ( Windows)
This release supports Oracle 11.2.0.2. For more information, see the white paper: Oracle RAC Support, which you can download from the SA 9.13 release section at: http://support.openview.hp.com/selfsolve/manuals.
As of SA 9.10, OS Provisioning Build Plan (OSBP) content is uploaded into the SA Core during installation and/or upgrade.
After installing or upgrading to SA 9.13, you must not download OSBP content from the HP Live Network (HPLN). Any automatic jobs you have created to download OSBP content from HPLN must be disabled. If you download OSBP content from HPLN after installing SA 9.13, you will overwrite the SA 9.10 OSBP content that was uploaded during install/upgrade with older SA 9.0 OSBP content, which will cause errors when you attempt to run OS Build Plan jobs under SA 9.13.
The HPSA Patch Importer for Oracle Enterprise Linux allows users to import packages for the subscribed channels from the Oracle Unbreakable Linux Network (ULN) and automatically create the corresponding software policies for each imported channel in HPSA. It can be run from the command line manually, or can be part of a cron job that performs the import on a recurring basis.
Note: For information about using the HPSA Patch Importer, download the whitepaper, SA Patch Importer for Oracle Enterprise Linux, from the HP Software Support Online at: http://support.openview.hp.com/selfsolve/manuals.
Duplicate patches can be inadvertently created in the SA database if you import the Microsoft Patch Supplement (MPS) and then run the SA Patch Import process using the Microsoft Offline Patch Catalog (wsusscn2.cab). These duplicates can cause conflicts during remediation and compliance checks.
To resolve this, a one-time de-duplication procedure is available that enables you to remove the duplicates and eliminate the source of these conflicts to prevent future duplication. Step-by-step instructions are provided in the whitepaper, Resolving Conflicts between SA Patching and the MS Patch Supplement, which is available on the HP Software Support Online site under the 9.13 Server Automation release.
IMPORTANT NOTE: The de-duplication procedure varies based on your version of SA. For this reason, there will be multiple versions of this whitepaper, each providing release-specific instructions.
To find the whitepaper:
Find filename: SA9.13_Windows_PatchingPatch_Supplement_wp.pdf
You can also do the same search directly from the SSO Product Manuals site: http://support.openview.hp.com/selfsolve/manuals
After SA 9.13 was released, it was discovered that the Remediation Performance and Scalability component introduced in SA 9.11 is missing from SA 9.13. Although there is no loss of SA functionality without this component, you might see reduced performance levels (as compared to SA 9.05/9.06/9.11/9.12) for both remediation and patching jobs against core-connected servers.
To check for the missing component, run the following command:
/etc/init.d/opsware-sas list | grep tsunami
If nothing is displayed, then the component is missing from the core.
A hotfix has been created to restore the missing component and is now available from HP Support. This hotfix must be installed to complete the 9.13 installation.
This table shows browser support information for this release.
Windows 2003 | Windows XP | Windows Vista | Windows 7 | Windows 2008 | |
---|---|---|---|---|---|
Chrome | x |
x |
x |
x |
x |
Firefox 3.0 | x |
x |
x |
x |
|
Firefox 3.5 | x |
x |
x |
x |
x |
Firefox 3.6 | x |
x |
x |
x |
x |
IE 6.0 | x |
x |
x |
||
IE 7.0 | x |
x |
x |
x |
|
IE 8.0 | x |
x |
x |
x |
x |
This section lists the operating-systems supported on the SA Client.
Note: This information applies to all Unix and Linux platforms.
To install the SA Command-line Interface (OCLI) on a Solaris 11 server, issue the following command:
<package_name> -d <installation_directory>
No other steps to setup the environment are necessary.
When using the SA Command-line Interface (OCLI) on a Solaris 11 server, the oupload and odownload scripts are run from the newly created /bin directory:
<installation_directory>/bin/oupload.sh
<installation_directory>/bin/odownload.sh
To support Solaris 11, the following classes were added to the UAPI:
Oracle Solaris 11 uses IPS packages to deliver software and software updates. IPS (Image Packaging System) is a network-based package management system that is used for the entire software lifecycle, including package installation, upgrade and removal.
Server Automation's Solaris 11 platform support for server patching allows you to update your managed servers to the latest versions of existing software without installing new software. This is a powerful way to keep your system up to date in an environment that no longer supports explicit patch units.
Solaris 11 patching support leverages the existing Solaris patching functionality, with a few differences to adapt to the new Solaris IPS package delivery structure. Additionally, there are setup requirements for setting up the initial IPS Package database.
Note: The SA 9.13 Solaris 11 Patching Whitepaper provides detailed instructions on setting up and maintaining the IPS Package database. This document also describes the Solaris 11 patch policy and remediation differences. This whitepaper is available via HP's Software Support Online at http://support.openview.hp.com/selfsolve/manuals.
Solaris 11 introduces Role-based Access Control (RBAC), which affects the functionality of SA Server Modules (SMOs). SMOs rely on the root user to get information from the managed servers. In previous versions of Solaris, the root account is created as a user by default, with the option to have it as a role. In Solaris 11, the root account is created as a role by default; not as a user.
In order to enable the SMO functionality, the system administrator must modify the root account (root) to make it a user with regular login capability. The best method for doing this is to run an ad hoc script on the managed server.
To enable root for regular login capability:
After the job completes the user will be able to run the SMOs on the selected servers.
Note: To change the root account to a role again, complete the same procedure but enter ‘usermod -K type=role root’ in the Script Contents.
Oracle Solaris 11 provides a service called the Automated Installer to handle OS Provisioning of bare-metal servers. The Oracle Solaris Automated Installer requires a new package type called IPS.
Solaris 11 OS Provisioning is somewhat different than the procedure used to provision Oracle Solaris 10.
The Oracle Solaris 11 Automated Installer no longer uses the Solaris JumpStart mechanism. Instead, SA components have been adapted to use the new Solaris 11 provisioning mechanism.
Installation MediaThe installation media represents an IPS repository.
In order to obtain this repository:
- Go to http://www.oracle.com/technetwork/server-storage/solaris11/downloads/index.html?ssSourceSiteId=ocomen and download the images under "Oracle Solaris 11 11/11 Repository Image". There are two ISO images: Part A SPARC, x86 (3.3 GB), and Part B SPARC, x86 (3.1 GB).
- Follow the instructions on the download page to concatenate the two images into a single full image.
- Mount the image and copy it to the SA Media server under the directory: /media/opsware/sunos
If you need to perform OS Provisioning on the same core for both oracle Solaris x86 and Oracle Sun SPARC, create a link to the repository on the SA Media Server. This link is required because a Media Resource Locator (MRL) cannot be created using the same path for two operating systems.
For example:
# cd /media/opsware/sunos/
# ls -lsa
4 lrwxrwxrwx 1 root root 15 Mar 26 14:39 solaris11_link -> solaris11_repo/
8 drwxr-xr-x 3 root root 4096 Mar 26 14:33 solaris11_repo
The Oracle Solaris Automated Installer uses two response files:
Note: Oracle documentation about creating ai.xml and sc.xml files can be found at: http://docs.oracle.com/cd/E23824_01/html/E21798/
Since SA can handle only one response file per operating system and the Solaris Automated Installer requires two files for provisioning Solaris 11, SA provides a script, join_ai_sc.py, located in /opt/opsware/buildscripts/solaris/tools that can join ai.xml and sc.xml into a single file. The output file this script creates is then used as the response file for Oracle Solaris 11 provisioning.
Limitations
- Choosing Between Solaris 10 SPARC and Solaris 11 SPARC
Because Oracle Sun SPARC provisioning uses the bootp protocol, there is no clear way to choose between Solaris 10 SPARC and Solaris 11 SPARC provisioning. Therefore, you must use the /opt/opsware/boot/jumpstart/tools/switch_solaris_sparc_miniroot command to tell SA which version of Solaris to choose.
Note: To perform both Solaris 11 and Solaris 10 OS provisioning, create an OS Sequence with an attached Software Policy containing a script that runs the switch_solaris_sparc_miniroot command on the SA Core. The script must contain required parameters automating the miniroot change.
- Enabling Oracle Solaris 11 x86 with the Manage Boot Client (MBC)
By default, the Manage Boot Client (MBC) utility can only automatically provision Solaris 10 X86. To enable Oracle Solaris 11 X86 in MBC, run the tool /opt/opsware/boot/js-x86/tools/switch_solaris_x86_default_pxe and choose the desired default option from the Solaris Preboot Execution Environment (PXE) menu. Once Solaris 11 is set as the default, all machines that will boot into Server Pool will boot the Solaris 11 X86 miniroot. Also all MBC jobs will use the Solaris 11 X86 miniroot, causing MBC jobs for Solaris 10 X86 to fail.
Customizations
The new Oracle Solaris 11 OS provisioning mechanism allows the use of custom attributes to simplify system configuration.
These are:
Custom Attribute Name | Use | Example Values |
---|---|---|
mrl | The MRL created by the import_media script. This is useful when you want to use an external repository. | http://pkg.oracle.com/solaris/release |
http_proxy |
Specify when you use HTTP repositories. | http://PROXY_IP:PROXY_PORT |
root_password | Specifies the root password in hash format. | Use a password from the /etc/shadow directory on the Solaris 11 machine. |
enable_root_user | Solaris 11 allows root to be defined as a role in the sc.xml file. Presence of this custom attribute will trigger the user creation, even if 'no' value is set. |
yes |
hostname | Specifies the hostname of the machine. | mysolaris11 |
SA briefly stops the VMware ESXi firewall in order to communicate the VM's IP Address, so that it can be managed by the SA core. SA restarts the firewall immediately afterwards.
If you do not want SA to stop and start the firewall, use static OS Provisioning instead.
For more information on any of these topics, see the SA UG: OS Provisioning and SA online help.
Support for OS Build Plans
HP ProLiant Gen8 models have two embedded service operating systems available: Linux and Windows Preinstallation Environment (Windows PE). SA can use either the Integrated Lights-Out (iLO) functionality of these embedded systems, or Preboot Execution Environment (PXE) booting.
To use the embedded systems' iLO functionality, you must add your server as an iLO device to the list of unprovisioned servers.
There are two ways to do this:
- Use the Run the OS Build Plan interface. In the second screen, where you select the target servers, you can add iLO devices one by one or through bulk comma-separated value (CSV) input.
- Right-click Devices > Servers > Unprovisioned Servers and select Add ILO Device.
In both these cases, you will need to manually provide iLO credentials and the corresponding SA realm (which is related to the SA Satellite or core available in the target network). This information defines the SA infrastructure to be contacted after the embedded service operating system is booted. When you use PXE booting, this information is provided by the PXE server automatically.
For more information on how to add your server, see" Installing (Provisioning) an Operating System" in the 9.10 UG: OS Provisioning.
Operating System Build-Plan Updates
The following 3 steps have been enhanced, or added, to build plans for HP ProLiant Gen8 support.
- Boot step - Enables SA to use the iLO library to boot the target server into its embedded service operating system. The step is transparent for non-HP ProLiant Gen8 target systems.
- Wait for Agent - Pauses the build plan until the SA Agent registers the target server and reports that it is in MAINTENANCE mode.
- Prepare Disks - Prevents the Windows installer from accidentally using the Virtual Installation Device (VID) as the target disk.
Virtual Installation Device (VID) handling
HP ProLiant Gen8 devices support special Virtual Installation Devices (VIDs).
SA handles VIDs as follows:
- If the BIOS VID setting is Hidden (default), SA detects the setting and handles VIDs automatically.
- If the BIOS VID setting is not Hidden, and you are using iLO with the HP ProLiant Gen8-specific embedded service operating systems (Linux or Windows PE), you must include the step "Prepare Disks on HP ProLiant Gen8" (see the HP ProLiant Gen8 online help) early in the build plan. Adding this step prevents the Windows installer from installing on the VID.
For more information on this topic and for specific use case, see "Prepare Disks on HP ProLiant Gen8" in the SA OS Provisioning README.
Running OS Build Plans on Servers
Build plans can now be run on servers that are in the pre-unprovisioned lifecycle.
If you are using SuSE Linux Enterprise Server (SLES) 11 SP1 and above, you must do the following to support the SuSE kernel & driver update method using kISO:
Note: If you experience driver problems with SLES 11 SP1 and kISO on HP ProLiant Gen8 target machines, see the following technical support documents:
http://h20000.www2.hp.com/bizsupport/TechSupport/Document.jsp?objectID=c03237879
http://h20000.www2.hp.com/bizsupport/TechSupport/Document.jsp?objectID=c03237878
For more information on kISO handling, see the kISO media documentation and the OS Provisioning README.
Custom attributes set on the build plan will now be resolved and used in parameters and scripts during the execution of the build plan. If a custom attribute with the same name exist anywhere else on or in the servers’ hierarchy, its value will take precedence over the value set on the build plan. This is the priority order from highest to lowest:
SA service operating-system drivers have been updated with the latest HP ProLiant SP drivers: SPP 2012.02 b51.
As of this release, SA provides multipath SAN support for the topics listed in this section.
OS Sequences
Red Hat 6 automatically identifies multipathing and enables kernel modules, but Red Hat 5 does not.
For Red Hat 5, before you run the Red Hat Enterprise Linux (REL) 5 OS Sequence, use the following steps to pass 'mpath' as a kernel argument for the server:
- Define a custom attribute kernel_arguments for the server, and set its value to mpath.
- Create the OS Installation Profile.
- Add a custom attribute to the Installation profile, and set its value to mpath.
- Set the value of the installation profile to mpath.
If you are performing multipath installation, it is also recommended that you add mpath as the kernel argument in the kickstart Red Hat 5 file. You can do this by creating a copy of Configure Default Red Hat 5 ks.cfg and modifying the file for use in your OS build plan.
SUSE Linux Enterprise Server 11
SUSE Linux Enterprise Server 11 does not identify multipathing by default. Use the procedures in this section to install SUSE Linux Enterprise Server on multipath LUNs.
You must copy the drivers into the media to enable multipathing. More information about the driver update disk (DUD) and the process of updating the media is available here:
http://www.novell.com/support/viewContent.do?externalId=7009981&sliceId=1
To configure SA, you must perform the following tasks:
To enable multipath installation, pass mpath as a value for the custom attribute kernel_argument by doing one of the following:
SUSE Linux Enterprise Server 11 requires that devices be specified in Mapper format during multipath environment installations. SA OS Provisioning verifies the AutoYaST profile and modifies it to suit the multipathing environments, based on the following rules:
Case | SA Solution |
---|---|
#1 - AutoYaST does not have the partitioning section. | SA will add the partitioning section with the drive and device. |
#2 - Has a single drive section, but does not have a device section. | SA will insert the boot drive expressed in Mapper format. |
#3 - Has multiple drive sections, but does not have device sections. Has multiple drive sections but no <device> in <drive> section. |
SA will not make any changes to AutoYaST. It will just print a warning and continue with the installations. This sort of configuration causes installation failure. To ensure a successful installation, fill each of the drive sections with devices in Mapper formats. |
#4 - Has devices in the configuration Single _OR_ multiple drive sections and a <device> specified for the <drive>. |
SA checks if the drive attributes in AutoYaST are in the correct Mapper format. If they are not, SA prints a warning and will not make any checks for correctness. If the drive attributes are not in the correct format, the installation fails. To ensure a successful installation, fill each of the drive sections with devices in Mapper formats. |
The /dev/mapper devices should only be used during the installation. The SLES Storage Administration Guide states that, in the running system, the multipath devices should be accessed through /dev/disk/by-id/.
SA will set up the scripts to convert the devices from /dev/mapper to /dev/disk/by-id when the user defines the custom attribute friendly_mpath_device to true.
Windows 2008/ Windows 2008 R2
Windows supports installations in multipath environments. Install multipath Device Specific Modules (DSMs) to further improve the capabilities of the target server. Import these DSMs into the SA Software Repository and install them using software policies attached to OS sequences.
You can modify the SAN Policy for the Windows Service OS (Winpe32, winpe64, winpe32-ogfs, winpe64-ogfs) shipped with SA. For more information about available options, see: http://technet.microsoft.com/en-us/library/cc749466(v=ws.10).aspx.
The default SAN Policy registry value in the SA-provided Windows pre-installation environment (WinPE) boot image is 1.
VMWare ESX/ESXi
VMWare ESX and ESXi support multipathing - no additional steps are required.
Use the Inject Multipath AutoYaST step
You must add the Inject Multipath AutoYaST Settings step to the OS build plan.
Perform the following tasks:
1. Create a new OS build plan by copying the default SUSE Linux Enterprise Server 11 Install Build Plan from the SA Client library.
2. Add the Inject Multipath AutoYaST Settings step after the Install And Configure SUSE Build Image step.
3. Add the Embed Files Initrd step after the Inject Multipath AutoYaST Settings step with the following parameters:-s /tmp/user.autoinst.xml:/
The drive injections in the AutoYaST will be performed based on the criterion defined in the OS sequences (see Partitioning Section in AutoYaST).
This section describes the SA 9.13 installation procedure.
As with previous SA releases, all SA Core installations and upgrades must be performed by HP Professional Services or HP-certified consultants. SA Satellite installations and upgrades performed by customers continue to be supported.
See the SA installation guides for more installation information.
patch_opsware.sh
script is used
both for installing and for uninstalling SA 9.13. patch_opsware.sh
. patch_database.sh
script is used both for installing and rolling back database schema
changes required for SA 9.13. patch_database.sh
script on the First Core and all Secondary Cores. Note that the
Model Repository and Oracle database can exist on a different host than the SA
host. patch_database.sh
. If any installed SA components (other than a previously installed patch) have a different build ID, you will not be allowed to install SA 9.13.
/var/opt/opsware/install_opsware/inv/install.inv
and find the section beginning with
%basics_
. Under this line, find thebuild_id
.
For example:
%basics_linux
build_id: opsware_45.0.6790.0
When you install an SA patch, the patch installation updates the
install.inv
file to record the patch installation and the patch build ID.For example:
%opsware_patch
build_id: opsware_45.0.17090.0
/etc/init.d/opsware-sas status
command), the patch operation will terminate. If you must roll back the SA 9.13 patch in a Multi-master Mesh, HP recommends that you roll back the secondary cores and satellites first, then the primary core.
spog.pkcs8
certificate must exist under /var/opt/opsware/crypto
(typically the
certificate is installed with the Shell, SAS Web Client, or Build
Manager). If the certificate does not exist, the patch operation will fail
with the following error:
Could not find spog.pkcs8
/var/opt/opsware/crypto/occ
Copy the certificate from another core machine
(for example, occ ) to
/var/opt/opsware/crypto/occ
and
retry this operation.
If this error is encountered, simply copy the certificate from another core machine to your core server and retry the operation.
/var/opt/opsware/crypto/spin
(typically the
certificate is installed with the Web Services Data Access Engine ( spin )). If the certificate does not exist, the patch
operation will fail with the following error:
Could not find spin.srv under
/var/opt/opsware/crypto/spin.
Copy the certificate from another core machine (such as occ) to
/var/opt/opsware/crypto/spin
and retry this operation.
You don't have permission to update the patch meta
database in HP SA.
Please re-run this command with a proper hpsa_user and
hpsa_pass.
The hpsa_user needs permission to write the folder
"/Opsware/Tools/Solaris Patching" and the Package Management
Client
Feature, "Manage Package" permission set "Read & Write".
There was a
problem with running update_supplements.
Please refer to section Patch
Management for Solaris of the Users Guide:
Application Automation manual
for details on how to set up Solaris
patching on your core.
You can safely disregard this error.
The pre-patch, database update and patch install scripts must be run in the following order:
SA Script Running Order - Upgrade
Upgrade From |
To |
Script Running Order |
---|---|---|
9.10, 9.10.01, 9.10.02, 9.11, 9.12 |
9.13 |
|
SA Script Running Order - Rollback
Roll back From |
To |
Script Running Order |
---|---|---|
9.13 |
9.10 |
|
Certain SA Client features (such as Run OS Build Plan or HP UX Provisioning) require the Adobe Flash Player. If you try to run these features, and you have not yet installed Adobe Flash Player, you will get an error.
To make sure Adobe Flash Player functions correctly and to avoid the error message, you should:
The error message should not display.
The script run during this procedure makes required changes to the Model Repository including adding required tables and objects.
patch_database.sh
on the Model Repository
host:
<distro>/opsware_installer/patch_database.sh --verbose -r <response file>
Where <
response file
> is the response file last used to install/upgrade the system.Usage:
patch_database.sh
[--verbose
] -r <response file
>
patch_database.sh
automatically detects if a database update is already installed and presents a corresponding menu:
Welcome to the Opsware
Installer.
It appears that you do not have a database update
installed on this system.
Press 'i' to proceed with
patch installation.
Press 's' to show patch
contents.
Press 'q' to quit.
Selection:
i
Welcome to the Opsware Installer.
It appears
that you have installed or attempted
to install a previous version of
the database
update on this system.
Press 'u' to upgrade the patch
to the current version.
Press 'r' to remove this patch.
Press 's' to
show patch contents.
Press 'q' to quit.
Selection: u
You chose to
upgrade the patch. Continue? [y/n]: y
Enter u
at
the prompt then Y
to begin the database
update.
[timestamp] Done with component Opsware SQL
patches.
[timestamp]
########################################################
[timestamp] Opsware Installer ran
successfully.
[timestamp]
########################################################
Note: After running the
patch_database.sh
script, you may see the following error when running the System Diagnostic test on your core:
Test Name: Model Repository Schema
Description: Verifies that the Data Access Engine's version of the schema matches
the Model Repository's version.
Component device: Data Access Engine (spin)
Test Results: The following tables differ between the Data Access Engine and the
Model Repository: local_data_centers, role_class_bridge.
This error is invalid and you can disregard it.
All SA installation DVDs require a DVD drive that supports dual layer.
Perform the following tasks to install SA 9.13:
patch_opsware.sh
on every host in the core/satellite facility
:
<distro>/opsware_installer/patch_opsware.sh --verbose
Usage :
patch_opsware.sh [--verbose]
patch_opsware.sh
automatically detects whether or not there is a patch already installed and presents a corresponding menu:
Welcome to the Opsware Installer. It appears that
you do not have any patches installed on this system.
Press 'i' to
proceed with patch installation.
Press 's' to show patch contents.
Press 'q' to quit.
Selection: i
Enter i at the prompt to begin the installation.
patch_opsware.sh
is invoked from a newer patch Welcome to the Opsware
Installer.
It appears that you have
installed or attempted
to install a previous version of
the patch on this system.
Press 'u' to upgrade the patch to the current version.
Press 'r' to remove this patch.
Press 's'
to show patch contents.
Press 'q' to
quit.
Selection: u
Enter u
at the
prompt to begin the upgrade.
The installer displays the following upon completion:
[<timestamp>] Done with component Opsware
Patch.
[<timestamp>]
########################################################
[<timestamp>] Opsware Installer ran
successfully.
[<timestamp>]
########################################################
This section details upgrades to the software repository content on the upload distribution (such as agent packages to be reconciled to managed servers).
patch_opsware.sh
from the upload
distribution, but will only update those packages that have changed since
the last major version. word
store
). word
store
) host, invoke the upgrade script::<distro>/opsware_installer/patch_contents.sh
--verbose -r <response file>
where <response
file
> is the response file last used to install/upgrade the SA
Core.
The following menu is displayed:
Welcome to the Opsware Installer. Please select the
components
to install.
1 ( ) Software Repository - Content (install
once per mesh)
Enter a component number to toggle ('a' for all, 'n' for
none).
When ready, press 'c' to continue, or 'q' to
quit.
Enter either 1 or a, and press c to begin the installation.
[<timestamp>] There are no components to
upgrade.
[<timestamp>] Exiting Opsware Installer.
To roll back SA 9.13 to SA 9.10, invoke the script:
<distro>/opsware_installer/patch_opsware.sh
--verbose
If this is a patched system, the following will be displayed:
Welcome to the Opsware Installer.
It
appears that you have previously
completed installation of this patch on
this system.
Press 'r' to remove this
patch.
Press 's' to show patch contents.
Press
'q' to quit.
Selection:
Enter r at the prompt to remove the patch.
<distro>/opsware_installer/patch_database.sh
--verbose -r <response file>
Where <response
file
> is the response file last used to install/upgrade the
system.
If the database has been updated, you see the following:
Welcome to the Opsware Installer. It appears that you have previously
completed the installation of this database update on this system.
Press 'r' to remove this patch.
Press 's' to show patch contents.
Press 'q' to quit.
Selection: r
Enter r at the prompt to begin the database schema update rollback.
This section lists the tasks that should be performed after you install SA 9.13. Some tasks might not be appropriate for your situation.
If you plan to install the SA Command-line Interface (OCLI) on a Windows Server after upgrading to SA 9.13, you must update the SA Agent on that server to the latest version. Errors occur during OCLI installation on Windows servers with earlier SA Agent versions.
The ability to import and install Windows multibinary patches—including those traditionally blocked by SA or requiring fixit scripts—was introduced in 9.06 and 9.11. You only need to perform these post-patch installation/upgrade steps once.
Note: If you have already performed these steps for a previous upgrade (SA 9.06, 9.11 or later), then you may disregard these instructions.
To take advantage of the patch-management enhancements in this release, the following steps are required after installing or upgrading to SA 9.11 or above:
# cd /opt/opsware/mm_wordbot/util
# export PYTHONPATH=/opt/opsware/pylibs2
# python clear_winpatch_prereq_override.pyc -c -d
The clear_winpatch_prereq_override.pyc script has the following options:
Option Description - c Clears out the blocked patches - d Displays the current value Note: You can run the script with just the –d value to see the current value.
This enables your server for Windows Multibinary Patch support.
(See the Windows chapter in the SA User's Guide: Server Patching for additional details.)
Note: To support the Windows multibinary patching functionality, the SA Agent must be from SA release 9.06 or 9.11 or later within those branches.
(See Upgrading SA Agents for details.)
Important: The SA Agent must be updated before you scan and remediate the servers.
After completing all of the above 3 steps, you should be able to scan and remediate the servers with the Windows Multibinary Patches.
Caution: Do not apply any SA-provided fixit scripts after upgrading to SA 9.11 or later. The fixit scripts are only applicable to previous versions of SA. If you encounter an issue with a patch for which there have been fixit scripts in the past, contact SA Customer Support. Applying fixit scripts after applying this SA update will not resolve the issue and could cause an error.
Tip: If you are installing Windows OS Service Packs (SPs), hotfixes, and update rollups, HP recommends that you remediate the server with the desired SP before installing the hotfixes and update rollups. Hotfixes are usually quickly released refinements that are eventually rolled up into the subsequent SP. Isolating the SP into its own policy before running the remainder of the vendor-recommended policy can prevent errors that might result from installing unnecessary hotfixes that are rolled up into the subsequent SPs.
(Back to the top of this document.)
If you install additional Slice Component bundle instances after patching the SA Core to version 9.1X, wayscript versions are set to version 9.10 rather than to the patch version. Use the following procedure to remedy the situation:
cd
/var/opt/opsware/OPSWpatch/OPSWwayscripts/scripts
Command #2:
./post_after_startup.sh
If you have customized such settings as Java heap settings, you must reapply your customizations after you install 9.13, as the settings are set to the SA default during installation or upgrade.
If you plan to import Windows multi-binary patches that have been traditionally blocked by SA or have required fixit scripts (such as MS11-078.pyc), see Windows Server Post-Patch Upgrade and Install Tasks in the Install section.
Back to the Table of Contents
This section describes issues from SA versions 9.0, 9.01, 9.02, 9.03, 9.10, 9.11, 9.12, and 9.13. The table lists issues first alphabetically by Subsystem, then numerically within each subsystem.
Known Issues
QCCR1D | Symptom/Description | Platform | Subsystem | Workaround |
---|---|---|---|---|
Agent | ||||
QCCR1D100660 | Windows ADT login fails for administrators that are not user Administrator due to security controls for Windows UAC. | Windows Server 2008 using UAC | Agent Deployment | Turn off UAC:
|
QCCR1D107917 | Failed SA agent installation returns gencache.py
error message in the agent log file
(%SystemDrive%\Windows\System32\opsware-agent-installer-
|
Windows | Agent Installer | Remove all the files from the following three directories, if they exist, and reinstall the agent. %SystemDrive%\Program
Files\opsware\agent\lcpython15 %TEMP%\gen_py %SystemDrive%:\Windows\temp\gen_p |
QCCR1D110347 | Windows servers running a pre-9.0 agent can perform a scan on an upgraded 9.0 core but not on a freshly installed 9.0 core because scanning Windows utilities are retained on the upgraded core but are absent from the freshly installed core. |
Windows | Agent | Upgrade the agent on the managed servers that you registered with SA 9.0. |
QCCR1D111593 | Agent fails to install when there is a gateway problem, but the agent
log reports that the installation was successful. |
Windows | Agent Installer | Make sure the gateway can be reached from the managed
server. |
QCCR1D118437 | Agent deployment fails several times due to timeout issues when the
uudecode utility is not installed on the machine to which
the Agent is deployed. |
Red Hat or SUSE Linux s390x | Agent Deployment/Upgrade | Install the uudcode utility on the machine(s) to which Agents are to be deployed to facilitate Agent deployment. |
QCCR1D123031 | Reprovisioning for Red Hat Enterprise Linux 6 with the ext4 file system is not yet supported. | Linux | Agent Deployment/Upgrade | None |
Application Configuration | ||||
QCCR1D50099 | Data-manipulation script that includes Japanese characters in the filename and content fails with errors. |
Windows | Application Configuration - Backend | Grant Read-Write privileges for Client Features - Manage Installed Configuration and Backups on Servers. |
QCCR1D111765 | You are unable to modify Application Configuration value sets for all scopes (Configuration, Facility, Customers) even when you have appropriate permissions. |
Independent | Application Configuration | Grant Read-Write privilege for Client Features - Manage Installed Configuration and Backups on Servers. |
APX | ||||
QCCR1D116476 | APX session timeout is not correctly implemented. | Independent | APX | To increase timeout for Web APX Proxy (Note: this must be done at all the Slices on all the Cores):
|
Audit and Remediation | ||||
QCCR1D81308 | An erroneous status is displayed when remediation of registry and COM+ objects fails. |
Windows | Audit and Remediation - Backend | None |
BSA Essentials Dataminer | ||||
QCCR1D112784 | In multimaster environments, Application Deployment data mismatches between SA and BSA Essention can occur if the data is replicated across the Model Repository Multimaster Component (vault) to where the BSA Essentials Dataminer is installed within the first few milliseconds of a minute. |
Independent | BSA Essentials Dataminer | None |
Database Scanner for Oracle | ||||
QCCR1D91143 | The status of an ASM Diskgroup shown in the Properties view (for one database instance only) is different than the status shown in the Database Configuration Assistant (DBCA) view (a CONNECTED status is displayed instead of a
MOUNTED status). |
Independent | Database Scanner for Oracle | None |
QCCR1D133226 | In order to monitor an Oracle 11G database with the SA Oracle Database Scanner, the XML DB and DBMS_NETWORK_ACL_ADMIN package must exist in the database.
The SA Oracle DB Scanner needs access to these objects in order to grant privileges and access for itself. If the objects do not exist, then the "pamuserprivilege.sql" will fail and the DB Scanner cannot be run. An application may or may not install these objects in its Oracle 11G database. The following error might be displayed under these circumstances: |
Independent | Database Scanner for Oracle | Before executing the SA DB Scanner "pamuserprivilege.sql" in the Oracle
database, first perform the following steps to install the XML DB and DBMS_NETWORK_ACL_ADMIN package in the Oracle 11G database.
|
Database and Middleware Automation (DMA) |
||||
QCCR1D133378 | Workflow cancellation indicates success in the DMA Console, but workflow runs to completion on SA managed server. | Independent | DMA | Use the appropriate process manager for the managed server's Operating System (for example, kill for UNIX and Task Manager for Windows) to manually terminate the process. |
QCCR1D146469 | If you run the da_baseline.sh -erase command manually in order to delete all of your Database and Middleware Automation data and reload the default data, the Discovery solution pack is not reloaded. | Independent | DMA | Follow these steps to import the Discovery solution pack:
|
Installer | ||||
QCCR1D111215 | Restoring OS Provisioning Stage 2 images fails on SUSE Enterprise Linux 9, which is a deprecated platform. | SUSE Enterprise Linux 9 | Installer | You can restore the OS Provisioning Stage 2 images by manually
running the restore_stage2.pyc script. This script is
located
in:<distro>/opsware_installer/tools/restore_stage2.pyc |
QCCR1D114639 | If you install additional Slice Component bundle instances after patching the SA Core to version 9.1X, wayscript versions are set to version 9.10 rather than to the patch version. | Independent | SA Installer | In the SA Web Client:
|
QCCR1D132004 | Health check fails with mismatched core component version errors (check_occ_up, lnxisoupload, new_platforms, opswbia_upload) | Independent | Installer | None |
ISM Tool | ||||
QCCR1D110511 | ISM tool fails to upload ISM into a software policy because the
tool mistakenly detects the registry key
HKLM\Software\Wow6432Nodeidentified , causing it to
identify a Windows Server 2008 server as a Windows Server 2008 x64
server. |
Windows Server 2008 | ISM Tool | If you are developing an ISM on a server that has this registry key, temporarily rename the registry during ISM development so that ISMtool will correctly identify the server as a Windows Server 2008 server. |
Jobs |
||||
QCCR1D121448 | When you run the dba_jobs script in the database, the following type of error is displayed: 8412:ORA-06512: at "SYS.DBMS_STATS", line 21271 8413:ORA-06512: at line 1 8433:ORA-12012: error on auto execute of job 68 8434:ORA-20003: Unable to gather system statistics |
Linux | Jobs | None |
OS Provisioning | ||||
QCCR1D100928 | RAID deployment fails when the RAID configuration is captured using
ACU version 8.35.7.0 (linux5 boot image) for valid RAID
configurations on machines with SCSI drives because the "pretty printing" of SCSI drive bus values uses 0-based index instead of 1-based index. |
Red Hat Enterprise Server 5 | OS Provisioning |
|
QCCR1D103362 | After a server is provisioned using a Red Hat DHCP-LESS image in the VLAN, attempting to reprovision the server causes the server to reboot, after which the reprovision process fails at the Anaconda Configure TCP/IP window, prompting for network information. |
Independent | OS Provisioning - SA Client | None |
QCCR1D103602 | After you choose Windows as the OS Family, you cannot use the
Manage Boot Client (MBC) to create winpexx-ogfs server
records in the MBC Single Form because winpe32-ogfs
and winpe64-ogfs image types are not displayed in the PXE
image drop-down box. |
Windows | OS Provisioning - Backend | Use the CSV import to generate the pre-unprovisioned devices. |
QCCR1D104194 | When RAID deployment fails after the RAID controller configuration
has been cleared, subsequent RAID captures or deployments will
fail unless RAID is first configured manually.
The following error displays:
|
Independent | OS Provisioning - Backend | Manually set the RAID controller configuration:
After the RAID controller has been manually configured, |
QCCR1D104739 | During a network boot of the server, the following error displays: "No driver found" screen appears: " |
Red Hat Enterprise Server IA64 | OS Provisioning | The missing driver is not required. Press F12 to bypass the driver. |
QCCR1D109077 | Assign Customer OS Build Plan content script fails if customer names contain quote characters. |
Windows | OS Provisioning Backend | Specify the customer ID instead of the actual customer name. |
QCCR1D114523 | When performing OS provisioning using Application Deployment Automation, if the SA OS Sequence in use includes a device group and multiple servers are provisioned simultaneously, some of those server provisioning jobs will fail with the following error: Unexpected general exception: com.opsware.device.DeviceGroupVO.modifiedDate |
Linux, Solaris, Windows | OS Provisioning | Do not include any device groups in your OS Sequence. |
QCCR1D129619 | ESXi DHCP-LESS: DNS Search Domain was not set via the DHCP-LESS value | VMware ESXi 4.1 | OS Provisioning Backend | None. Anything entered in the "DNS Search Domain" field is inserted into the custom attribute as domain=. Therefore, the DNS domain will not be inserted to the ESXi 4.1 system. |
QCCR1D131265 | Manage Boot Client utility: WAPX sometimes fails at the second step with: 503 Service Unavailable Error | Independent | OS Provisioning Backend | In the Global File System (GFS), use the /opsw/apx/bin/osprov/manage_boot_clients_script. |
QCCR1D132628 | SUSE Linux Enterprise Server provisioning fails on an HP ProLiant server with a CCISS disk. | SLES | OS Provisioning | For post-SUSE Linux Enterprise Server 10 SP2 provisioning on HP ProLiant servers with CCISS disks, ensure that the <code1>autoyast.xml</code1> file has the partitioning configuration specified as shown below:
<partitioning config:type="list"> |
QCCR1D132825/130199 | During provisioning in a non-DHCP environment, Red Hat Enterprise Linux 6 provisioning can become interactive if the HPSAlinux_boot_cd.iso remains connected. | Red Hat Enterprise Linux 6 | OS Provisioning Backend | Eject the boot CD when the server reaches the restart phase or after provisioning becomes interactive. |
QCCR1D133153 | VMware ESX 4.1 OS Sequence with OS firewall enabled fails with a broken pipe exception during the Agent download phase. | VMware ESX 4.1 | OS Provisioning | The firewall should be disabled during OS Provisioning and re-enabled after. |
QCCR1D136144 | Running a single OS Sequence job against two servers with the same MAC address in different networks causes an exception. | Independent | OS Provisioning - Backend | If you have servers with the same MAC address in different networks, you must provision them sequentially. For example, complete the Run OS Sequence job on the first server before you power on the second server. |
QCCR1D136423 | Provisioning dupe-mac Linux servers sequentially into a managed server, then reprovsioned the servers simultaneously should result in a Run OS Sequence job error. |
Linux | OS Provisioning | None |
QCCR1D136445 | If you attempt to register another server with the same MAC address as a mini agent after a server is in the OS Provisioning lifecycle, the Run OS Sequence job on the first server fails. | Independent | OS Provisioning | Follow these instructions:
|
QCCR1D136592 | After a server is registered with SA as miniagent, if you then attempt to register a second server with the same MAC address (but with a different service operating system than the first server) as a miniagent, the registration fails. Then, if you shut down the second server as instructed, the Run OS Sequence job against the first server fails with an error. | Linux | OS Provisioning | When this problem occurs, you must
1. Ensure the second server remains shut down |
QCCR1D146347 |
Linux 6-OGFS boot failed on HP ProLiant DL380 G4 and DL385 G1. | Linux | OS Provisioning | None |
QCCR1D144427 | When creating an iLO device using the Add iLO operation, the server history mistakenly shows that the device was created by user opsware, and not the user actually performing the operation. | Independent | OS Provisioning | None |
QCCR1D145658 | OS sequence provisioning for SuSE Linux 11 x64 on HP ProLiant DL385G5 fails randomly. |
SuSE Linux | OS Provisioning | Set the mpath kernel argument on the server record before running the OS sequence. |
QCCR1D147226 | An iLO's realm information is not updated after the iLO is moved from one satellite to another. | Independent | OS Provisioning | Remove the iLO server, then add it again, this time into the correct realm. |
Patch Management for Solaris | ||||
QCCR1D100566 | Although the reboot is performed correctly, when you preview remediating a patch policy on a server, or view the job status for a patch policy that is already remediated, the reboot setting might incorrectly display “Install and Reboot Later” when it should display “Install and Reboot”. |
Solaris | Patch Management - Solaris | A workaround is not required because the reboot is performed
correctly, even though the display may be incorrect. |
QCCR1D111342 | Importing clusters with old names using the solpatch_import command will fail because Sun has renamed their Solaris patch clusters - the clusters in your SALibrary may not match the cluster names from Sun on Sunsolve. | SunOS 5.6 - 5.10 and SunOS 5.10 x86 | Patch Management - Solaris | Download the latest supplement file from HP LNc. For complete
instructions, see “Obtaining the Solaris Patch Supplementary Data File” in the SA User Guide: Application Automation. |
QCCR1D114156 | The solpatch_import –filter option does not display
recommended and/or security patches if they had previously been marked obsolete. This became an issue on June 4, 2010 when Oracle changed
the criteria for recommended and security patches (described here: http://blogs.sun.com/patch/entry/merging_the_solaris_ recommended_and). |
Solaris | Patch Management - Solaris | Users with an existing metadata database (solpatchdb )
must delete the solpatchdb.zip , solpatchdb-old.zip and files and run create_db to have support for recommended obsolete patches. You must recreate the Solaris patch metadata database ( solpatchdb ) if the following are true:
After you have installed SA 9.13, perform
these tasks to recreate
Follow the steps to create a new metadata database ( solpatchdb ) as described in the SA User Guide: Application Automation, Patch Management for Solaris. |
QCCR1D130601 | solpatch_import: Import "Fujitsu Solaris 10 OS Recommended Patch Cluster" fails with 'KeyError: 'patch_order.SS' on a Solaris core. The following error displays when importing a large cluster on an older Solaris core:
|
Solaris | Patch Management - Solaris - Backend | Import a specific patch on your core. See the article from Oracle at https://support.oracle.com/CSP/main/article?cmd=show&type=NOT&id=1020109.1. |
QCCR1D132332 | After upgrading to SA 9.13, you must manually run update_supplements to see new clusters, such as Fujitsu clusters. | Solaris | Patch Management - Solaris | SA 9.10 provides new clusters (Fujitsu clusters) in the Solaris collection file. This workaround applies to a mesh that contains different SA cores (such as SA 7.88 or SA 9.13), where the mesh has Solaris patching configured (solpatchdb). After the mesh has been upgraded to SA 9.13, the following command must be manually run on a core server (slice): /opt/opsware/solpatch_import/bin/solpatch_import -a update_supplements This command loads the new collection file into solpatchdb. To see all available clusters, including the new clusters, run the following command: /opt/opsware/solpatch_import/bin/solpatch_import -a show --available_clusters |
QCCR1D138928 | Solaris Recommended Clusters cannot be imported into SA when the SA core server has a version of unzip that does not support large Solaris Cluster zip files. | SunOS | Solaris server patching | To resolve this issue, update the unzip tool to v.6.0 on the SA core server before importing the recommended Solaris Clusters into SA.
To update the unzip tool on the SA core server:
|
Patch Management - Windows | ||||
QCCR1D102713 | If a patch policy contains one or more superseded patches, the number of total rules counted after remediating or patch installation is incorrect. Note that the compliance state of the server is accurate before and after remediating. | Windows | Patch Management - Windows - Backend | None |
QCCR1D105098 | Automatically installing recommended Microsoft patches Q934041
(2000) and Q924883 (MS07-014) fails. Manually installing the
patches results in a Non-Compliant server status and a dialog that
is displayed: Update has already been applied or is included in an update that has already been applied. |
Windows | Patch Management - Windows - Backend | Since the patch is not required, set a “never install” exception on the patches. |
QCCR1D108451 | Windows Patch 944036 (installer for IE 8) reports install failure (and returns a non-zero exit code) on Windows Server 2008 x64 managed servers, even though the patch is installed correctly. Subsequent patch and compliance scans will indicate the patch is installed and compliant. | Windows Server 2008 x64 | Patch Management - Windows - Backend | A workaround is not required because the patch job succeeded, even though an error displayed. |
QCCR1D110257 | If you try to install the March 2010 (or later) MBSA patch database, and then attempt to install MS10-015 (KB977165) and MS10-021 (KB979683) on to a Windows Server 2008 x86 using a Windows Patch policy, the Patch install job results will incorrectly indicate success. After a patch compliance scan, SA will still report the patches as recommended for the server, and the server will be listed as non-compliant. | Windows Server 2008 x86 | Patch Management - Windows - Backend |
“side effect” messages. These progress messages may not be accurate. The true indicator of whether these patches installed or not is when after the Install Patch job completes (with reboot), MS10-015 (KB977165) and MS10-021 (KB979683) are no longer recommended, and the compliance indicators for these patches no longer show a red X. Note: The SA Client will show the mpsyschk.exe version of these patches as recommended or installed. |
QCCR1D108451 | Attempts to install the patch for KB944036 (Installer for IE8) using SA will fail. | Windows Server 2008 x86 | Patch Management - Windows - Backend | Log on to the managed server and install the patch for KB944036 manually. |
QCCR1D111397 | After remediating the vendor recommended policy on a Windows managed server, the server's recommended patch list shows additional patches that need to be installed, requiring the vendor recommended policy to be remediated again. This sometimes occurs when older patches introduce problems that newer patches already fixed. | Windows | Patch Management - Windows - Backend | None |
QCCR1D136373 | Windows systems that use .cab files for remediation time out during the installation process for a vendor recommended policy. When installing multiple .cab files without reboot, calls to the Windows API UpdateSearcher.Search do not return in a timely manner, causing the timeout. | Windows | Patch Management - Windows - Backend | After the remediation job fails (due to the timeout), reboot the server to complete the timed out job and then run the remediation job again. |
QCCR1D130284 | Install patch Q979309 failed on a Windows 2008 R2 server with exit code 4. | Windows | Patch Management - Windows - Backend | Install KB978601 and KB979309. |
QCCR1D131240 | If you do not have the SA 9.10 Agent installed on a managed server, the Reboot Server option will not change the server's Reboot Required state from "Yes" to "No". | Windows | Patch Management - Windows - Backend | Upgrade the Agent on your managed server to the SA 9.10 Agent and then reboot the server. The server's Reboot Required state will then accurately display the state as "No". |
QCCR1D140751 | Attempts to install the Windows KB929300 multi-binary patch without including all the binaries in the SA patch policy remediation job will fail with the following error and stack trace:
The operation to install or uninstall software failed.Execution error: Traceback (most recent call last): |
Windows | Patch Management - Windows | To resolve this issue, include ALL binaries from the multi-binary patch in the SA patch policy remediation job. |
QCCR1D153156 | Duplicate patches can be created in the SA database if you import the Microsoft Patch Supplement (MPS) and then run the SA Patch Import process using the Microsoft Offline Patch Catalog (wsusscn2.cab). These duplicates can cause conflicts during remediation and compliance checks. | Windows | Patch Management - Windows | A one-time de-duplication procedure is available to remove the duplicates and eliminate the source of these conflicts to prevent future duplication. See Patch De-duplication Steps for Windows Patching for more information. |
SA Client (Framework) | ||||
QCCR1D105671 | The SA Client cannot be installed under a localized (I18N) directory if the directory path contains non-ascii characters. | Windows | SA Client Framework | Install the SA Client Launcher in a path containing only ascii characters. |
QCCR1D115211 | Launching the SA Client under Windows Vista or Windows 7 causes flickering/redrawing of the screen. | Windows Vista, Windows 7 | SA Client Framework | Disable the Windows Aero graphical user interface and/or use the Windows Classic display. |
QCCR1D115403 | The SA Client freezes when any of the following items is launched for the first time: Application Deployment, Web Automation Platform Extension (APX) processes (including OS Build Plans and HP Virtualization), and reports. | Windows | SA Client Framework | Use the Applications tab in the Windows Task Manager to stop the SA Client. Restart the SA Client. |
SA Client (Search) | ||||
QCCR1D66448 | Some discovery dates for SAN arrays, NAS filers, and switches will
not display as expected, although they are technically correct, if the user profile is set to a non-UTC timezone for Storage System Discovery Date. |
Independent | SA Client (Search) | None |
SA/SAR Reports | ||||
QCCR1D82108 | In the SA Client in an NA-enabled core, if you run the Connections by Network Device report and set the parameter to Device Name Equals [Any Value], the search returns no results. | Independent | SA Client Reporting | Run the report using the following parameters: Device Contains |
QCCR1D107293 | Scheduled reports exported to .xls do not display charts or graphs
(“Unsupported Image error” is displayed instead). However, tables are sent correctly. Graphs are not visible in the .xls file, but the report should not display empty image blocks. |
Independent | SA Client Reporting | None |
SA Web Client | ||||
QCCR1D118067 | Chinese Date format: Commonly used date formats, like yyyy ’year’ mm ‘month’ dd ‘date’ and ‘yyyy-mm-dd’ are not listed in the user profile setting, so users cannot select a favorite date format. | Win2008 Win2008 R2 | SA Web Client | None |
QCCR1D131019 | After locales are enabled, the SA Web Client fails to start. | Windows | SA Web Client | None |
Satellites | ||||
QCCR1D91097 | In some cases, in the SA Web Client Manage Gateway page, changes to bandwidth throttling values between two gateways do not persist. | Independent | Satellites | You must apply any bandwidth changes between gateways on the Core Gateways only. Changes made on other gateways will not take effect. |
QCCR1D97659 | Network scans to a satellite realm fail for hosts with the error:
XML document structures must start and end within the same
entity . |
Windows | Satellites | In the SA Client Options select Tools -> Options->
Unmanaged Servers -> Advanced and remove the argument
-S %GATEWAY_IP% from the NMAP parameters. The network scan
should complete successfully. |
Script Execution | ||||
QCCR1D79545 | Exporting a Run Server Script Job output containing multi-byte
characters (Japanese/Korean) to .csv results in question
marks in place of the multi-byte characters. |
Windows | Script Execution | Export the job results in .txt format to eliminate the
garbled text. |
SE Connector | ||||
QCCR1D88755 | There is no Target and Target Volume information displayed in the storage volume access path view for a LUN. | Independent | SE Connector | None |
QCCR1D91582 | When you perform a provisioning operation for an HP EVA array (such as create, delete, or modify a volume or pool), the changes for the volume or pool might not be immediately available in the corresponding Inventory view in SA after running the "Update from Storage Essentials" process. | Independent | SE Connector | After 30 minutes has lapsed, run the "Update from Storage Essentials" process again. See the Storage Essentials SRM Software User Guide for information about provisioning EVA arrays. |
QCCR1D105778 | When a server on which SE Connector is running is directly
deactivated and deleted, selecting Administration -> Storage
Scanners -> <The names of the SE Scanner on "Host">
shows stale entries of Storage scanners in the Scanner
panel. The stale entries count will increase, depending on how many
times the server is deactivated and deleted from the core. |
Independent | SE Connector | Manually delete the inactive Storage Scanner entries from the Storage Scanner panel by using the Remove menu option provided for each entry. |
Software Management | ||||
QCCR1D101517 | After performing a software remediation, the compliance status might incorrectly show servers as being out of compliance, because there is a caching delay in the Web Services Data Access Engine (twist). | Independent | Software Management | Run a Software Policy Compliance scan. This will show the correct compliance status. For more information, see "Software Compliance" and "The Software Policy Compliance Scan" in the SA User Guide: Application Automation. |
QCCR1D115665 | Migrating a Windows Server 2008 R2 x64 server, attached to a dynamic device group with the rule Operating System=Windows Server 2008 x64 specified, prevents all Application Configurations, remediated using a Software Policy, from being detached from the server. | Windows Server 2008 R2 x64 | Software Management | None |
QCCR1D118747 | The install/uninstall software status is Failed for all packages, even for packages that were successfully installed. | Linux | Software Management - Backend | Set the way.remediate.yum=0 in the SA Command Center system configuration. |
QCCR1D132577 | When you remediate a software policy that contains patches, you see a red cross on the software-policy icon in the Management Policies -> Software Policies window.
The following message is displayed: "There is nothing to be done on this job." |
Solaris, HPUX, and Windows | Software Management | None |
QCCR1D133097 | Running a compliance scan using a software policy that contains a large number of software policy items might generate errors if the following conditions exist:
There is no exact number of items in the software policy that will cause this issue; the higher the number, the higher the probability of an error. |
Software Management | None | |
QCCR1D133123 | Script permissions are not updated correctly in the SA Client. | Windows | Software Management | Reset the Manage Server Script feature permission to restore the permissions as follows:
|
Software Repository |
||||
QCCR1D131751 | HP Server Automation 9.10 configurations that use a Solaris machine as the Network File System (NFS) server may encounter unexpected behavior and unreliable performance. Due to known instability issues with Solaris NFS, this configuration should be avoided. The recommended configuration is a Linux NFS server, which works reliably with either Solaris or Linux clients. | Solaris | Software Repository | The recommended configuration is a Linux server as the NFS, which works with either Solaris or Linux machines as clients. |
Storage Host Agent Extension | ||||
QCCR1D93630 | LUNs that are multipathed by EMC PowerPath are shown as both "ROOT" and "LUN" in the Inventory -> Storage -> Volumes Panel alongside LUNs that are correctly displayed. The SCSI Bus number provided by PowerPath (using the powermt command) does not match the bus number of the disks (LUNs). | Windows | Storage Host Agent Extension | None |
QCCR1D93690 | The Server -> Relationships -> SAN
Switches panel on a virtual server is empty/only displays SAN switches to which the given server is directly connected. In some cases, a server may depend on SAN switches that are not displayed in this panel. For example, a virtual server may be using storage allocated from a hypervisor that was allocated storage from a SAN. |
Independent | Database Scanner for Oracle | None |
QCCR1D105382 | When you run the storage snapshot specification, the disk capacity
information is changed in the Inventory -> Storage -> Disk panel. This occurs if there is a mismatch in the disk names, as reported by the hardware registration script and the storage snapshot specification. |
Windows 2008 | Storage Host Agent Extension | Reboot the server after changing disk information (such as
installing or uninstalling multipathing software, presenting new LUNs,
deleting LUNs, and so on), and run the hardware registration before running the storage snapshot specification. |
QCCR1D105953 | An EMC Symmetrix array that is discovered through SE Connector can
report more than one storage volume with the same LUN number
presented to a managed server. Running the storage snapshot
specification on the managed server will succeed; however, the
Inventory -> Storage -> File Systems and Inventory -> Storage -> Managed Software panels will be empty. In addition, some host volumes with a LUN service type will not be displayed in the Storage -> Inventory -> Volumes panel. |
Independent | Storage Host Agent Extension | None |
QCCR1D106400 | The right-hand panel under Inventory -> Storage -> Volumes does not display the supply chain and composition information
properly (or the information is missing altogether) for some volumes that were created out of LUNs |
AIX | Storage Host Agent Extension | Delete the stale volumes on the AIX managed server and run the storage inventory snapshot specification. |
QCCR1D106699 | For managed servers with mirrored volumes, if one of the disks that
is part of a mirrored volume fails or is removed, the state of the
volume is shown as "Failed Redundancy" in the Disks Management panel. However, in the Inventory -> Storage -> Volumes panel for the managed server, the status of this volume is shown as "OK", even after running storage snapshot specifications. |
Windows 2008 | Storage Host Agent Extension | None |
QCCR1D107944 | Running a storage snapshot specification on unsupported ESX 3.0.x servers returns an error message that indicates unsupported namespace in content of SOAP body. Storage related information for ESX 3.0.x is not stored and displayed in the SA Client. | VMware ESX 3.0.x | Storage Host Agent Extension | None |
QCCR1D111724 /111727 |
Host Storage Inventory does not support VMware servers that have VMDK on NFS datastore. Therefore, no storage information will be collected on VMWare servers with this configuration. |
All VMware servers | Storage Host Agent Extension | None |
Bug ID: 149406 / QCCR1D 60760 | Solaris LVM RAID on Soft Partition on slices stops responding, resulting in a defective storage supply chain. |
Independent | Storage Host Agent Extension | None |
Bug ID: 149707 / QCCR1D 61061 | The SNIA v1 HBAAPI Extension reports two single port cards when a single dual port card is present in a multi-port card. |
Independent | Storage Host Agent Extension | None |
Bug ID: 151921 / QCCR1D 63275 | When you add a mirror to concatenated or stripe, there is no distinction between the volume types "Mirror Concatenated" and "Mirror Striped" in the Volume Manager labels. Note that "Mirrored Concatenated" and "Mirror Striped" are distinct on the volume manager on the host, such as on the Veritas Volume Manager. | Independent | Storage Host Agent Extension | None. The type of volume manager might not match the native tool,
such as the Veritas Volume Manager. The STORAGE_TYPE
value is the immediate nodein the supply graph, which is the storage type of the most decendent volume. |
Bug ID: 152016 / QCCR1D 63370 | The value stored in STORAGE_COMPONENTS.STORAGE_DRIVE
is a different format on Solaris 5.10 than on Solaris 5.8 and 5.9.
The different format for 5.10 causes a broken storage supply chain on
affected servers. |
Unix | Storage Host Agent Extension | Check the version number in the /etc/format.dat file on the server. If it is less than 1.28, update the file. |
Bug ID: 152942 / QCCR1D 64296 | On a Windows 2003 server with the SNIA library from QLogic 9.1.4.15 HBAAPI, Fibre Channel Adapter and storage volume information might not be discovered by the Storage Host Agent Extension, causing fibreproxy.exe to stop responding. | Windows | Storage Host Agent Extension | For Windows Server 2003 operating
system, use the native Microsoft SNIA library instead of the SNIA that is provided by the QLogic driver. Download the Fibre Channel Information Tool to add the Microsoft HBAAPI support to the operating system. For Windows 2003 SP1 or later, the Microsoft HBAAPI support is built in. If the SNIA's version of hbaapi.dll is installed on the operating system, remove it. |
Bug ID: 154418 / QCCR1D 65772 | When you snapshot a Unix server that has a QLogic driver installed,
there is no:
|
Independent | Storage Host Agent Extension | Install patches 108434 and 108435 on Solaris 8 SPARC servers. The
Storage Host Agent Extension on Solaris 5.8 SPARC requires these
patches. Note: There is no known workaround for Red Hat 3 or Red Hat 4 servers using QLogic controllers. |
Bug ID: 154971 / QCCR1D 66325 | Veritas Storage Foundation 4.3 with QLogic 9.1.4.15 results in invalid fibre proxy SCSI addresses. The SAN storage volume displays both LUN and Root as the Service Type. There are two lines for the physical drives: One line displays LUN and the other displays Root. | Independent | Storage Host Agent Extension | None |
Bug ID: 155476 / QCCR1D 66830 | The file system is not displayed on the server storage file system panel when the partition and format on the Windows server is mounted to an empty NTFS folder. There is no support for mounting Windows 32 file systems on non-drive letter locations. | Windows | Storage Host Agent Extension | None
Note: The Storage Host Agent Extension does not report file systems
that have non-drive letter mount points. The Storage Host Agent
Extension does not report file systems |
Bug ID: 157579 / QCCR1D 68933 | When you run take a Storage Host Agent Extension snapshot by running fibreproxy on a Windows server where Emulex LP850, LP952, LP9002, or LP9402 is installed, three FibreChannelTargetMappings arereturned, two of which are duplicates. This symptom does not occur with Emulex driver 1.30a9. |
Windows | Storage Host Agent Extension | None |
Bug ID: 158923 / QCCR1D 70277 | If you run the chpath command as shown below to take a Storage Host Agent Extension snapshot for each available path to the device, all the MPIO paths to a logical device become disabled. In this state, the system calls used by the diskproxy and mpioproxy will stop responding. chpath -l hdisk2 -p fscsi0 -s disable
xx |
AIX | Storage Host Agent Extension | None |
Bug ID: 159156 / QCCR1D 70510 | After you update a LUN mapping (and remove the initiator ports) from a mounted SAN volume on a server, the old LUN mapping information still displays in the SAN array volume view and in the server storage volume view. An additional access path is displayed in the SAN array volume view (Access Path subview) for the volume for which LUN mapping was removed. The access path that shows no initiator device and/or initiator port information is the correct one. The incorrect access path is removed from the display when the next Storage Host Agent Extension snapshot is run. |
independent | Storage Host Agent Extension | Take a snapshot of the server to which the volume was mapped or partitioned. |
Bug ID: 159580 / QCCR1D 70934 | The SAV displays incorrect information after adding a zone to a fabric. The zone is not associated to the correct server/port/WWN. |
independent | Storage Host Agent Extension | None |
Bug ID: 164951 / QCCR1D 76305 | The multipath information is not reported correctly for a server that has HP-UX 11iv2 OS installed and Veritas DMP managing the multipathing in the SA Client. The SNIA library does not support HBA_GetFcpTargetMappingsV2r . |
HP-UX 11iv2 | Storage Host Agent Extension | None |
Bug ID: 167103 / QCCR1D 78457 | If you perform a core upgrade to SA 7.50 and ASAS 7.50 and then
run the customer extension to upgrade a Storage Host Agent Extension
on the host, the host disappears from the INTERFACE table and the host's STORAGE_DRIVE does
not appear in the STORAGE_COMPONENT table. |
Independent | Storage Host Agent Extension | It may take one to two hours for the host and drives to
repopulate their tables. Verify that the host is present in the INTERFACE table and that the STORAGE_DRIVE element is present in theSTORAGE_COMPONENT table. |
Bug ID:168889 / QCCR1D 80243 | If you disable a volume in Veritas DMP and subsequently take a new Storage Host Agent Extension snapshot, it appears as if the updated volume is still managed by Veritas DMP. | Independent | Storage Host Agent Extension | When constructing LVM modules on the HP-UX 11.31 platform, use agile DSF devices. There is no workaround for other platforms. |
Usability |
||||
QCCR1D133123 |
Script permissions are not updated correctly in the SA Client. | Windows | Usability | Reset the Manage Server Script feature permission to restore the permissions as follows:
|
Virtualization | ||||
QCCR1D90019 | If a system has more than one virtual network with the same name, even if those networks are managed by different hypervisors, scanning for virtual servers fails due to a violation of unique name constraints. | Windows Server 2008/Hyper-V | Virtualization | Do not use duplicate virtual network names. |
QCCR1D104418 | The reported OS property text for ESX servers is inconsistent between direct (SA Agent) and indirect managed (vCenter) cases. | VMware ESX (all versions) | Virtualization - Backend (VMware) | None |
QCCR1D105999 | After cloning an SA-managed virtual machine, when the new clone starts
up for the first time but cannot obtain network connectivity, agent revival will fail to create a new server record for the new clone. |
Independent | Virtualization - Backend (VMWare) | Restart the agent on the new clone after network issues are resolved and the agent will correctly register as an SA-managed virtual machine. |
QCCR1D106085 | If your Hyper-V server has more than one IP address, SA might change the Management IP address from the one you registered to one of the other IP addresses. | Windows Server 2008 pre-R2 server | Virtualization - Hyper-V | To prevent this problem, you need to manage your Windows 2008
pre-R2 server from a Windows 2008 R2 server and make sure the option to
allow the management operating system to share the network adapter
is not selected. The following gives the basic steps to accomplish
this, however, see your Microsoft Hyper-V documentation for complete
details. More information may also be available by searching the
internet for "New in Hyper-V Windows Server 2008 R2" and "Hyper-V
Remote Management: You do not have the required privileges to complete
this task."
|
QCCR1D106909 | Clone Virtual Machine job will fail the Registering Server step if the Windows Shutdown Event Tracker is enabled on the source virtual machine. This is because the Shutdown Event Tracker waits for user input before it completes rebooting, so the SA Agent registration cannot complete. | Windows | Virtualization | Disable the Shutdown Event Tracker on the clone source virtual server. Windows 2003 x64 cloning requires a manual reset to resume virtual machine images customization. |
QCCR1D109887 | The snapshot view is not available on an ESX server that is managed by vCenter. |
Red Hat Linux | Virtualization | Manage ESX directly snapshot view. |
QCCR1D110035 | When a hypervisor that is dual-managed (through Virtualization Service and SA Agent) loses one of its management paths (such as when the Virtualization Manager or VCenter is removed from SA) the Login Credentials panel does not display in the server browser panel and the hypervisor credentials do not display in the Properties view. | VMWare ESXi | Virtualization | Right-click on the hypervisor, select "Refresh Server", and then press F5 (Refresh) to refresh the client so that the Login Credentials panel displays in the server browser. |
QCCR1D111307 | If you add an ESXi server with larger hardware configuration data, the add operation suspends processing. | ESXi | Virtualization - Backend (VMWare) | Reconfigure the server with fewer CPU counts in hardware information. |
QCCR1D111780 | Hypervisors are not loading. | Windows | Virtual Center | Check that the hypervisor version you are using is supported. Or
check to see if the hypervisor is in a NOT RESPONDING
state. |
QCCR1D111789 | Adding two vCenters concurrently results in one of the automatically triggered reload data to suspend processing. | Independent | Virtualization - Backend (VMWare) | Add the vCenters separately. |
QCCR1D111922 | The Create Virtual Machine operation on an SA managed ESX 3.5
server fails with an error message of
com.vmware.vim25.VirtualMachineConfigSpec. |
ESX 3.5 | Virtualization - Backend (VMWare) |
After the SA 7.5 release, one of the VMM library jar files was
changed from OPSWvmm-vmware.jar to vmm-vmware.jar. When an
upgrade is performed from SA 7.5 directly to SA 9.0, or from SA 7.5 to
SA 7.8 to SA 9.0, and any virtualization operation is invoked on the
ESX hypervisors, the VMM package gets remediated first with the new package but the OPSWvmm-vmware.jar file is left
untouched. This causes the consecutive virtual machine create operations to fail. The following workaround is intended for any ESX
hypervisor that is managed by an Agent in SA and whose
virtualization aspect will still be handled through the same route (not
via VS)
|
QCCR1D111972 | Create Virtual Machine fails on a directly managed ESX hypervisor if the virtual machine's datastore name contains special characters. | ESX | Virtualization - Backend (VMWare) | Change the datastore name so that it does not contain special characters. |
QCCR1D116276 | After an SA Agent is installed and before hardware registration has completed, you can create or modify a virtual machine with a memory value that is larger than the hypervisor's physical memory. The job completes successfully. | Windows 2008, Windows 2008 R2 | Virtualization - Hyper-V |
Run a full hardware registration on the Windows server. After hardware registration is completed, SA will not allow you to use memory that is greater than the maximum memory allowed on a virtual machine for the actions to create and modify a virtual machine. |
QCCR1D146111 | Permission denied to run APX v12n.apx.utility when you try to add an ESXi hypervisor. | ESX | Virtualization - Backend (VMWare) | In order to add an ESXi hypervisor (or to run ESXi5 Build Plan OS Provision) you must also have Allow Install Agent (SA Client) or Allow Managed Server (SA Web Client) action permission. |
Web Services Data Access Engine |
||||
QCCR1D111039 | An out-of-memory error is encountered in the Web Services Data Access Engine | Red Hat Linux/Solaris | Web Services Data Access Engine (twist) | The default maximum heap size for Web Services Data Access Engine has been increased to 2560MB from 1280MB. |
QCCR1D112222 | The Web Services Data Access Engine does not start and records the error:
Could not reserve enough space for object heap in:
|
Platform: Red Hat Enterprise Linux AS3 32-bit |
Web Services Data Access Engine (twist) | Before upgrading, edit the file:/etc/opt/opsware/twist/twistOverrides.conf and add the following entry: twist.mxMem=<memory size in
Megabytes> The value must be 2000 megabytes or less. Example:
twist.mxMem=1960m |
The Fixed Issues table includes issues that:
The table lists issues first by subsystem, then numerically within each subsystem.
QCCR1D | Symptom/Description | Platform |
---|---|---|
Agent |
QCCR1D141249 | Some Windows physical hosts are reporting as Hyper-V Virtual Machines (VMs). | Windows |
Agent Deployment/UI |
||
QCCR1D133423 | The ‘Choose Automatically’ option in the Protocol drop-down list will attempt the deployment for all detected protocols, not just the protocols that are available for all scanned hosts. The rest of the options in the drop-down list will be the intersection of detected protocols for all scanned hosts. The Agent Deployment Tool backend uses the detected protocols, instead of all available UNIX-specific protocols. | Independent |
QCCR1D142492 | Agent Upgrade Tool fails and displays a runtime error. | Linux |
Application Deployment Manager |
||
QCCR1D142218 | You can now specify a base directory for Configuration File components | Independent |
QCCR1D144882 | You can now copy and paste an existing Component into the same or a different application. | Independent |
Audit and Compliance |
QCCR1D137342 | Audits with compliance checks or custom scripts fail due to memory limits. | UNIX |
BSAE Integration |
QCCR1D134089 | The maximum value for the SAS_SERVER_CUST_ATTRIBUTES.ATTRIBUTE_NAME column is 255 bytes. If a larger value is added, the system returns an exception error similar to the following:
2011-01-30 06:32:20,285 [omdb.loader] ERROR Exception thrown while processing task: ORA-12899: value too large for column "CMDB_DATA"."SAS_SERVER_CUST_ATTRIBUTES"."ATTRIBUTE_NAME" (actual: 259, maximum: 255) The maximum value has been increased to 4000 bytes. |
Linux |
QCCR1D133815 | The audit Description column is not long enough. | Independent |
QCCR1D144611 | SA displays an error if you try to load certain data files into BSAE if the value is too large for the corresponding column. | Independent |
Command-Line Interface (CLI) |
||
QCCR1D118249 | When we upload a package like thisC:\Program
Files\Opsware\ocli\scripts> we get an error SUMMARY: |
Win2008 Win2008 R2 |
Data Center Markup Language (DCML) Export Tool (DET) |
QCCR1D144429 | Full export/import in the DCML Export Tool (DET) does not display a full set of data. | Linux |
Database and Middleware Automation |
||
QCCR1D130699 | Previously, the number of members in a Smart Group appeared to be zero whenever the group’s membership was being re-evaluated. Now, “Evaluating” is displayed when membership is being evaluated. | Independent |
QCCR1D134082 | The deployment name is now displayed on the History page in the web UI. Note that the deployment name is only displayed for deployments that are executed after SA 9.13 is installed. | Independent |
QCCR1D134082 | The deployment name is now displayed on the History page in the web UI. | Independent |
QCCR1D136903 | The SELECT ALL action on Run page now only selects targets that are visible in the Target check list. | Independent |
QCCR1D140035 | It is now possible to create "bridged execution" workflows in which different steps are executed on different targets. The targets are specified at run time. | Independent |
QCCR1D140572 | New users now correctly appear in the Roles preview pane on the Setup > Permissions page. | Independent |
QCCR1D140594 | The Targets view on the Deployment page can now filter by organization. | Independent |
QCCR1D140794 | Steps are now permitted to contain embedded CDATA sections. | Independent |
QCCR1D141826 | For Windows targets, you can now specify an administrator account name other than Administrator (the default). For configuration details, see http://support.openview.hp.com/selfsolve/QCCR1D141826. |
Independent |
QCCR1D141832 | Environment entities (servers, instances, or databases) no longer have a Policies tab. It is not possible to associate a policy with these entities, so the presence of the tab was confusing. | Independent |
QCCR1D142326 | The IMPORT SCRIPT link on the Step > Action page now works correctly. | Independent |
QCCR1D142572 | After a Save action, you now return to the page from which you initiated the Save. | Independent |
QCCR1D142887 | A file reading error was fixed. | Independent |
QCCR1D142946 | A scripting error was fixed. | Independent |
QCCR1D143013 | The USAGE pane on the Environment > Custom Fields page now displays the correct Custom Field values. | Independent |
QCCR1D143036 | A user can now execute a deployment against a target if that user has Execute permission on the deployment and Read permission on the organization that includes the target. | Independent |
QCCR1D143293 | Previously when a workflow is imported using Import Workflow, all of the script steps contained in the workflow were duplicated (with a numeric suffix) if a step with the same name already existed. The steps were duplicated even if the step that was imported was exactly the same as the existing step. Now, the import algorithm searches for steps with similar names in addition to identical names. In this case, similar means names that start with the same string but have additional information appended to it. If it is necessary to rename a step, a time stamp like (2012-03-05T12:45:00) is appended to the end of the step name. |
Independent |
QCCR1D143714 | Previously, when querying the results for a specific workflow run, the workflow status API calls only returned results for workflows that were currently running or had completed during the last minute. This time constraint was removed. | Independent |
QCCR1D143864 | A web service API was added to return a list of the servers that are not assigned to an organization. Note that only Database and Middleware Automation Administrators are allowed access to this web service. | Independent |
QCCR1D143949 | A problem with the Database and Middleware Automation configuration files was fixed. | Independent |
QCCR1D144115 | You can now view the members of a Smart Group on the Smart Groups page (previously, this information was only available on the Run page). You can also filter the list of members that you view by name. | Independent |
QCCR1D144669 | Smart Group membership is now checked before a scheduled deployment runs to ensure that the deployment uses the correct targets. | Independent |
QCCR1D144959 | The Server.OS property is now set correctly when a server is added to an organization using the web services interface | Independent |
Global Filesystem/Shell Backend |
QCCR1D137505 | UAPI method calls in the Global Shell sometimes return incorrect or unexpected results. | Independent |
Installation | ||
QCCR1D135514 | SA failed to reload Platforms UnitTypes list. An ocli.badParameters error is displayed. | Linux | QCCR1D141992 | SA incorrectly stores patch data in /var/opt/opsware. | Independent |
QCCR1D146493 | Suspended users have their DEFAULT_USER_GROUP role stripped during an upgrade. | Independent |
Model Repository |
QCCR1D140360 | SA will not start if the database on the RAC node used to install SA is unavailable. | Independent |
OS Provisioning |
||
QCCR1D102830 | Cannot enter timeout value for pre/post remediate scripts while creating a new OS sequence. | Independent |
QCCR1D111445 | Run OS Sequence does not escalate device group privilege. Run an OS Sequence that has an attached device group results in an exception and the job is not created. | Independent |
QCCR1D111245 | When you upgrade, unless you have customized the bm.reprovision_attributes_to_preserve system configuration value, your custom values will not be replaced by the new values required for this SA release - instead, old values are retained. |
Independent |
QCCR1D111781 | When provisioning Windows Server 2003 or 2008 using an OS Sequence to a non-C drive using a WIM image:
|
Windows Server 2003 |
QCCR1D111845 | Provisioned machines using MBC may continually reboot into the un-provisioned server pool after initial provisioning because the sequence_id and MAC link were not removed. | Solaris |
QCCR1D118372 | SA fails to complete OS provisioning if the net or subnet from which the machine obtains its IP address does not have PTR records associated with the IP address. | Red Hat Linux 4 and 5 Power |
QCCR1D122908 | Multipath IO OS Sequences failing on Linux RHEL5 (physical system). | Red Hat Enterprise Linux 5 |
QCCR1D129544 | SA does not support OS Sequence SuSE Linux Enterprise Server (SLES) Dynamic Host Configuration Protocol (DHCP)-less provisioning. | SLES |
QCCR1D137148 | If autoinst.xml already contains pre- and post-processing scripts, the resulting configuration file is not formed correctly. | Linux |
QCCR1D137944 | WinPE file-system agents connect to the wrong agent gateway. | Independent |
QCCR1D141114 | Setting up the Master Boot Code (MBC) for a decommissioned server does not place it into the server pool until provisioning begins. | Independent |
QCCR1D141762 | When multiple network parameter lines are used, some of the lines are dropped from the ks.cfg file. | Red Hat Linux |
QCCR1D141782 | Run OS sequence for OS install fails if the Server ID value is large. A traceback error is recorded in the buildmgr log. | Independent |
QCCR1D141438 | Run Manage Boot Client (MBC) jobs fail if the facility has not been set when you run the MBC Web APX. | Independent |
QCCR1D142556 | WinPE IS not working with G7 AMD Blades. | Independent |
QCCR1D143225 | Not able to Provision Windows OS on HP BL680c G5 from SA. | Windows |
QCCR1D143465 | You receive an error message ('mgmt failure') during ESXi5 build plan OS Provisioning. | ESXi |
QCCR1D145697 | OS build plans with hpsa_netconfig custom attribute fail on ESXi. | Independent |
Patch Management | ||
QCCR1D136541 | AIX 6.1 Patch policies do not function correctly.
The AIX patch installation was modified to install all patches in a policy in one bundle. |
AIX 6.1 |
QCCR1D146433 | If the managed server is low on storage space, sol_ips_import will display a stack trace error. |
Solaris |
QCCR1D141251 | Patch policy remediation process fails due to incorrect calculation of required space for staging and deploying patches.
The size calculation for the amount of space needed to stage and deploy packages is now a much closer approximation to the actual space needed. |
Solaris |
SA Client |
||
QCCR1D112476 | JRex is not fully W3C compatible, causing client executable codes failures. | Independent |
QCCR1D142731 | If you do not have permission to set device-group conditions, and you try to set the dynamic rule on a device group using the Library/Integrations/Hewlett-Packard/Server Automation/Device Groups/Set Device Group Dynamic Rule, the operation adds a space in front of each device's platform name, and the rule does not return any records. |
Independent |
SA Reports |
||
QCCR1D110658 | The “Application Deployment” and “Patch” folders and their reports in the BSA Essentials/SAR Client 2.0/2.01 installation patch are missing. These folder definitions are required to view the Application Deployment and Windows patch reports. | Independent |
SA-UCMDB Integration |
||
QCCR1D141512 | If you are running Discovery and Dependency Mapping Advanced Edition (DDMA), and both SA and DDMA are reporting data to UCMDB, some duplicate configuration Items might be created in UCMDB. These duplicates might include Node, Hypervisor and Installed Software.
DDMA automates discovery and dependency mapping of services, applications, and underlying infrastructure. |
Independent |
SA Web Client |
||
QCCR1D118068 | User’s preferred datetime format is: “星期五 十月 22, 2010 15:16 CST”, but in the SA Web Client, all dates and times are displayed as “Fri Oct 22,2010 14:04 CST” |
Win2008 Win2008 R2 |
QCCR1D131019 | After locales are enabled, the SA Web Client fails to start. | Windows |
Software Management Backend |
||
QCCR1D100754 | You cannot set the timeout value for the time it takes to install
or remove software or execute scripts to anything other than the
default value of 5 hours - the job times out at 5 hours. This
timeout value is specified by "way.remediate.action_timeout" in the SA Web Client. The job fails with the message "The request to retrieve information from the Agent failed because it timed out. If the problem persists, please contact your HP Server Automation Administrator." |
Independent |
QCCR1D119222 | Software remediations are stopped and do not progress until Command Engine scripts are restarted. | Linux;Solaris |
QCCR1D133403 | When scanning software compliance of a device group or multiple servers, the scan fails on Solaris servers with the following exception:
"SEVERE Exception in polling scan compliance status: java.rmi.ServerException: RemoteException occurred in server thread; nested exception is: " .... |
Solaris |
Software Management UI |
||
QCCR1D137723 | Remediation jobs become unresponsive when device groups in policies are empty. | Independent |
Web Services |
||
QCCR1D139539 | When trying to call the configureAutomaticProvisioningRule method on the com.opsware.osprov.OSSequenceService JBOSS web service interface, the following exception is thrown:
No such operation 'configureAutomaticProvisioningRule'. |
Independent |
QCCR1D141935 | Web Services Data Access Engine stops working during a JAVA Remote Method Invocation (RMI) job. The Web Services Data Access Engine includes the Oracle Weblogic patch-14391267 ( to fix issue: “BEA-101020 JAVA.LANG.ILLEGALARGUMENTEXCEPTION: PORT OUT OF RANGE”). |
Back to the table of contents
When a feature or platform is identified as deprecated for a release, it means that you (the SA customer) are notified of its future removal. Deprecated features are still fully supported in the release they are deprecated in, unless specified otherwise. The intent is that deprecated features or platforms will have support removed in the next major or minor SA release; however, eventual removal is at the discretion of HP.
The following sections list features that were deprecated in the SA 7.80 release. Current and future availability of these features is also described here.
In SA 9.0, moving application configurations and configuration templates into folders changed the behavior of the associated classes and services.
Attributes folder and lifecycle of the VO-s are required for the following API calls:
com.opsware.acm.ConfigurationService#create
com.opsware.acm.ConfigurationService#update
com.opsware.acm.CMLService#create
com.opsware.acm.CMLService#update
The following methods have been deprecated, but they still perform as expected. Method
com.opsware.folder.FolderVO#getCustomers should be used instead.
com.opsware.acm.ConfigurationVO#getCustomers
com.opsware.acm.CMLVO#getCustomers
The following methods have been deprecated and they have a void implementation. Method
com.opsware.folder.FolderVO#setCustomers must be used instead.
com.opsware.acm.ConfigurationVO#setCustomers
com.opsware.acm.CMLVO#setCustomers
Code Deployment and Rollback (CDR) was deprecated in the SA 7.80 release, but is still supported in SA 9.0. In a future release, this feature will not be supported. The new Application Deployment Manager in SA 9.0 is intended to replace CDR. See Application Deployment Manager in the SA Application Deployment User Guide for more information.
If you upgrade to SA 9.13 and also upgrade to an SA 9.13 agent, you will no longer be able to use CDR because it is only supported if you are using SA 9.01 (or earlier) agents. However, you will be able to use CDR with SA 9.13 if you uninstall the SA 9.13 agent and reinstall an SA 9.01 (or earlier) agent (not recommended, as you will be unable to use other SA 9.13 features).
For more information about the deprecation of CDR, contact your HP Technical Support representative.
In SA 9.0, HP announced the deprecation of the Server Automation Agent for the following versions:
• Server Automation 7.01
• Server Automation 7.00
• Server Automation 6.xx
• Server Automation 5.xx
Although releases of HP Server Automation prior to 9.0 will continue to support these versions of the Agent, it is recommended that customers migrate to newer versions of the Agent on their managed platforms.
HP SA will stop supporting these versions of the Agent in an upcoming major or minor release. Agents from Server Automation 4.xx and earlier are not supported in this release.
DOS-based OS Provisioning was deprecated in the SA 7.80 release and is not supported in SA 9.0 or later. For more information, contact your HP Technical Support representative.
As of SA 7.80, the following scripts are no longer supported:
start_opsware.sh
stop_opsware.sh
In SA 9.0 and later you must use the unified start script:
/etc/init.d/opsware-sas
If you have any applications or scripts that depend on this script, you must rewrite them to use the unified start script.
The Open Console action (for VMware virtual machines of ESX agent-managed hypervisor) was removed in the SA 9.0 release.
The Open Web Access action (for ESX agent-managed hypervisors) was removed in the SA 9.0 release.
As of SA 9.0, virtualization actions for agent-managed ESX 3.0 hypervisors are no longer supported, unless the ESX 3.0 hypervisor is managed by a Virtualization Service. It cannot be directly added to the Virtual Servers view (through Add Hypervisor); it needs to be vCenter managed.
This section discusses documentation information for this release.
In addition to the 9.10 guides on the Self Solve website, the following white papers and guides are relevant to this release:
This section describes documentation errata.
Property Name | Required? | Description | Example |
---|---|---|---|
agent_recert.using_cdr=<0 | 1> |
Yes | Specifies that the Agent CA push will be required when the value is set to 1. Also, the old Agent CA cleanup phase will be skipped unless the agent_recert.cleanup_old_agent_ca property is set to 1. Valid values are 1 (true) or 0 (false). Any other value will result in an invalid property error. Default: 1. Note: CDR refers to Code Deployment and Rollback which has been deprecated but may still be in use. |
agent_recert.using_cdr=1 |
agent_recert.cleanup |
No | Indicates whether to clean up the old AgentCA after Core Recertification.
Cleanup of old Agent CA phase is not necessary and can be disabled. The valid values are 1 (true) or 0 (false). Any other value will result in an invalid property error. This is an optional property. Default: 0. |
agent_recert.cleanup_old_agent_ca=0 |
In the Details: Start Order for SA Components list, item 16, opswgw-lb, should be removed. The opswgw-lb component was incorporated into the Core Gateway (opswgw-cgw) component.
In the Application Deployment Manager, context-sensitive online help is
provided for numerous dialogs, including the Manage Applications and Manage
Targets dialogs.
To view a context-sensitive help topic, click the question mark icon in the
dialog. Note that the F1 key does not open online help for the Application
Deployment Manager.
To view the portion of the SA online help that pertains to application
deployment, select Help -> Help in the Application
Deployment Manager.
Refer to the HP Server Automation Application Deployment Manager User Guide for additional information.
On page 25, Network Reports section, disregard the following reports:
The only reports available in SA are "Connections by Network Device" and "Connections by Server".
Enabling IBM POWER6 Provisioning with the DHCPD Tool
Due to hardware constraints, IBM POWER6 servers cannot be provisioned using the normal OS Provisioning procedures. However, you can enable provisioning for these servers using the dhcpdtool provided with SA.
This involves using the dhcpdtool to set a parameter in the initial subnet declaration which enables Linux OS provisioning on the IBM POWER6 server's hardware. This workaround also requires that SA be in Full DHCP Management mode.
Note: Setting this parameter prevents Solaris SPARC OS provisioning. You will see a warning message to this effect when you change the parameter.
Perform the following tasks to enable Linux OS Provisioning on IBM POWER6 hardware:
1. Start dhcpdtool.
/opt/opsware/dhcpd/sbin/dhcpdtool
2. Select Full DHCP Management.3. Select the network for which the workaround is to be applied. An editing menu is displayed:
Opsware DHCP Network Configuration
Editing DHCP information for 192.168.208.160/27 (255.255.255.224)
1) gateway : 192.168.208.161
2) DHCP range : 192.168.208.163 - 192.168.208.190
3) DNS servers: 192.168.194.4
4) DNS domain : dev.opsware.com
5) Power6 provisioning override: No
!!! WARNING: Option 5) breaks Solaris SPARC OS provisioning !!!
!!! Make sure to deactivate this option before trying to provision a Solaris SPARC machine !!!1..5 to edit option.
d)elete network and return to main menu.
k)eep network and return to main menu.Choice [1..5, d, k]:
4. Select option 5 and press ENTER
5. The message “Are you sure that you want to toggle POWER6 provisioning although it breaks Solaris SPARC provisioning (Enable/Disable)?” is displayed.
6. Press E to Enable POWER6 provisioning, D to Disable, select K to keep the settings, then S to save them.
7. Exit the DHCPD tool. Upon exiting, the DHCPD server is restarted and the new setting takes effect.
Note: If you are not using the SA-provided DHCPD server, you must add the following line to the subnet in which POWER6 you will perform OS provisioning:
filename "yaboot";
Step 4: The following bullet should be removed, as the truth_data.tar.gz is no longer a valid file:
Chapter 8: Satellite Installation
Remove the following bullet:
- word_tmp_dir: directory where Package Repository will temporarily place content during uploads
- word.store.host: the host name of the server where Software Repository content is stored.
- media_server.windows_share_name: the share name to use for the Windows media sharing server (note: share names that are longer than 8 characters may give errors while browsing or may not be accessible to some older clients.)
- media_server.windows_share_password: the password to write-protect the Windows media share. Import_media tool will prompt for this password each time it is run.
Under the heading "Uninstall a Single Core in a Multimaster Mesh", remove step 1a. This step is no longer required.
The DHCP setup step should be removed from the HP-UX OS Provisioning section of the SA 9.10 User Guide: OS Provisioning. As of SA 9.05 and 9.11, DHCP is enabled by default and the OS Provisioning menu now includes HP-UX.
Complete HP-UX OS Provisioning instructions are provided in the SA HP-UX OS Provisioning Whitepaper, which can be downloaded for any relevant SA 9.x release from the HP Support Self-Solve website: http://support.openview.hp.com/selfsolve/manuals or http://support.openview.hp.com/selfsolve/documents
In chapter 4: Patch Management for Solaris, page 108, Fujitsu Clusters, SA Commands, replace the following statement:
Use the following command to display additional information about cluster commands:
/opt/opsware/solpatch_import -manual
With this statement:
Use the following command to display additional information about cluster commands:
/opt/opsware/solpatch_import/bin/solpatch_import --manual
Installing the opsh RPM places the OGFS shell and the Agent Upgrade Tool in the directory /opt/opsware/opsh/bin.
# cd /opt/opsware/opsh/bin
Please note the following documentation errors:
Back to the Table of Contents
This web site provides contact information and details about the products, services, and support that HP Software offers. For more information, visit the HP Support web site at: HP Software Support Online.
HP Software support provides customer self-solve capabilities. It provides a fast and efficient way to access interactive technical support tools needed to manage your business. As a valued support customer, you can benefit by being able to:
To access the Self-Solve knowledge base, visit the Self-Solve knowledge search home page.
Note: Most of the support areas require that you register as an HP Passport user and sign in. Many also require an active support contract. To find more information about support access levels, go to: Access levels.
To register for an HP Passport ID, go to: HP
Passport Registration.
The only warranties for HP products and services are set forth in the express warranty statements accompanying such products and services. Nothing herein should be construed as constituting an additional warranty. HP shall not be liable for technical or editorial errors or omissions contained herein.
The information contained herein is subject to change without notice.
Confidential computer software. Valid license from HP required for possession, use or copying. Consistent with FAR 12.211 and 12.212, Commercial Computer Software, Computer Software Documentation, and Technical Data for Commercial Items are licensed to the U.S. Government under vendor's standard commercial license.
© Copyright 2000-2012 Hewlett-Packard Development Company, L.P.
Adobe® is a trademark of Adobe Systems Incorporated.
Intel® and Itanium®
are trademarks of Intel Corporation in the U.S. and other
countries.
Microsoft®, Windows®‚ Windows® XP are U.S. registered trademarks
of Microsoft Corporation.
Oracle and Java are registered trademarks of Oracle
and/or its affiliates.
UNIX® is a registered trademark of The Open
Group.
New Publication Date | Change |
06-11-2012 | Added Information on Patch Importer for Oracle Enterprise Linux (OEL). Added information about Remediation Performance and Scalability component. |
06-28-2012 | Added paragraph: New Delivery Method for OS Provisioning Build Plan Content Added paragraph on Remediation Performance and Scalability (missing component) |
07-23-2012 | Corrected G8 references to Gen8, Proliant references to HP ProLiant. |
08-09-2012 | Added Remediation Performance and Scalability hotfix clarification (hotfix should be obtained from HP Support). |
10-30-2012 | Added information about patch de-duplication steps for Windows patching in the What's New section. |
To check for recent updates or to verify that you are using the most recent
edition of a document, go to:
http://support.openview.hp.com/selfsolve/manuals
This site
requires that you register for an HP Passport and sign in.
Or click the New
users - please register link on the HP Passport login page. You will also
receive updated or new editions if you subscribe to the appropriate product
support service.
Contact your HP sales representative for details.
Back to the Table of Contents