1 环境概要
环境概要说明:因此次资源不足,仅有两台服务器进行Oracle OVM测试安装,故部分功能无法实现,如在线迁移、异常OVS异常模拟等。
1.1 硬件信息
物理服务器 2 台, HP DL 360 G7,颗 4核 CPU(型号不详),32G内存,2×300G SAS 盘,服务器没有 HBA 卡,存储不支持 iSCSI,因此没办法提供 SAN 存储。
1.2 安装配置
此次我们采用linux下的KVM方式进行ovm和oem的虚拟配置,如下:
实体服务器 | 实体机ip | 实体机系统 | 虚拟机ip | 用途 | 操作系统 |
服务器一 | 10.0.57.11 | ORL6.5(OVS) | 10.0.57.12--17 | hypervisor | 随虚拟机而定 |
服务器二 | 10.0.57.7 | ORL6.3 | 10.0.57.8 | OVM 管理机 | ORL6.5 |
10.0.57.9 | OEM 管理机 | ORL6.5 |
申明:本文档基于ORACLE OVM 3.2版本编写,与部分版本在界面上存在一定的差异,功能上类似。
OVM以及OVS的安装过程此处不做介绍,下面直接进入使用介绍。
2 界面窗体概要
2.1 登陆窗口
登陆OVM控制台https://10.0.57.8:7002/ovm/console,在配置中,端口信息采用默认端口7002.数据账号、密码即可,默认管理账号admin,密码为安装时设定密码。
登陆后最初显示
2.2 Servers and VMs 界面
2.3 Respositories 界面
2.4 Networkings界面:
2.5 Storage界面:
2.6 Tools and resource界面:
2.7 Jobs界面:
3 各界面使用详解
3.1 health界面
显示为绿色表示OVS运行ok,负载等均在阀值范围内,橙色表色警告,红色表示负载等严重超过阀值,假设有两台OVS则会显示如下图示表示其中一台出现严重问题。
例如:
3.2 Statistics介绍
点解界面的statistics出现如下界面信息,可以看到服务器池,服务器池上的虚拟机、刷新频率、状态信息。
3.3 Servers and VMs界面介绍
左侧为服务池以及创建服务池等按钮,依次为:discovery servers、create VNCs、create server pool、create virtual machine、find。
同样右击serverpools也可以看到相应的信息。
右击server pool中相应的server pool可以得到如下信息:
同样在server pool池中的OVS上右击看到如下信息:
3.4 Respositories介绍
在资源库中可以看到VM模板、上传的ios镜像文件,虚拟磁盘、VM文件等信息。
同时也可以创建新的资料库:
在右侧主窗口中可以看到上传ISO文件选项:
添加新的镜像文件,这里支持HTTP和FTP 协议的导入方式:
3.5 Networks介绍
网络情况点击Networks,显示如下信息:
点击VLAN Groups 显示如下信息,此外+信息为create New Vlan Goup如下:
点击Virtual NICs显示如下信息,可以看到该处就可以虚拟多个网卡信息,可批量创建:
3.6 存储池介绍
左侧窗口中+依次表示:
右击相应的SAN存储有如下信息:
3.7 tools and sources 介绍
在tools and sources中有如下信息下图为Tags信息:
点击NTP按钮有如下信息:
Yum管理:
Preferences信息:
3.8 Jobs信息
Recurring:
4 虚拟机的创建
4.1 模板创建虚拟机
4.2 镜像文件安装
在使用镜像文件安装虚拟机时需要在disk选项中加入CD/ROM,加载ISO文件如下所示
4.3 利用RAC模板创建RAC环境
4.3.1 创建RAC虚拟机
在repositories下的模板库中选定模板,然后点击clone or move template按钮即可:
4.3.2 修改网卡
因默认根据模板创建的虚拟机中只包含一个网卡信息、此时我们需要增加额外的网卡信息已符合oracle RAC创建的需求。
4.3.3 添加共享存储
同样在RAC模板应用安装中,我们需要给各节点添加共享存储,以边ORACLE RAC votedisk以及数据存储使用,首先创建共享存储:
分配共享存储给个节点服务器,如下图所示:
根据同样步骤一次添加其他节点的共享存储,即可完成共享存储的创建,
注意:全虚拟化PVM的情况下,最多添加的存储个数不能超过3
4.3.4 各文件配置
接着在管理机中编辑配置文件,完成oracle RAC的模板创建,注意此处需要我们自行到官方网址下载相应的安装脚本以完成ORACLE RAC 的配置安装过程,如下所示:
注意根据文件名判断自己需要的配置文件模板:
[root@ovmm utils]# more netconfig-sample64-11g.ini
# Node specific information -------------节点网络配置
NODE1=test13 --------------节点名称
NODE1IP=192.168.1.231 --------------节点ip
NODE1PRIV=test13-priv ------------节点私有名称
NODE1PRIVIP=10.10.10.231 ------------节点私有ip
NODE1VIP=test13-vip ------------节点vip名称
NODE1VIPIP=192.168.1.233 ------------节点vip
NODE2=test14
NODE2IP=192.168.1.232
NODE2PRIV=test14-priv
NODE2PRIVIP=10.10.10.232
NODE2VIP=test14-vip
NODE2VIPIP=192.168.1.234
# Common data
PUBADAP=eth0 ----------节点通用网卡2
PUBMASK=255.255.255.0 ----------子网掩码
PUBGW=192.168.1.1 ---------网关
PRIVADAP=eth1 ----------网卡2
PRIVMASK=255.255.255.0
RACCLUSTERNAME=crs64bitR2
DOMAINNAME=localdomain # May be blank
DNSIP= # Starting from 2013 Templates allows multi value
# Device used to transfer network information to second node
# in interview mode
NETCONFIG_DEV=/dev/xvdc ---------网络配置设备
# 11gR2 specific data
SCANNAME=test13-14-scan ----------scan 名称
SCANIP=192.168.1.235 ----------scan ip
# Single Instance (description in params.ini)
# CLONE_SINGLEINSTANCE=yes # Setup Single Instance -------
# CLONE_SINGLEINSTANCE_HA=yes # Setup Single Instance/HA (Oracle Restart)
根据自己的需求修改该文件中的信息,建议修改之前事先备份该文件或者cp一个自己的配置文件:例如:
4.3.4.1 网络文件的配置
[root@ovmm utils]# more netconfig-my.ini ------------选择网络的配置文件.
# Node specific information
NODE1=RAC-1
NODE1IP=9.9.9.1
NODE1PRIV=RAC-1-Priv
NODE1PRIVIP=11.11.11.1
NODE1VIP=RAC-1-VIP
NODE1VIPIP=9.9.9.3
NODE2=RAC-2
NODE2IP=9.9.9.2
NODE2PRIV=RAC-2-Priv
NODE2PRIVIP=11.11.11.2
NODE2VIP=RAC-2-VIP
NODE2VIPIP=9.9.9.4
# Common data
PUBADAP=eth0
PUBMASK=255.255.255.0
PUBGW=9.9.9.10
PRIVADAP=eth1
PRIVMASK=255.255.255.0
RACCLUSTERNAME=my-cluster
DOMAINNAME=localdomain # May be blank
DNSIP= # Starting from 2013 Templates allows multi value
# Device used to transfer network information to second node
# in interview mode
NETCONFIG_DEV=/dev/xvdc
# 11gR2 specific data
SCANNAME=SCAN-my-cluster
SCANIP=9.9.9.11
# Single Instance (description in params.ini)
# CLONE_SINGLEINSTANCE=yes # Setup Single Instance
# CLONE_SINGLEINSTANCE_HA=yes # Setup Single Instance/HA (Oracle Restart)
4.3.4.2 参数文件的配置
[root@ovmm utils]# more params-sample11g.ini ----------选择11g参数文件配置
#
#/* Copyright 2009-2013, Oracle. All rights reserved. */
#
#
# WRITTEN BY: Oracle.
# v1.6: Jul-2013 Add Single Instance, Policy managed DB, Low memory support & DB on Filesystem
# v1.5: Aug-2012 Add resolver options
# v1.4: May-2012 Add colored logfile & unlock accounts
# v1.3: Aug-2011 Document Clusterware only
# v1.2: Jun-2011 Relink on major OS change & Post SQL scripts
# v1.1: Feb-2011 Added options for multicast checking
# v1.0: Jul-2010 Creation
#
#
# Oracle DB/RAC 11gR2 OneCommand for Oracle VM - Generic configuration file
# For Single Instance, Single Instance HA (Oracle Restart) and Oracle RAC
#
##############################################
#
# Generic Parameters
#
# NOTE: The first section holds more advanced parameters that
# should be modified by advanced users or if instructed by Oracle.
#
# See further down this file for the basic user modifiable parameters.
#
##############################################
#
# Temp directory (for OUI), optional
# Default: /tmp
TMPDIR="/tmp"
#
# Progress logfile location
# Default: $TMPDIR/progress-racovm.out
LOGFILE="$TMPDIR/progress-racovm.out"
#
# Must begin with a "+", see "man 1 date" for valid date formats, optional.
# Default: "+%Y-%m-%d %T"
LOGFILE_DATE_FORMAT=""
#
# Should 'clone.pl' be used (default no) or direct 'attach home' (default yes)
# to activate the Grid & RAC homes.
# Attach is possible in the VM since all relinking was done already
# Certain changes may still trigger a clone/relink operation such as switching
# from role to non-role separation.
# Default: yes
CLONE_ATTACH_DBHOME=yes
CLONE_ATTACH_GIHOME=yes
#
# Should a re-link be done on the Grid & RAC homes. Default is no,
# since the software was relinked in VM already. Setting it to yes
# forces a relink on both homes, and overrides the clone/attach option
# above by forcing clone operation (clone.pl)
# Default: no
CLONE_RELINK=no
#
# Should a re-link be done on the Grid & RAC homes in case of a major
# OS change; Default is yes. In case the homes are attached to a different
# major OS than they were linked against, a relink will be automatically
# performed. For example, if the homes were linked on OL5 and then used
# with an OL6 OS, or vice versa, a relink will be performed. To disable
# this automated relinking during install (cloning step), set this
# value to no (not recommended)
# Default: yes
CLONE_RELINK_ON_MAJOR_OS_CHANGE=yes
#
# The root of the oracle install must be an absolute path starting with a /
# Default: /u01/app
RACROOT="/u01/app"
#
# The location of the Oracle Inventory
# Default: $RACROOT/oraInventory
RACINVENTORYLOC="${RACROOT}/oraInventory"
#
# The location of the SOFTWARE base
# In role separated configuration GIBASE may be defined to set the location
# of the Grid home which defaults to $RACROOT/$GRIDOWNER.
# Default: $RACROOT/$RACOWNER
RACBASE="${RACROOT}/oracle"
#
# The location of the Grid home, must be set in RAC or Single Instance HA deployments
# Default: $RACROOT/11.2.0/grid
GIHOME="${RACROOT}/11.2.0/grid"
#
# The location of the DB RAC home, must be set in non-Clusterware only deployments
# Default: ${RACBASE}/product/11.2.0/dbhome_1
DBHOME="${RACBASE}/product/11.2.0/dbhome_1"
#
# The disk string used to discover ASM disks, it should cover all disks
# on all nodes, even if their physical names differ. It can also hold
# ASMLib syntax, e.g. ORCL:VOL*, and have as many elements as needed
# separated by space, tab or comma.
# Do not remove the "set -/+o noglob" options below, they are required
# so that discovery string don't expand on assignment.
set -o noglob
RACASMDISKSTRING="/dev/xvd[c-g]1"
set +o noglob
#
# Provide list of devices or actual partitions to use. If actual
# partition number is specified no partitioning will be done, otherwise specify
# top level device name and the disk will automatically be partitioned with
# one partition using 'parted'. For example, if /dev/xvdh4 is listed
# below it will be used as is, if it does not exist an error will be raised.
# However, if /dev/xvdh is listed it will be automatically partitioned
# and /dev/xvdh1 will be used.
# Minimum of 5 devices or partitions are recommended (see ASM_MIN_DISKS).
#注意修改下面的信息,防止磁盘不够或者名称不一致。
ALLDISKS="/dev/xvdc /dev/xvdd /dev/xvde /dev/xvdf /dev/xvdg"
#
# Provide list of ASMLib disks to use. Can be either "diskname" or
# "ORCL:diskname". They must be manually configured in ASMLib by
# mapping them to correct block device (this part is not yet automated).
# If you include any disks here they should also be included
# in RACASMDISKSTRING setting above (discovery string).
ALLDISKS_ASMLIB=""
#
# By default 5 disks for ASM are recommended to provide higher redundancy
# for OCR/Voting files. If for some reason you want to use less
# disks, then uncomment ASM_MIN_DISKS below and set to the new minimum.
# Make needed adjustments in ALLDISKS and/or ALLDISKS_ASMLIB above.
# Default: 5
#ASM_MIN_DISKS=5 -------注意安装时一定要手动修改,防止磁盘不够。
#
# By default, whole disks specified in ALLDISKS will be partitioned with
# one partition. If you prefer not to partition and use whole disk, set
# PARTITION_WHOLE_DISKS to no. Keep in mind that if at a later time
# someone will repartition the disk, data may be lost. Probably better
# to leave it as "yes" and signal it's used by having a partition created.
# Default: yes
PARTITION_WHOLE_DISKS=yes
#
# By default, disk *names* are assumed to exist with same name on all nodes, i.e
# all nodes will have /dev/xvdc, /dev/xvdd, etc. It doesn't mean that the *ordering*
# is also identical, i.e. xvdc can really be xvdd on the other node.
# If such persistent naming (not ordering) is not the case, i.e node1 has
# xvdc,xvdd but node2 calls them: xvdn,xvdm then PERSISTENT_DISKNAMES should be
# set to NO. In the case where disks are named differently on each node, a
# stamping operation should take place (writing to second sector on disk)
# to verify if all nodes see all disks.
# Stamping only happens on the node the build is running from, and backup
# is taken to $TMPDIR/StampDisk-backup-diskname.dd. Remote nodes read the stamped
# data and if all disks are discovered on all nodes the disk configuration continues.
# Default: yes
PERSISTENT_DISKNAMES=yes
#
# This parameter decides whether disk stamping takes place or not to discover and verify
# that all nodes see all disks. Stamping is the only way to know 100% that the disks
# are actually the same ones on all nodes before installation begins.
# The master node writes a unique uuid to each disk on the second sector of the disk,
# then remote nodes read and discover all disks.
# If you prefer not to stamp the disks, set DISCOVER_VERIFY_REMOTE_DISKS_BY_STAMPING to
# no. However, in that case, PERSISTENT_DISKNAMES must be set to "yes", otherwise, with
# both parameters set to "no" there is no way to calculate the remote disk names.
# The default for stamping is "yes" since in Virtual machine environments, scsi_id(8)
# doesn't return data for disks.
# Default: yes
DISCOVER_VERIFY_REMOTE_DISKS_BY_STAMPING=yes
#
# Permissions and ownership files, EL4 uses PERMISSIONFILE, EL5 uses UDEVFILE
UDEVFILE="/etc/udev/rules.d/99-oracle.rules"
PERMISSIONFILE="/etc/udev/permissions.d/10-oracle.permissions"
#
# Disk permissions to be set on ASM disks use if want to override the below default
# Default: "660" (owner+group: read+write)
# It may be possible in Non-role separation to use "640" (owner: read+write, group: read)
# however, that is not recommended since if a new database OS user
# is added at a later time in the future, it will not be able to write to the disks.
#DISKPERMISSIONS="660"
#
# ASM's minimum allocation unit (au_size) for objects/files/segments/extents of the first
# diskgroup, in some cases increasing to higher values may help performance (at the
# potential of a bit of space wasting). Legal values are 1,2,4,8,16,32 and 64 MB.
# Not recommended to go over 8MB. Currently if initial diskgroup holds OCR/Voting then it's
# maximum possible au_size is 16MB. Do not change unless you understand the topic.
# Most releases default to 1MB (Exadata's default: 4MB)
#RACASM_AU_SIZE=1
#
# Should we align the ASM disks to a 1MB boundary.
# Default: yes
ALIGN_PARTITIONS=yes
#
# Should partitioned disks use the GPT partition table
# which supported devices larger than 2TB.
# Default: msdos
PARTITION_TABLE_GPT=no
#
# These are internal functions that check if a disk/partition is held
# by any component. They are run in parallel on all nodes, but in sequence
# within a node. Do not modify these unless explicitly instructed to by Oracle.
HELDBY_FUNCTIONS=(HeldByRaid HeldByAsmlib HeldByPowerpath HeldByDeviceMapper HeldByUser HeldByFilesystem HeldBySwap)
#
##### STORAGE: Filesystem: DB/RAC: (shared) filesystem
#
# NOTE1: To not configure ASM unset RACASMGROUPNAME
# NOTE2: Not all operations/verification take place in a
# FS configuration.
# For example:
# - The mount points are not automatically created/mounted
# - Best effort verification is done that the correct
# mount options are used.
#
# The filesystem directory to hold Database files (control, logfile, etc.)
# For RAC it must be a shared location (NFS, OCFS or in 12c ACFS),
# otherwise it may be a local filesystem (e.g. ext4).
# For NFS make sure mount options are correct as per docs
# such as Note:359515.1
# Default: None (Single Instance: $RACBASE/oradata)
#FS_DATAFILE_LOCATION=/nfs/160
#
# Should the database be created in the FS location mentioned above.
# If value is unset or set to no, the database is created in ASM.
# Default: no (Single Instance: yes)
#DATABASE_ON_FS=no
#
# Should the above directory be cleared from Clusterware and Database
# files during a 'clean' or 'cleanlocal' operation.
# Default: no
#CLONE_CLEAN_FS_LOCATIONS=no
#
# Names of OCR/VOTE disks, could be in above FS Datafile location
# or a different properly mounted (shared) filesystem location
# Default: None
#CLONE_OCR_DISKS=/nfs/160/ocr1,/nfs/160/ocr2,/nfs/160/ocr3
#CLONE_VOTING_DISKS=/nfs/160/vote1,/nfs/160/vote2,/nfs/160/vote3
#
# Location of OCR/VOTE disks. Value of "yes" means inside ASM
# whereas any other value means the OCR/Voting reside in CFS
# (above locations must be supplied)
# Default: yes
#CLONE_OCRVOTE_IN_ASM=yes
#
# Should addnodes operation COPY the entire Oracle Homes to newly added
# nodes. By default no copy is done to speed up the process, however
# if existing cluster members have changed (patches applied) compared
# to the newly created nodes (using the template), then a copy
# of the Oracle Homes might be desired so that the newly added node will
# get all the latest modifications from the current members.
# Default: no
CLONE_ADDNODES_COPY=no
#
# Should an add node operation fully clean the new node before adding
# it to the cluster. Setting to yes means that any lingering running
# Oracle processes on the new node are killed before the add node is
# started as well as all logs/traces are cleared from that node.
# Default: no
CLONE_CLEAN_ON_ADDNODES=no
#
# Should a remove node operation fully clean the removed node after removing
# it from the cluster. Setting to yes means that any lingering running
# Oracle processes on the removed node are killed after the remove node is
# completed as well as all logs/traces are cleared from that node.
# Default: no
CLONE_CLEAN_ON_REMNODES=no
#
# Should 'cleanlocal' request prompt for confirmation if processes are running
# Note that a global 'clean' will fail if this is set to 'yes' and processes are running
# this is a designed safegaurd to protect environment from accidental removal.
# Default: yes
CLONE_CLEAN_CONFIRM_WHEN_RUNNING=yes
#
# Should the recommended oracle-validated or oracle-rdbms-server-*-preinstall
# be checked for existance and dependencies during check step. If any missing
# rpms are found user will need to use up2date or other methods to resolve dependencies
# The RPM may be obtained from Unbreakable Linux Network or http://oss.oracle.com
# Default: yes
CLONE_ORACLE_PREREQ_RPM_REQD=yes
#
# Should the "verify" actions of the above RPM be run during buildcluster.
# These adjust kernel parameters. In the VM everything is pre-configured hence
# default is not to run.
# Default: no
CLONE_ORACLE_PREREQ_RPM_RUN=no
#
# By default after clusterware installation CVU (Cluster Verification Utility)
# is executed to make sure all is well. Setting to 'yes' will skip this step.
# Set CLONE_SKIP_CVU_POSTHAS for SIHA (Oracle Restart) environments
# Default: no
#CLONE_SKIP_CVU_POSTCRS=no
#
# Allows to skip minimum disk space checks on the
# Oracle Homes (recommended not to skip)
# Default: no
CLONE_SKIP_DISKSPACE_CHECKS=no
#
# Allows to skip minimum memory checks (recommended not to skip)
# Default: no
CLONE_SKIP_MEMORYCHECKS=no
#
# On systems with extreme memory limitations, e.g. VirtualBox, it may be needed
# to disable some Clusterware components to release some memory. Workload
# Management, Cluster Health Monitor & Cluster Verification Utility are
# disabled if this option is set to yes.
# This is only supported for production usage with Clusterware only installation.
# Default: no
CLONE_LOW_MEMORY_CONFIG=no
#
# By default on systems with less than 4GB of RAM the /dev/shm will
# automatically resize to fit the specified configuration (ASM, DB).
# This is done because the default of 50% of RAM may not be enough. To
# disable this functionality set CLONE_TMPFS_SHM_RESIZE_NEVER=yes.
# Default: no
CLONE_TMPFS_SHM_RESIZE_NEVER=no
#
# To disable the modification of /etc/fstab with the calculated size of
# /dev/shm, set CLONE_TMPFS_SHM_RESIZE_MODIFY_FSTAB=no. This may mean that
# some instances may not properly start following a system reboot.
# Default: yes
CLONE_TMPFS_SHM_RESIZE_MODIFY_FSTAB=yes
#
# Setting CLONE_CLUSTERWARE_ONLY to yes allows Clusterware only installation
# any operation to create a database or reference the DB home are ignored.
# Default: no
#CLONE_CLUSTERWARE_ONLY=no
#
# As described in the 11.2.0.2 README as well as Note:1212703.1 mutlicasting
# is required to run Oracle RAC starting with 11.2.0.2. If this check fails
# review the note, and remove any firewall rules from Dom0, or re-configure
# the switch servicing the private network to allow multicasting from all
# nodes to all nodes.
# Default: yes
CLONE_MULTICAST_CHECK=yes
#
# Should a multicast check failure cause the build to stop. It's possible to
# perform the multicast check, but not stop on failures.
# Default: yes
CLONE_MULTICAST_STOP_ON_FAILURE=yes
#
# List of multicast addresses to check. By default 11.2.0.2 supports
# only 230.0.1.0, however with fix for bug 9974223 or bundle 1 and higher
# the software also supports multicast address 244.0.0.251. If future
# software releases will support more addresses, modify this list as needed.
# Default: "230.0.1.0 224.0.0.251"
CLONE_MULTICAST_ADDRESSLIST="230.0.1.0 224.0.0.251"
#
# The text specified in the NETCONFIG_RESOLVCONF_OPTIONS variable is written to
# the "options" field in the /etc/resolv.conf file during initial network setup.
# This variable can be set here in params.ini, or in netconfig.ini having the same
# effect. It should be a space separated options as described in "man 5 resolv.conf"
# under the "options" heading. Some useful options are:
# "single-request-reopen attempts:x timeout:x" x being a digit value.
# The 'single-request-reopen' option may be helpful in some environments if
# in-bound ssh slowness occur.
# Note that minimal validation takes place to verify the options are correct.
# Default: ""
#NETCONFIG_RESOLVCONF_OPTIONS=""
#
##################################################
#
# The second section below holds basic parameters
#
##################################################
#
# Configures a Single Instance environment, including a database as
# specified in BUILD_SI_DATABASE. In this mode, no Clusterware or ASM will be
# configured, hence all related parameters (e.g. ALLDISKS) are not relevant.
# The database must reside on a filesystem.
# This parameter may be placed in netconfig.ini for simpler deployment.
# Default: no
#CLONE_SINGLEINSTANCE=no
#
# Configures a Single Instance/HA environment, aka Oracle Restart, including
# a database as specified in BUILD_SI_DATABASE. The database may reside in
# ASM (if RACASMGROUPNAME is defined), or on a filesystem.
# This parameter may be placed in netconfig.ini for simpler deployment.
# Default: no
#CLONE_SINGLEINSTANCE_HA=no
#
# OS USERS AND GROUPS FOR ORACLE SOFTWARE
#
# SYNTAX for user/group are either (VAR denotes the variable names below):
# VAR=username:uid OR: VAR=username
# VARID=uid
# VAR=groupname:gid OR: VAR=groupname
# VARID=gid
#
# If uid/gid are omitted no checks are made nor users created if need be.
# If uid/gid are supplied they should be numeric and not clash
# with existing uid/gids defined on the system already.
# NOTE: In RAC usernames and uid/gid must match on all cluster nodes,
# the verification process enforces that only if uid/gid's
# are given below.
#
# If incorrect configuration is detected, changes to users and groups are made to
# correct them. If this is set to "no" then errors are reported
# without an attempt to fix them.
# (Users/groups are never dropped, only added or modified.)
# Default: yes
CREATE_MODIFY_USERS_GROUPS=yes
#
# NON-ROLE SEPARATED:------------默认的角色不分离
# No Grid user is defined and all roles are set to 'dba'
RACOWNER=oracle:1101
OINSTALLGROUP=oinstall:1000
GIOSASM=dba:1031
GIOSDBA=dba:1031
GIOSOPER=dba:1031
DBOSDBA=dba:1031
DBOSOPER=dba:1031
#
# ROLE SEPARATION: (uncomment lines below)---角色分离,需手动撤销注销
# See Note:1092213.1
# (Numeric changes made to uid/gid to reduce the footprint and possible clashes
# with existing users/groups)
#
##GRIDOWNER=grid:1100
##RACOWNER=oracle:1101
##OINSTALLGROUP=oinstall:1000
##GIOSASM=asmadmin:1020
##GIOSDBA=asmdba:1021
##GIOSOPER=asmoper:1022
##DBOSDBA=dba:1031
##DBOSOPER=oper:1032
#
# The name for the Grid home in the inventory
# Default: OraGrid11gR2
#GIHOMENAME="OraGrid11gR2"
#
# The name for the DB/RAC home in the inventory
# Default: OraRAC11gR2 (Single Instance: OraDB11gR2)
#DBHOMENAME="OraRAC11gR2"
#
# The name of the ASM diskgroup, default "DATA"
# If unset ASM will not be configured (see filesystem section above)
# Default: DATA
RACASMGROUPNAME="DATA"
#
# The ASM Redundancy for the diskgroup above
# Valid values are EXTERNAL, NORMAL or HIGH
# Default: NORMAL (if unset)
RACASMREDUNDANCY="EXTERNAL"
#
# Allows running the Clusterware with a different timezone than the system's timezone.
# If CLONE_CLUSTERWARE_TIMEZONE is not set, the Clusterware Timezone will
# be set to the system's timezone of the node running the build. System timezone is
# defined in /etc/sysconfig/clock (ZONE variable), if not defined or file missing
# comparison of /etc/localtime file is made against the system's timezone database in
# /usr/share/zoneinfo, if no match or /etc/localtime is missing GMT is used. If you
# want to override the above logic, simply set CLONE_CLUSTERWARE_TIMEZONE to desired
# timezone. Note that a complete timezone is needed, e.g. "PST" or "EDT" is not enough
# needs to be full timezone spec, e.g. "PST8PDT" or "America/New_York".
# This variable is only honored in 11.2.0.2 or above
# Default: OS
#CLONE_CLUSTERWARE_TIMEZONE="America/Los_Angeles"
#
# Create an ACFS volume?
# Default: no
ACFS_CREATE_FILESYSTEM=no
#
# If ACFS volume is to be created, this is the mount point.
# It will automatically get created on all nodes.
# Default: /myacfs
ACFS_MOUNTPOINT="/myacfs"
#
# Name of ACFS volume to optionally create.
# Default: MYACFS
ACFS_VOLNAME="MYACFS"
#
# Size of ACFS volume in GigaBytes.
# Default: 3
ACFS_VOLSIZE_GB="3"
#
# NOTE: In the OVM3 enhanced RAC Templates when using deploycluster
# tool (outside of the VMs). The correct and secure way to transfer/set the
# passwords is to remove them from this file and use the -P (--params)
# flag to transfer this params.ini during deploy operation, in which
# case the passwords will be prompted, and sent to all VMs in a secure way.
# The password that will be set for the ASM and RAC databases
# as well as EM DB Console and the oracle OS user.
# If not defined here they will be prompted for (only once)
# at the start of the build. Required to be set here or environment
# for silent mode.
# Use single quote to prevent shell parsing of special characters.
RACPASSWORD='oracle'
GRIDPASSWORD='oracle'
#
# Password for 'root' user. If not defined here it will be prompted
# for (only once) at the start of the build.
# Assumed to be same on both nodes and required to be set here or
# environment for silent mode.
# Use single quote to prevent shell parsing of special characters.
ROOTUSERPASSWORD='ovsroot'
# 上面信息是root的默认秘密,注意时候修改或者提前设定
# Build Database? The BUILD_RAC_DATABASE will build a RAC database and
# BUILD_SI_DATABASE a single instance database (also in a RAC environment)
# Default: yes
BUILD_RAC_DATABASE=yes
#BUILD_SI_DATABASE=yes
#
# Allows for database and listener to be started automatically at next
# system boot. This option is only applicable in Single Instance mode.
# In Single Instance/HA or RAC mode, the Clusterware starts up all
# resources (listener, ASM, databases).
# Default: yes
CLONE_SI_DATABASE_AUTOSTART=yes
#
# Comma separated list of name value pairs for database initialization parameters
# Use with care, no validation takes place.
# For example: "sort_area_size=99999,control_file_record_keep_time=99"
# Default: none
#DBCA_INITORA_PARAMETERS=""
#
# Should a Policy Managed database be created taking into account the
# options below. If set to 'no' an Admin Managed database is created.
# Default: no
DBCA_DATABASE_POLICY=no
#
# Create Server Pools (Policy Managed database).
# Default: yes
CLONE_CREATE_SERVERPOOLS=yes
#
# Recreate Server Pools; if already exist (Policy Managed database).
# Default: no
CLONE_RECREATE_SERVERPOOLS=no
#
# List of server pools to create (Policy Managed database).
# Syntax is poolname:category:min:max
# All except name can be omitted. Category can be Hub or Leaf (12c only).
# Default: mypool
CLONE_SERVERPOOLS="mypool"
#
# List of Server Pools to be used by the created database (Policy Managed database).
# The server pools listed in DBCA_SERVERPOOLS must appear in CLONE_SERVERPOOLS
# (and CLONE_CREATE_SERVERPOOLS set to yes), OR must be manually pre-created for
# the create database to succeed.
# Default: mypool
DBCA_SERVERPOOLS="mypool"
#
# Database character set.
# Default: WE8MSWIN1252 (previous default was AL32UTF8)
# DATABASE_CHARACTERSET="WE8MSWIN1252"
#
# Use this DBCA template name, file must exist under $DBHOME/assistants/dbca/templates
# Default: "General_Purpose.dbc"
DBCA_TEMPLATE_NAME="General_Purpose.dbc"
#
# Should the database include the sample schema
# Default: no
DBCA_SAMPLE_SCHEMA=no
#
# Certain patches applied to the Oracle home require execution of some SQL post
# database creation for the fix to be applied completely. These files are located
# under patches/postsql subdirectory. It is possible to run them serially (adds
# to overall build time), or in the background which is the default.
# Note that when running in background these scripts may run a little longer after
# the RAC Cluster + Database are finished building, however that should not cause
# any issues. If overall build time is not a concern change this to NO and have
# the scripts run as part of the actual build in serial.
# Default: yes
DBCA_POST_SQL_BG=yes
#
# An optional user custom SQL may be executed post database creation, default name of
# script is user_custom_postsql.sql, it is located under patches/postsql subdirectory.
# Default: user_custom_postsql.sql
DBCA_POST_SQL_CUSTOM=user_custom_postsql.sql
#
# The Database Name
# Default: ORCL
DBNAME="ORCL"
#
# The Instance name, may be different than database name. Limited in length of
# 1 to 8 for a RAC DB & 1 to 12 for Single Instance DB of alphanumeric characters.
# Ignored for Policy Managed DB.
# Default: ORCL
SIDNAME="ORCL"
#
# Configure EM DB Console
# Default: no
CONFIGURE_DBCONSOLE=no
#
# Enable HA (high availability) for EM DB Console by starting up
# a dbconsole instance on each node of the cluster, so that if one
# is down, others can service the requests, default: No
# Default: no
DBCONSOLE_HA=no
#
# DB Console port number. If left at the default, a free port will be assigned at
# runtime, otherwise the port should be unused on all network adapters.
# Default: 1158
#DBCONTROL_HTTP_PORT=1158
#
# SCAN (Single Client Access Name) port number
# Default: 1521
SCANPORT=1521
#
# Local Listener port number
# Default: 1521
LISTENERPORT=1521
#
# Allows color coding of log messages, errors (red), warning (yellow),
# info (green). By default no colors are used.
# Default: NO
CLONE_LOGWITH_COLORS=no
#
# END OF FILE
#
[root@ovmm utils]#
4.3.5 开始安装
上述信息配置完成后,可执行如下命令进行安装:
[root@ovmm deploycluster]# ./deploycluster.py -u admin -p Oracle123 -H localhost -M RAC-1,RAC-2 -P utils/params-my.ini -N utils/netconfig-my.ini
监控安装过程
在目标节点上进行安装过程的监控
4.3.6 监控日志信息
在日志信息中,此处贴出全部的日志以供观察研究安装的详细过程。
4.3.7 验证安装结果
5 模板的clone
在模板clone之前,我们应事先创建好需要制作成模板的服务器、数据库服务器、中间件服务器等模板源。
5.1 模板clone
此处以已安装的数据库VM虚拟服务器DB-Template-11gR2-OL6u5 为例进行,详细步骤如下:
然后点击OK即可完成模板制作。
5.1.1 模板Clone注意事项:
1、 在模板clone中一定要将光驱消除或者调整引导顺序改为disk最先引导,否则通过模板clone的服务器仍会提示安装
2、 Clone模板前需要将样本的网络信息注销,防止ip冲突。
推荐本站淘宝优惠价购买喜欢的宝贝:
本文链接:https://hqyman.cn/post/6296.html 非本站原创文章欢迎转载,原创文章需保留本站地址!
休息一下~~