• Facebook
  • RSS Feed
  • Instagram
  • LinkedIn
  • Twitter
Jul 012008
 

I was setting up EV 2007 Service Pack 2 on a Windows 2003 2-node cluster for the last couple of days (so I could work on a repro that I have to do).   One interesting thing that I found was that when I was trying to configure the failover node, it wasn’t giving me a list of all the cluster resource group containing my EV resources.

What I mean is that when I run the Enterprise Vault Configuration Wizard I see this :

If I select the middle option to add a new to an existing cluster I see this :

I should also see my other resource group called EVGroup.

I read through the notes on the wizard screen:

Does my group contain all those resources? Yes.

Is it online on the node in the cluster? Yes.

Why then is EVGroup not listed?  A little snippet of information in the Installing and Configuring Guide led me to find out what the problem was :

Configuring a failover node

Perform this procedure on the nodes that are to act as failover nodes.

To configure a failover node

1 On the node’s Windows Start menu, click All Programs > Enterprise Vault > Enterprise Vault Configuration. The first page of the Enterprise Vault Configuration wizard appears.

2 Click Configure the node as a failover node for an existing clustered server, and then click Next.

3 The wizard prompts you for the name of the resource group for which you want to add the node as a failover node. Select any resource group that is configured to fail over to this node. The resource group must be online on one of the nodes that you have configured as an Enterprise Vault primary node, and its  resources must all have the failover node as a possible owner.


Select the name of the resource group, and then click Next.

4 On the next wizard page, enter the password for the Vault Service account, and then click Next.

5 The next wizard page lists the actions the wizard will take if you proceed. To continue click Next, then click and then click OK to confirm the actions taken.

6 The final wizard page displays a list of the actions the wizard has performed,and the results. Click Finish to exit the wizard.

When I double checked the EV resources didn’t have this node listed as a possible owner :

Now one way to fix this up would be to edit each of the resources and add the second node as a possible owner.  Another way is to take the resources offline on the active node, fail the group over to the passive node (Node B in my case), then fail the resources back to the active node  (Node A in my case).  If you then bring the resources back online on NodeA, the configuration wizard can go through on the passive node !

Oct 052007
 

I must say it’s a LOT simpler than it used to be when I worked at Microsoft.  Okay there are still some hoops that you have to jump through, but overall, it’s not that difficult to do, and to get a “good” working cluster up in 2-3 hours or work, from scratch!  Below is a list of the steps that I took when installing a 2 node Active / Active (yes I know that’s bad!) cluster for Windows 2003 SP 2, and Exchange 2003 SP 2.

Creating the Base VMs

Create a custom virtual machine for Windows 2003 Enterprise, and try to allocate at least 768 Mb of RAM

Ensure that the disk that is created is IDE (Not SCSI)

Ensure that you have two network cards (both mine were configured to be host only)

You will need the .flp file from here, in order to install the VMWare SCSI adapter during Windows installation :

http://www.vmware.com/download/ws/drivers_tools.html

Installing Windows is as normal except that you should press F6 lots of times when the installation starts, so that you can specify the .flp file downloaded from the above link.  In order for this to work you add the virtual floppy drive to the server, but do not connect it until you get to the blue screen where you are asked about additional drivers.  After the first reboot, don’t be worry if you get an error about not being able to find NTLDR…  just disconnect the floppy drive, and restart!

The rest of the Windows installation is as normal, just carry on accepting all the defaults, and setting the language options.

Cluster Admin Account

I normally create a cluster administrator account, and then either make the account a member of the local administrators group on each node in the cluster, or more normally I just add the cluster administrator account to the domain admins group for the domain.

Configuring Windows

Firstly you should format the disks as Basic NTFS and assign drive letters.  I always like assigning Q to the Quorum disk.

Secondly you should ensure that the network cards are renamed (one should be MainLan, and the other HeartBeat).  Ensure that the HeartBeat network card is “out of the way” I used 10.0.0.1/255.255.255.0 to make sure it didn’t communicate with my ordinary VM’s

You then need to join the node to the domain, and preferably login using the Cluter Administrator account (I normally create one called clusteradmin)

Configuring Node 1

You run Cluster Administrator, and opt to create a new cluster.  You will be asked for username and password for the cluster administrator account, IP address of the cluster, and the network name of the cluster.  Afterwards you will be left with :

A cluster group called : Cluster Group

It contains the resources :

Cluster IP Address

Cluster Name

Disk Q:

Configuring Node 2

This time when you run Cluster Administrator, you opt to join an existing cluster.  All you need then is the name of the cluster, and the cluster administrator password.

Doing a failover

From Cluster Administrator you literally pick the group you want to failover, and then right click, and choose Move Group.  It can take 30-60 seconds for the resources to fully move over to the other node, on my environment the “slow” resource appears to be the Network Name resource in each group – I’ve not looked in to why that might be.

Applying Windows 2003 SP 2

I do this on the inactive node, and then restart it.

Installing a DTC

In the cluster group, you should add a cluster resource of the type Distributed Transaction Co-ordinator, otherwise you are not able to install Exchange.  This is described here :

http://technet.microsoft.com/en-us/library/be6c39e1-7432-4a5e-b638-57a7826f13f1.aspx

Installing Exchange

Before installing Exchange I made sure that the cluster administrator was delegated an Exchange Full Administrator role.  I just think that will save any problems in the future (I’m not sure it is necessary).

When installing Exchange on each node, the first process is to just lay down the binaries, in other words run Setup.exe, and follow the installation through.  This doesn’t create ANYTHING, just all the registrations, and file copying necessary.

You do that on each node, before going on to the next step.

Configuring Exchange

There isn’t really a lot that you have to do for this.  I always make sure that my cluster group has a nice name, so I can recognise it from the general cluster group.  Next you add in to that group a disk resource, an IP address and a network name.  The IP and network name are going to end up being the names and IP addresses that Outlook users and Exchange will use, so ensure it meets some sort of naming standings.

Once those resources are present, and online, you create a System Attendant resource and make sure it is dependent on the disk and network name resources.

You are then ready to online the resource, which might take a minute or two the first time you try it.

Installing the Exchange Server Pack 2

Installing the Exchange Service pack is good as well.  When you try to run the update.exe, and you get to the component selection screen, you can’t continue if the current node owns any Exchange Virtual Servers.  You get a nice warning saying that you must perform a failover of those groups (it doesn’t tell you which groups they are though).

image

After the installation of the binaries there is the obligatory restart required.

Now know doubt this is in the readme, or other instructions, but I was a bit purplexed after installing SP 2 that the node wouldn’t “hold” any of the Exchange resources.  So I paused the active node (to stop fail back), and then brought the whole virtual server group offline, before moving it to the node which I had just applied the service pack on.  I then brought the IP and Disk resource online, followed by the Network Name.  I then tried the SA resource, and got this message :

image

After a small amount of panicking, I checked the Application event log, and saw this :

image

So nothing is/was corrupted (as I first thought).  A simple right click in Cluster Administrator, and you can chose the option to Upgrade the Virtual Server.  Around 15 seconds later you get a confirmation pop-up saying it was successful.

You have to do the upgrade on each Exchange Virtual server on the cluster.  Once you’ve down the binary upgrade on one node, you can then do these steps to upgrade the virtual servers and move them to that node, followed by an upgrade of the binaries on the other nodes in the cluster.