Recovery :: Windows 2012 Clustering With Quorum Disk
Jun 20, 2015
I have a windows 2012 cluster environment that consists of two SQL servers nodes with Quorum disk configured as witness.
Manual failover between nodes is working fine, however the sql instance virtual is not seeing the Quorum disk.
Moreover the Quorum disk has the same number as another cluster storage disk, is that considered a problem?
When I move the SQL instance from a node to anohter, should the Quorum Disk change ownership as well to that destination node ? if it is not changing ownership what would be the problem??
I am trying to setup clustering for SQL 2005. I initially want to setup a 2 node cluster in Active/Active Configuration.
I am trying to understand the Quorum disk in SQL 2005. As I understand the quorum disk is a shared resource. How is this resource configured? Would I need to have an iSCSI or fibrechannel connection from each of the nodes to the shared disk?
As well, does each node have a separate data drive? Or do all the nodes use only the shared storage?
I am trying to setup a test cluster and am having an issue. When I try to create the resource of a physical disk it takes both the drive e: and drive q: and doesn't seperate them into two physical disks as resources. This means when I try to associate the quorum disk it links the to physcial disk resource of drive e and q. Then when I try to install SQL2k5 I get the warning about installing SQL on the quorum disk. Am I missing something? Is there a way to seperate e and q onto two physical disk resources so I can specifically associate the quorum to q and the sql to e or should I be setting the quorum disk to a majority node set? Thanks in advance.
I'm trying to install a server cluster to implement an SQL Server 2005 cluster. No other services (I think this is important).
I've a dual SCSI channel Smart Array with 4 disks configured in a 400Gb RAID 5.
I do not need to move different resource groups from one node to the other, I need only one group with all the resources IP, Network name, MSDTC, and SQL Server..., when a node fails, all services should failover to the other node.
Is it possible to have only one physical disk (RAID 5) for Quorum disk and shared disk?
It would be the following configuration:
[Groups] Cluster Group IP Address Network Name Physical disk (used for quorum and shared storage) Distributed Transaction Coordinator SQL Server SQL Server Agent Generic Service (SQL Server Fulltext)
The other option would be having a 1 physical disk Raid 0 for Quorum (146Gb wasted) and another physical disk Raid 5 (3 disk) for Shared Storage, but this schema will have a a flaw point that if Quorum disk fails, the cluster fails....
I've the quorum disk of my cluster on win2003 full and I cannot use The cluster administrator because the Service Cluster cannot going up.
Obviously the shared disk (included Quorum) and MSDTC are not visible and I'm wondering if is possible to solve the problem without rebuild the cluster.
I want to install sql 2012 in a non-cluster windows servers, then set up Always-on between these 2 servers. Every article mentions Windows Clustering, and we dont have clustering for now.
Can I set up Always On, in a NON_CLUSTER environment?
Also where I can find instructions step-by-step installations of SQL Server 2012 and most importantly ALWAYS-ON.
I have configured two node cluster environment on vmware with windows 2008 r2 os.I want to configure sql clustering over it.but prior to that I ran failover clustering manager on node V1 to create two node cluster(node v1,node v2 and third node as a domain controller with shared storage using iscsi target).After I successfully created cluster with name SqlClus1 I shut down all my vmware.Now when I start them all and see clustering status I get error RPC server not found and Cluster server can not be connected.On my cluster manager page I see cluster events marked as critical for both nodes.and cluster staus down.
I'm just starting to work with AlwaysOn Availability and WSFC.
I have in my environment (in Azure) a DC, WSFC and to SQL instances, so I have 3 nodes in my Failover Cluster:
WSFC SQL1 SQL2
If I simulate failure by shutting down one of the SQL boxes my Availability group seamlessly fails over to the other SQL instance - which is great.
However, I'm starting to look into the workings of the Quorum, my envt has the default settings and when I shutdown both of my SQL servers I expected the Cluster itself to go offline as 2 out of the 3 votes will be negative, but the Cluster is still up - Screenshot below when SQL1 and SQL2 are shutdown:
Going through the Wizard (but not changing anything) it shows following config:
I am wondering what would be the best disk/RAID setup for a Windows server 2008 R2 OS and SQL Server 2012 database that has heavy read/write. I have the following disks I can use:
4x 15k 146GB 2x 10k 600GB
According to the server build requirements for the application, I need 100GB for the OS and 290GB for the drive containing the SQL mdf there are no stated requirements for the ldf, but would like to know if it should be allocated elsewhere?I should do RAID 10 for the 15k drives for SQL and RAID 1 for the OS on the 10k.
I tested the failover clustering for SQL 2008R2.When I stop the SQL server services manully, the failover did not fail to another resource. At the Faiolover cluster manager, SQl server(MSSQLSERVER) only show the status for offline. I think it should be move to another owner intead of just show as offline.
I have setup a WSFC that contains 3 nodes ( node 1,2 belongs to subnet A and node3 to subnet B), lets say node 1 is the primary host in the wsfc.
Each node has their own sql server instance, that i configure an Availability Group that will replicate between node1 and node 3), created an Availability Group Listener that will contain 2 vips for each subnet.
My question here and I am not sure if i missed something when setting up the Availability Group. If i failover an Availability group from sql server from node 1 subnet a to nde 3 subnet b, then the listener will be on the vip that belongs to subnet b and the primary instance will be the one in node 3, but this wont cause the WSFC to move the primary host from node 1 to node 3.
So at this point dont see what is the dependency between setting up the AG with WSFC. when fail over an AG doesn't cause the primary host in wsfc to fail over as well.
Why availability group relies on WSFC and how can i make it to depend so the primary host moves as well?
I am writing up a general purpose document regarding Disaster Recovery and Backup and Restore, etc. I have a question that I've been unable to find any definitive answer on, and I was hoping that someone might have an answer.
Here is the scenario and what I have so far, and my question:
Scenario - A simple database, with one MDF and one LDF, each on it's own disk, with no mirroring or filegroups etc.
Disaster - the LOG disk fails.
To simulate the disaster, I shut down SQL Server, and delete the LDF manually.
When I restart SQL Server, the database is marked as suspect. As expected :)
So, now the recovery begins..... I've read that in this case there is no data loss, but I find that hard to believe. Can anyone confirm?
Anyway, how to restore the database to usable status? Here is what I've come up with so far:
1) Do a backup of "the tail of the log." This works even though there is no log. I have no idea why really, but I can't proceed otherwise.
2) Detach the database.
3) Reattach the database. This auto-creates a new log file (albeit not in the location I want, which makes for more work after).
At this point the database appears to be fine. I have not seen any lost data or problems, but of course I'm working on a test database, not with live users hitting it constantly. If the database were under heavy use, I expect that there might indeed be data lost?
Is this the recommended way to recover from a disaster of this type? Or, would it be better to just go back to your last good backups and forget about "up to the minute" recovery in this case? Or is there yet another means?
We are trying to build HA and DR environment, combining AG and FCI. So we have 3 servers added to a cluster, 2 of them sharing the same storage and working as ACTIVE-PASSIVE cluster instance. The other server has a stand alone instance with its local storage, with the same drivers letters as the cluster instance.
Between the Active FCI and the Stand Alone, we set up an Availability Group. They are running fine, except by the listener. When the stand alone instance owns the Primary Replica, I am able to access the SQL Server instance by the listener, but when the Primary Replica is owned by the FCI, I can't reach the database by the listener (name or IP).
I've opened the same ports in both servers, it hasn't worked though.
I am trying to cluster two sql node server on hyper-v both virtual and my situation is below :
two hyper-v host connected to vnxe storage two hyper-v are mapped to 3 LUN ( SQL Data , SQL logs , Quorurm) connectivity between hyper-v and storage direct fiber three LUN are mapped to the virtual SQL01 and SQL02 as VHDX and configure it on virtual machine as IDE below is the picture
LUNs are presented to the guest SQL01 and SQL02 and visible through disk management
When create the fail over cluster i am not getting the disk to add them to cluster
Getting the below error when validate the storage :
No Disks suitable for cluster disks were found, For diagnostic information about disks available run the Validate config wizard’..
Also , regarding this LUN how i can make them shared on the host side ?
Does anyone have any information or know where I can get information about setting up SQL2000 and Windows2000 clustering? We want to set up 3 SQL 2000 servers on Windows2000 clustering. We cannot find any technical documentation about the 2 together. If you know of any books or sites please let me know.
is SQL Clustering possible with Windows 2003 Standard Edition? I have read, we need Enterprise Edition but it is too expensive. is there any alternate solution?
We are planning to setup SQL Express on a windows os clustered environment and trying to shoot for a active-passive configuration. We know that SQL server 2005 has this ability and will be leveraging that for mission critial production applications, however there are several other apps that we intend to use internally and would like to leverage windows clustering. Has anyone done this?
- Should we share the INSTALLSQLDATADIR? If yes, then how to specify the SQL Express installtion on cluster node#2 that, "hey..utilized this dir as the data dir".
- Or Should we not worry about it, and only the databases that we create for our apps be on the quorum/clustered resource drive?
We've got a problem with our server in that Windows has totally died - a routine upgrade appears to have gone wrong and the repair function of the Windows installer keeps getting stuck in a loop so I've given up on that front. What I was wondering is:
1. If I install a fresh copy of Windows 2003 R2 over the top of the existing install will that kill MSSQL Sever 2005 and/or Windows Sharepoint Services? 2. If so, is there a simple way to get them back? 3. If I install another copy of Windows on the D: drive is there a recommended way to move the MSSQL databases & Sharepoint to the newer install or a way of pointing the new install at the old databases?
I'm not an admin person (i'm a developer normally) so I haven't got a lot of background in this stuff so any tips, pointers or ideas are welcome. If this is covered somewhere on the MS site I'm happy to read but I'm just getting lost at the moment.
And, yes, I realise that backups are a good thing but the latest backups are not new enough for some items. The backup process has been adjusted accordingly.
Hello,I am trying to recover a SQL Server 7 database from another hard diskdrive that has a corrupted Windows 2000 Advanced Server installation. Iam not able to repair the corrupted Windows 2000 Advanced Serverinstallation but the file system is intact. I have installed a new copyof SQL Server 7 onto a new hard disk and have used the sp_attach_dbsystem stored procedure to attach the database from the old hard driveinto the new installation of SQL Server 7 on the new hard drive. Thedatabase shows up in the Enterprise Manager View and all the data andstructure are there, but when I try to run a Web site that uses ODBCconnections to the database I get the following error:ODBC Error Code = 08001 (Unable to connect to data source)[Microsoft][ODBC SQL Server Driver][Named Pipes]Specified SQL servernot found.I have double-checked that all the NTFS permissions, ODBC DSNs, and WebServer settings are correct and when I run the SQL Server ODBC drivertest at the completion of each ODBC DSN setup, it is successfull.1) Why am I getting this ODBC error?2) What steps have I missed in the recovery process?3) If I have done something wrong recovering my database, can I stillrecover it since I still have: a) The old master database file and logfile; b) The old database (non-master) file and log file; c)A backed-up(Enterprise Manager -> Right Click A Database -> All Tasks -> BackupDatabase) copy of the old database (non-master) file; ?My System Specifications:-Windows 2000 Advanced Server-SQL Server 7-IIS 5.0-Cold Fusion Server 4.5 Professional Full EditionThanks For Your Time,Nathan
hello,all I am new to Sql 2000,I installed sql 2000 database in C disk,but Now I found my C disk space is smaller than before,So I want to move my databse(include data and structure) from C Disk to D Disk(its space is very large) . is it possible to do it ? if its can be done ,do I need to change my asp.net program source code (exp: chaneg my crystal report connectstring ) ? thanks in advanced!
I have a Windows 2008 R2 Always on Cluster with 3 nodes (two in the primary site and one in the DR site).
Primary Site: -Primary Site Server1 -Primary Site Server2
DR Site 1 (to be decommed): -DR Site Server1
Our company is planning on decommissioning the DR site. But before we do this, we want to add a 4th site to the cluster. Migrate the data...and then decommission the original DR Site.
Is it possible to have this configuration:
Primary Site: -Primary Site Server1 -Primary Site Server2
DR Site 1 (to be decommed): -DR Site Server1
DR Site 2 (NEW DR Site): -DR Site Server1
IF this is possible, do I simply add the new DR site to the existing cluster (same steps as adding the first DR node to the cluster when the cluster was originally configured? or are there special steps?
We have a couple of 200GB databases that are recreated each night on a SQL2012 server connected to a disk array. The SQL disk is an auto-tiered combo of 10k and 7k drives in a RAID1 lun, and both data and log files are there.
Recently, some room has opened up on an older array that contains smaller, but many, 15k drives that I could use in a RAID1 config. Being that I'd like to split up the mdf and ldf files, which would you put onto the new (faster) disk?
EDIT: Add'l info: the only current performance issues I see in the SQL Log are FlushCache messages occuring throughout the night, when all activity happens for this DB. Things like this: "FlushCache: cleaned up 388690 bufs with 23474 writes in 409743 ms (avoided 179747 new dirty bufs) for db 47:0"
I am practicing on vmware 2 node failover cluster for doing Sql failover clustering understanding.I did proper Ad Dc setup and configured san storage through iscsi software that I downloaded from net.I did target and initiator configuration for san storage on same Domain controller vm.Name of target is Target1 and has two initialtor for node1 and node2.It has two virtual disk assigned Quorumdisk.vhd and sqldisk.vhd.ON vmnode1 I have initialzed and formatted my both disk with proper volume and label.
Node which are participating in cluster are node1 and node2.while my windows clustering went fine,while Sql server installation on node 1,i got error IP Address conflict occured.
my DC has IP address:192.168.1.10 my Node1 is 192.168.1.20 node2:192.168.1.30 Cluster IP address is 192.168.1.35
What has cause this situation.What configuration is needed.How do I troubleshoot this problem as now I can not install my sql for clustering as also gave me Invalid network name error.Is my storage configuration not proper or IP address is in correct?Do I need to put my virtual disk in Clustered shared volume?