Logging To Second Disk -- What Happens When It Crashes?
Jul 20, 2005
Currently we have SQL Server running on a single RAID5 array.
The data and the logs are all written to this array. We don't
have huge volume of activity, but it is growing. Most activity
is from our website that interacts a lot with the database.
We have read that it is a good idea to separate the data files
from the log files, having each on separate disks (or in the
case of RAID, separate arrays). So, we are considering adding
a separate SCSI hard drive, or maybe a RAID 1 system, and have
sql server log to that instead.
My question is, what happens if that secondary drive or array
completely fails. Let's say it melts. Assuming the OS doesn't
crash (Windows 2003 Server), what will SQL Server 2000 do when
the drive stops working and it is trying to write its data
logging to it? Will it continue to function and write alert
events? Will it shutdown? If it shuts down, how do we alter its
configuration to tell it to use another drive for the logs?
Also, what kind of throughput would we need to be doing to see
any kind of benefit to making this change? It is easier to
justify the expense of another disk or RAID setup if it is
actually going to matter. ;-)
hello,all I am new to Sql 2000,I installed sql 2000 database in C disk,but Now I found my C disk space is smaller than before,So I want to move my databse(include data and structure) from C Disk to D Disk(its space is very large) . is it possible to do it ? if its can be done ,do I need to change my asp.net program source code (exp: chaneg my crystal report connectstring ) ? thanks in advanced!
I am developing a package on my local workstation. I have defined two logging service providers. One is for SQL Server and the other is for the Windows Event Log. I am using the Dts.Log method in a script task to write log entries.
Logging is working properly with the SQL Server provider and rows are being inserted into the sysdtslog90 table. However, the only events that are being logged in the Windows Event Log are the package start and end events which I believe SSIS is doing automatically anyway.
Is there something I need to do to enable WIndows Event Log logging other than defining a log provider and making sure it is checked active? Won't SSIS write to two different logs with one Dts.Log call? Any ideas on what might be going wrong with my approach?
If I return the Average, Minimum, and Maximum values for the counter Physical Disk: Avg. Disk Queue Length, and those values are 10, 0, 87 respectively, which value do I use to compute the Avg. Disk Queue Length for a 4 disk array(RAID 10): Average, Minimum, or Maximum? The disk(lun) is on a SAN.
Hi, I decided to use the SQL Server log provider to store logging data of all my Integration Services packages. I also created some reports about this data for operating purposes. I have a problem occurs the name of the executing package is not always written to the log,but the name of the single task which failed. But that is not very useful information for operating, because I do not see any chance to get the name of the package by the information which is logged in the sysdtslog90 table in the database which I defined for SSIS Logging.
How do I configure the package to always log the package information into the table, too?
-- Initialize Control Mechanism DECLARE@Drive TINYINT, @SQL VARCHAR(100)
SET@Drive = 97
-- Setup Staging Area DECLARE@Drives TABLE ( Drive CHAR(1), Info VARCHAR(80) )
WHILE @Drive <= 122 BEGIN SET@SQL = 'EXEC XP_CMDSHELL ''fsutil volume diskfree ' + CHAR(@Drive) + ':'''
INSERT@Drives ( Info ) EXEC(@SQL)
UPDATE@Drives SETDrive = CHAR(@Drive) WHEREDrive IS NULL
SET@Drive = @Drive + 1 END
-- Show the expected output SELECTDrive, SUM(CASE WHEN Info LIKE 'Total # of bytes : %' THEN CAST(REPLACE(SUBSTRING(Info, 32, 48), CHAR(13), '') AS BIGINT) ELSE CAST(0 AS BIGINT) END) AS TotalBytes, SUM(CASE WHEN Info LIKE 'Total # of free bytes : %' THEN CAST(REPLACE(SUBSTRING(Info, 32, 48), CHAR(13), '') AS BIGINT) ELSE CAST(0 AS BIGINT) END) AS FreeBytes, SUM(CASE WHEN Info LIKE 'Total # of avail free bytes : %' THEN CAST(REPLACE(SUBSTRING(Info, 32, 48), CHAR(13), '') AS BIGINT) ELSE CAST(0 AS BIGINT) END) AS AvailFreeBytes FROM( SELECTDrive, Info FROM@Drives WHEREInfo LIKE 'Total # of %' ) AS d GROUP BYDrive ORDER BYDrive
I am trying to setup a test cluster and am having an issue. When I try to create the resource of a physical disk it takes both the drive e: and drive q: and doesn't seperate them into two physical disks as resources. This means when I try to associate the quorum disk it links the to physcial disk resource of drive e and q. Then when I try to install SQL2k5 I get the warning about installing SQL on the quorum disk. Am I missing something? Is there a way to seperate e and q onto two physical disk resources so I can specifically associate the quorum to q and the sql to e or should I be setting the quorum disk to a majority node set? Thanks in advance.
I recently read the project real ETL design best practices whitepaper. I too, want to do custom logging as I do today, and also use SSIS logging. The paper recommended using the variable system::PackageExecutionId to tie the 2 logging methods together.
I have a three tier system using SQL server 2000, we are currently experiencing IO bottle necks on our SCSI Raid 10 array, which holds the Data and the logs in separate partitions.
So my options as I understand it are:
Get Enterprise edition
or
Get another physical raid 10 array and separate the logs and data i.e. data on one array and logs on the other array.
I would like to try the latter but I am totally unsure how much difference this will make or whether it will make any difference at all.
Does anyone know how much performance increase I will get from using two arrays as opposed to one?
Any other advice on this scenario would be greatly appreciated.
We have 2 servers when you try to change something in a transfer for sql to a text file, it crashes and kicks you out of enterprise manager. I have seen this before. MS recommends applying sp4 or MS03-031: Cumulative Security Patch for SQL Server (815495) This is on a cluster. Has anyone seen this before
When we migrated from SQL 6.5 to 7.0 our users that had DBO privileges moved fine. When we went in and added a new user and tried to give them the same priviledges as the old user, the user looks different and does not have the same priviliges. When the user tries to run the SETUSER stored procedure it tell the user that was newly setup that they do not have the rights to perform the operation. Any help with be greatly appreciated.
We are network admins at a college, with no database experience at all. We use an Altiris product which uses SQL to maintain a database. Our database is OK, but the transaction log has grown to 24 Gb and has caused a slowdown. When we try to back up the database in preparation for shrinking it to the recommended 100 Mb, the server seems to start the backup but crashes (restarts) after a few minutes. Has the situation got out of hand i.e. with files so big, is our server never going to be able to work with it? Or are we missing something obvious?
Any information on resolving the situation gratefully received, thanks in advance.
I am using the DTS wizard to import tables from an AS/400. Some of the tables in the job are quite large and Sql Server either hangs up or completely crashes. My current configuration is WIN NT, SP4.0 SQL Server 7.0 no Service pack. Any suggestions?
I've jumped over from the sister site xtremevbtalk.com to ask the question:
Has anyone ever seen an issue where when installing MSDE, during the installation the pc reboots of its own accord?
After the MDAC is installed as part of the installation the pc reboots and upon boot up, MSDE is not in the program files directory. Instead it's sitting in the C: root. This occurs for a straight MSDE install or with an Instancename install.
There does exist a Microsoft SQL Server folder in the Program Files but it is incomplete. It's missing the Binn folder. Also, there is no reference to MSDE in the Add/Remove programs tool but ther MSDE service does show up in the Services. However I cannot start these as they are pointing to the correct location.
To date, I have tried removing the instance and installation from the registry. I have turned off all AntiVirus software. I have ensured I am using an Admin user profile. I have tried a blank sa passowrd for the installation. all of this is to no avail.
I am seriously stuck with this now and I am in dire need of a resolution so if anybody has any ideas or has seen this one before, I love some help.
I can open a single connection to the database engine successfully but if I try to open another connection ('New Query' or 'query with current connection') Management Studio just goes away without any errors.
I am utilizing a remote desktop connection to the server but if I open a connection from my desktop through Management Studio I can establish multiple connections (database engine, query, etc...) with no problems. There are no errors in the sql logs, event viewer or windows log.
Has anyone seen this? Any suggestions would be appreciated.
My application uses sqldmo to retrieve the list of sql servers in thenetwork, it works fine in some machines, but it is crashing theapplication in others, no error and no exception, it shuts down theapplication. It fails in the following line:SQLDMO.NameList sqlServers = sqlApp.ListAvailableSQLServers();any ideas?thanksmusachy
I was looking to improve the performance of an ASP.NET application bycreating the an indexed view that could be used instead of some of theroot tables.What I didn't realize is that it would affect any future conenctionsto the root tables.This of course crashed the application on any type of insert, update,or delte from the root tables.Funny thing is when I removed the indexed view -- it didn't help.I tried flipping the offending options to the opposite of way theywere set on the database and the errors wouldn't go away.I ended up restoring the database from before my mess up to fix it.I was hoping that the SQL experts out there might shed some light onmy problem before I try again.
I'm using SQL 2005 Reporting services on Windows 2003 server, my development machine is Windows XP Pro. Attempting to report from a Progress 10.0B database with the DataDirect 4.2 SQL92 driver.
A simple static query executes brilliantly (i.e. SELECT Col1 FROM Tab1 WHERE Col1 = 200) When I change the named parameter to an unnamed (i.e. SELECT Col1 FROM Tab1 WHERE Col1 = ?), then execute the query I'm prompted for the parameter value, I enter a valid value, then receive the error: "Attempted to read or write protected memory. this is often an indication that other memory is corrupt." Then VS crashes. obviously this is bothersome.
An ODBC trace shows the SQL statement (SELECT Col1 FROM Tab1 WHERE Col1 = ?), note: the parameter value I entered is not in the trace. Should it be?
I have created a calculated data field for one of my data set. I use it for a table in my report. But when I try to preview it, visual studio crashes and I get the following error message.
"An error occured during local report processing. An internal error occured on the report server.see the error log for more details."
I have not installd servers on my machine(only the sql tools). So I can not see error logs either.
Every time I do a full backup of my databases, it will go through most of them, then get to the largest database I have (13 Gig) which is the Sharepoint database, and as it starts to back it up, the machine will then freeze up and need to be hard rebooted. I confirmed it was the backup by moving the backup to a different time, which is when the server crashed. Also, if I manually attempt to do a full backup on it, the machine will then crash.
I'm running Windows Server 2003, SP2, fully patched, SQL server 2005 SP2 (32-bit) fully patched. 4 Gigs memory. Have tried setting it to use 3 Gig itself, with and without AWE, but it didn't make a difference. This machine is a dedicated SQL Server machine. Microsoft has had us run perfmon and send them perfmon reports and dumps. They don't see anything that would cause our issues. Nothing shows up in the event viewer. This server has been running just fine since last June when we installed SQL 2005. The problems started occurring on March 8 of this year. I don't see any patches that correspond to that date.
Hardware is all reporting fine according to the Gateway Hardware tool.
I'm the Sharepoint Admin, we have no SQL Admin, so now this is another job of mine. So any advice, anything I should look for, would be extremely helpful.
Will the database get locked if my ASP.Net application that is calling a stored procedure in which a HOLDLOCK for table1 is issued to SQL Server, suddenly crashes just after the stored procedure is called?
I ran a DBCC SHOWCONTIG on a large database. Because the execution was taking longer than I wanted I cancelled and then killed the process running the DBCC. This action seems to have abnormally ended SQLServer and subsequently crashed the entire NT server. Have you experienced any behavior similar to this when using DBCC and/or killed processes?
I am creating an SSIS Package to import some data. The package contains a big for each loop with 3 parallel running nested for each loops. The package crashes within the big loop without returning any error information. It starts up the sqldumper and genreates an dump - but does not report any error to the UI (Visual Studio or dtexec). While I am running only one for each loop and disable the others the package executes successfully. Has anyone an idea what might be wrong ? or is the only chance to submit a bug report ?
I'm trying to use a user defined alias type in a table creation, and everything works as expected, right until the moment where I rightclick on [Indexes] and click [new index] to create a new index.
Then I get this error:
TITLE: Microsoft SQL Server Management Studio ------------------------------ Cannot show requested dialog. ------------------------------ ADDITIONAL INFORMATION: Cannot show requested dialog. (SqlMgmt) ------------------------------ Object reference not set to an instance of an object. (SqlManagerUI) ------------------------------ BUTTONS: OK ------------------------------
When I try to manually run one of my scheduled job, DR. Watson comes up and SQL Server crashes. the following error message is received in the NT event log. "Unable to read local Event log, reason the data area passed to a system call is too small". I am new to SQL and this is causing me some degree of stress!!!! Also please recommend a good book for a newbie like me.
i didnt think my sql qeury was that complicated that it would crash my webapp. all im trying to do is filter data between two tables. heres my query<cfquery name="GetResults" datasource="#datasource#">SELECT *FROM Content, Content_SitesWHERE Content.ContentID <> Content_Sites.ContentIDORDER BY Content.ContentID DESC</cfquery>equals works, but when i try not equals, it all goes haywire. any ideas?TIA
Hello,I am having a problem when using access xp as a frontend for sql server2000.I have been trying to update a number of stored procedures (Just simpleadding fields etc) which results in access crashing with event ID 1000 and1001.Does anyone have any ideas as to what could be the problem?Thanks in advance..
I have a package with numerous data flows in it. (I know - this is not recommended as a best practice.) Anyway, one of the Data Flows just started giving me some trouble. When I opened that particular Data Flow, then BIDS would crash. I tried loading the project again, and opening that Data Flow task, and it just kept crashing. I rebooted - same deal.
The Data Flow was itself very simple. There was an OLE DB Source, and OLE DB Destination, a conditional split, a derived column, a data conversion and a Lookup. Pretty routine stuff. Out of the blue really, and long after I created it - it just starting crashing BIDS when I opened it.
My solution was to copy the Data Flow control flow component and paste it back into the same package. This worked perfectly fine, and didn't see any issues after. So, apparently something in the meta data became corrupt, and caused it to crash.
Anyway, just curious if others have seen this behavior, but also wanted to post this as a solution in case someone else has the same problem.