I am constantly getting this error message in the Application log after installing SQL 2005 last night followed by SP1 (say 5 times a minutes). See below:
Q1: Any way to control which db loads when, and perhaps stop this error from happening? Again- when all is said and done, the report server functions, and I am able to access just fine.
Q2: What kind of recovery is being referred to? Given the message type- I assume it's nothing of concern- probably more to do with a system that doesn't run all the time.
Okay- so I get these two event messages each time I boot.
Below are two error message I see daily with each system boot- the services mentioned do "actually" function, and are running- as not 10 seconds after these messages are logged, the successfully loaded messages appear. It's almost as if report services tries to connect before the databases are up and running.
Basically the sequence (today) was:
1. 05:49:01 Failure Audit (MSSQL$SQLEXPRESS) 2. 05:49:01 Error (Report Server Windows Service (SQLEXPRESS) 3. 05:49:02 Recovery is complete. (EventID 3408 MSSQL$SQLEXPRESS) 4. 05:49:06 Starting up database 'ReportServer' 5. 05:49:09 Server Resumed execution (MSSQL$SQLEXPRESS)
I'm assuming based on the events that Report Server Service attempts to connect to the ReportServer database before it's actually started. Doesn't really explain the failure audit-
This is almost immediately followed by this (which I consider a no-brainer, considering the above):
2. Error:
Event Type: Error Event Source: Report Server Windows Service (SQLEXPRESS) Event Category: Management Event ID: 107 Date: 8/9/2007 Time: 05:49:01 User: N/A Computer: <MACHINENAME> Description: Report Server Windows Service (SQLEXPRESS) cannot connect to the report server database.
I was wondering if anyone knew of a way to disable the following, SQLISPackage start/finish events sent to windows event logs everytime a SSIS package is executed and completed.
Something strange has happened with Reporting Services on the server. When the ReportingServicesService windows service is runnig, it loads CPU (up to 100%) and allocates more and more virtual memory (up to 1,5Gb). There aren't subscriptions for the Report Server and nobody permanently executes reports on it. How can I fix that memory leak?
Log during memory leak: ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing ConnectionType to '0' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing IsSchedulingService to 'True' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing IsNotificationService to 'True' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing IsEventService to 'True' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing PollingInterval to '10' second(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing WindowsServiceUseFileShareStorage to 'False' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing MemoryLimit to '60' percent as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing RecycleTime to '720' minute(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing MaximumMemoryLimit to '80' percent as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing MaxAppDomainUnloadTime to '30' minute(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing MaxQueueThreads to '0' thread(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing IsWebServiceEnabled to 'True' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing MaxActiveReqForOneUser to '20' requests(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing MaxScheduleWait to '5' second(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing DatabaseQueryTimeout to '120' second(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing ProcessRecycleOptions to '0' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing RunningRequestsScavengerCycle to '60' second(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing RunningRequestsDbCycle to '60' second(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing RunningRequestsAge to '30' second(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing CleanupCycleMinutes to '10' minute(s) as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing DailyCleanupMinuteOfDay to default value of '120' minutes since midnight because it was not specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing WatsonFlags to '1064' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing WatsonDumpOnExceptions to 'Microsoft.ReportingServices.Diagnostics.Utilities.InternalCatalogException,Microsoft.ReportingServices.Modeling.InternalModelingException' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing WatsonDumpExcludeIfContainsExceptions to 'System.Data.SqlClient.SqlException,System.Threading.ThreadAbortException' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing SecureConnectionLevel to '0' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing DisplayErrorLink to 'True' as specified in Configuration file. ReportingServicesService!library!4!4/12/2007-15:39:48:: i INFO: Initializing WebServiceUseFileShareStorage to 'False' as specified in Configuration file. ReportingServicesService!servicecontroller!9!4/12/2007-15:39:49:: Total Physical memory: 3220504576 ReportingServicesService!servicecontroller!4!4/12/2007-15:39:49:: i INFO: RPC Server started. Endpoint name ='ReportingServices$MSSQL.3' ReportingServicesService!library!a!4/12/2007-15:39:49:: i INFO: Catalog SQL Server Edition = Enterprise ReportingServicesService!resourceutilities!a!4/12/2007-15:39:49:: i INFO: Reporting Services starting SKU: Enterprise ReportingServicesService!resourceutilities!a!4/12/2007-15:39:49:: i INFO: Evaluation copy: 0 days left ReportingServicesService!crypto!a!4/12/2007-15:42:46:: i INFO: Initializing crypto as user: NT AUTHORITYSYSTEM ReportingServicesService!crypto!a!4/12/2007-15:42:46:: i INFO: Exporting public key ReportingServicesService!crypto!a!4/12/2007-15:42:47:: i INFO: Performing sku validation ReportingServicesService!crypto!a!4/12/2007-15:42:52:: i INFO: Importing existing encryption key ReportingServicesService!servicecontroller!1b!4/12/2007-15:45:46:: i INFO: RPC Server stopped ReportingServicesService!servicecontroller!1b!4/12/2007-15:46:06:: Can't unload domain, trying again ReportingServicesService!servicecontroller!1b!4/12/2007-15:46:23:: Can't unload domain, trying again ReportingServicesService!servicecontroller!1b!4/12/2007-15:46:39:: Can't unload domain, trying again ReportingServicesService!servicecontroller!1b!4/12/2007-15:46:56:: Can't unload domain, trying again ReportingServicesService!servicecontroller!1b!4/12/2007-15:47:13:: Can't unload domain, trying again ReportingServicesService!servicecontroller!1b!4/12/2007-15:47:29:: Can't unload domain, trying again ReportingServicesService!servicecontroller!1b!4/12/2007-15:47:46:: Can't unload domain, trying again
when I restart the computer always come a massage telling me that the SQLDUMPER library failed initialization. Your installation is either corrupt or has been tampered with. Please uninstall then re-run setup to correct this problem.
We want to implement a cold DRP strategy here, Is there any info where:
1. To transfer syslogins to DR Server (is BCP in/out sufficient). 2. SQL to extract DTS/Agent jobs from msdb. 3. I am aware of how to fix broken logins.
I have a package which reads data from set of tables, goes thru a script component for some data changes and written to a flat file. To speed up the process, I have created 6 threads of this same sequence, but instead of writting to different files, I am joining the outputs of all 6 scripts components using 'UNION ALL' and writting it to a single file. I have to execute this for different input years and so calling this data flow within a 'Foreach loop' container for multiple years. When I ran this package for single year, it ran without any error. But when I ran it for multiple years (in debug mode), after processing 3 years, the package just stopped running. It does not abort, but just created a SQLDUMPER Error log file with the following contents.
Does anyone know if this SQL is part of MS Office or is it XP... I haven't figured that out. It says reinstall the program but doesn't say what program. I know I never intentionally installed an SQL program. At this time I get the same message as everyone seems to be getting - SQL Dumper Library failed... reinstall .... The problem for me started when I added Office 2007 and added the business contacts program. Any ideas are appreciated!
If I wanted to search for Jobs as a particular status (e.g. 0130) and wanted to keep the jobs at this status until it has reached 0500, 0125, or 0900 in it's subsequent status log entry, how can I write the SQL for it to achieve it?
I have the following SQL which searches for the Jobs at 0130, but don't know how to develop it further to search on the requirement above.
------ SQL ------- SELECT job.job_number, (SELECT MAX(jsl.job_log_number) FROM job_status_log jsl WHERE job.job_number = jsl.job_number AND jsl.status_code = '0130') as Last_Early_Warning_Status_Entry
[code].....
In the job_status_log table above, there is a job_log_number field which increments by 1 when there is a new status log entry.
I just added a post but I don't see it so here I go agian, sorry if it shows up later and there ends up being two of them. I started getting the error message - SQL Dumper Library failed ... Please reinstall... it doesn't say what to reinstal. Does anyone know if this is part of Office or XP? My problem started when I added Office 2007 and the additional Business contacts program from the web site. I did get SQL 2005 and installed it but it says it isn't working. When I click to repair it, it wants the disk, again I don't know what disk or program it wants, although it will not fix it from the XP disk. Any help would be appreciated!
select dssid,cpt,count(*) from enc_vis_cpt group by dssid,cpt will give me the cpt's and their frequency for each dssid.
dssid cpt count SPINAL CORD INJURY 9934120 AMB SURGERY EVAL BY NON-MD 622703 PSYCHOSOCIAL REHAB - GROUP 993414 SPINAL CORD INJURY 983419 AMB SURGERY EVAL BY NON-MD 6327031 PSYCHOSOCIAL REHAB - GROUP 9734114 SPINAL CORD INJURY 9934280 AMB SURGERY EVAL BY NON-MD 6227353 PSYCHOSOCIAL REHAB - GROUP 9934524
How do I limit the output to just display the 10 most frequent cpt's for each dssid. Thank you...
Hello all I'm not sure if this is anything to be concerned about, but I'd appreciate some input. I've created an alert that invokes a log backup scheduled job to start if one of the production database log becomes over 80% full. I've noticed that the log gets backed up every couple of minutes- sometimes even more frequently- while a series of scheduled jobs are in the process of execution. Otherwise, it runs and backs up the log every hour as scheduled. I've also noticed that some of the jobs are taking longer to complete (they've been running for about 6 months now). Each job truncates the table that it populates with data, so I'm not sure what the cause of the delay is. It doesn't look like there is any fragmentation. Am I missing anything? Thanks
I have used Base SAS for analysis for a while and it was really great.. everything is easy just with a simple command.. I am sure it's not the same in SQL Server but I need some help on how to start with the following:
I have a field called call_country and another field called call_minute. Each call will be saved with the destination country and the total number of minutes..
and I want to run a query to see what are the TOP frequent destinations in this format:
There must be a way to do this simply. We're running SQL Server 2000. I'm looking for some generic SQL statement that I can apply.
If I have a table with a person column and a location column and multiple records for the same person / locatioin combination, how do I select the person with the location they most frequently visited? Say George visits Mexico 5 times, and the Bahamas twice and costa rica once. I would have 8 records in my table for George. The data looks something like this:
I have a reviews table where all reviews are submitted. On the main page I want to display the 10 most reviewed products. I have a Product_ID column in this table which identifys the product. How can i write a query which will select the product_ID of records which have the most frequent product_ID's?
I came up with something like this: "Select Top 10 Product_ID, COUNT(*) AS Occurances FROM reviews GROUP BY Product_ID ORDER BY occurances DESC"
But it does not work.?? It says "Declaration expected" as error
We have 7 databases mirrored on two servers which are mirroring partners. 3 of the 7 databases are live on server1 and mirrored on server2; and the remaining 4 databases are live on server2 and mirrored on server1. The data is exposed through .NET Widows Application.
The configurations of the servers are as follows.
System: Microsoft Windows Server 2003 R2
Standard x64 Edition
Service Pack 2
Computer: Intel(R) Xeon (R) CPU
5130 @ 2.00 GHz
2.00 GHz, 32.0 GB of RAM
SQL Version: Microsoft SQL Server 2005 - 9.00.3175.00 (X64) Jun 14 2007 11:45:39
Copyright (c) 1988-2005 Microsoft Corporation Enterprise Evaluation Edition (64-bit)
on Windows NT 5.2 (Build 3790: Service Pack 2)
Problem:
The databases for no apparent reason keep randomly failing over to one server quite frequently. At least twice a day. There is no pattern associated for me to make out as to why this is happening.
My Questions:
1. Is it a good practice to divide the databases on each server, the way it is now? Or should all the databases be kept on one server and mirrored on other all the time?
2. From the above mentioned scenario, do you find the reason for database to 'failing over' so frequently? Could the Win Application which is used to expose the data be responsible for the failovers?
3. What steps can be taken to check for the reason which is causing the databases to failover? Alternatively and most importantly, how can I this problem of 'Databases Failing over randomly' be solved?
What function(s) can be used to find the mode of data? I have a column that is populated with codes and I'd like to summarize the data by the code that occurs the most frequently. Any help is appreciated!!
How can I make a statement that will return the 10 most frequently occuring values in a column?
I have no idea if that is even possible, if you have an idea on how I could do that I would really appreciate it.
Im trying to make a page that would show some statistics on a table I have.
Im also trying to make something that would show the count of the number of records inserted in the last 24 hours, week, month, year etc. The table has a column called "DateInserted" as SmallDate, right now i can use a Where DateInserted > '20080310' to get the count, but its not dynamic, is there anyone to merge all these results into one row with each column being a diffrent time period?
I know this a lot of questions, but I would really appreciate any pointers.
My SQL Server 2005 SP4 on Windows 2008 R2 is flooded with the below errors:-
Date 10/25/2011 10:55:46 AM Log SQL Server (Current - 10/25/2011 10:55:00 AM) Source spid Message Event Tracing for Windows failed to send an event. Send failures with the same error code may not be reported in the future. Error ID: 0, Event class ID: 54, Cause: (null).
Is there a way I can trace it how it is coming? When I check input buffer for these ids, it looks like it is tracing everything. All the general application DMLs are coming in these spids.
I have been testing with the WMI Event Watcher Task, so that I can identify a change to a file. The WQL is thus:
SELECT * FROM __InstanceModificationEvent within 30 WHERE targetinstance isa 'CIM_DataFile' AND targetinstance.name = 'C:\Program Files\Microsoft SQL Server\MSSQL.1\MSSQL\Backup\AdventureWorks.bak'
This polls every 30 secs and in the SSIS Event (ActionAtEvent in the WMI Task is set to fire the SSIS Event) I have a simple script task that runs a message box).
My understanding is that the event polls every 30 s and if there is a change on the AdventureWorks.bak file then the event is triggered and the script task will run producing the message. However, when I run the package the message is occurring every 30s, meaning the event is continually firing even though there has been NO change to the AdventureWorks.bak file.
Am I correct in my understanding of how this should work and if so why is the event firing when it should not ?
In my case I have to log the errors raised by any task in a package to either windows event log, text file or SQL server. Also I need to send an email notifications to a group of people telling them about the error.
Now can I use SSIS package logging for logging the errors into the required destinations. I mean right clicking on the package and selecting Logging, then adding the required log providers and enabling the events for logging into those. I think I have to upfront select the log providers to log the error, I will not have the liberty to log the error to the destination, the name of which is passed as a variable to the package. This is okay with me though.
Now what will a custom log provider help me to do in this case. Also can I somehow configure my package to call the send mail task everytime an error is raised.
Also, one more option can be developing a package that only does the error handling. It will take in the paramters or the error codes and descriptions, the destination to write to and a flag to send mail or not for that particular type of error.
we have a handful of developers and each of us are responsible for laying out and creating our own database backends. This often leads to inconsistencies in table and column structures.One obvious situation that comes up often is whether or not the other developers are building in history into their primary tables, using history/archive tables or (usually in cases of helper tables) no historical data at all.
My thought on how to alleviate this a little was to suggest that we all build a IS_DELETED computed column into our tables so that someone else trying to work with their data doesn't have to play the guessing game. In most cases, this column would just be running date comparisons on an Expiration Date and either checking to see if it's in the future (usually 12/31/9999) or NULL.
I have read that computed columns can be a performance hit if used/returned unnecessarily but is that also the case on fields where their main use would be filtering? It just seems that the calculation that the computed column is doing would be necessary for the WHERE anywhere so it seems like a wash ... and worth the benefit of not having to decipher someone else's work.
I am working on a text mining application wherein I need to detect unusual/anomalous sentences in text. Certain sentences, that I know occur very frequently, are given a likelihood of 0.2 by PredictCaseLikelihood. Other sentences that are just as frequent get a much higher likelihood (>0.9). I am using the NORMALIZED option. The only significant difference between these sentences is their length. The one with the lower likelihood has only 2 words in it, whereas the one with the higher likelihood has more than 10 words. The problem is that the shorter sentences end up being interpreted as anomalous, when in fact they are'nt. Any suggestions?
Server 2003 SE SP1 5.2.3790 Sql Server 2000, SP 4, 8.00.2187 (latest hotfix rollup) We fixed one issue, but it brought up another. the fix we applied stopped the ServicesActive access failure, but now we have a failure on MSSEARCH. The users this is affecting do NOT have admin rights on the machine, they are SQL developers. We were having
Event Type: Failure Audit Event Source: Security Event Category: Object AccessEvent ID: 560 Date: 5/23/2007 Time: 6:27:15 AM User: domainuser Computer: MACHINENAME Description: Object Open: Object Server: SC Manager Object Type: SC_MANAGER OBJECT Object Name: ServicesActive Handle ID: - Operation ID: {0,1623975729} Process ID: 840 Image File Name: C:WINDOWSsystem32services.exe Primary User Name: MACHINE$ Primary Domain: Domain Primary Logon ID: (0x0,0x3E7) Client User Name: User Client Domain: Domain Client Logon ID: (0x0,0x6097C608) Accesses: READ_CONTROL Connect to service controller Enumerate services Query service database lock state
I upgraded my Databases today with SQL SERVER 2005, initially this stopped the SQL Agents from working on port 1433, I changed the port to 1434 and now it works fine!!
Howerver I am gettin lots of error messages in my event viewer as follows.