I'm running SQL Server 6.5, with the latest Service pack on NT 4.0 with SP4. A couple of days ago I started getting a warning message in my event log right after a scheduled task ran. The scheduled task calls three stored procedures (data dumps for master, msdb, and a production database) it also runs a consistency check on each before dumping them. I can interactively run each stored procedure and get NO error messages. Prior to getting the warning message (which occurs after the scheduled task runs) I get normal informational entries that the dumps occurred without a problem and the files are written to the proper area. The error is in the Task Engine category and are Warning type messages. Here is the text:
SQL Server Scheduled Task: 27, 'Proddata Dump' -- Status: Failed
-- Task Invoked on: 5/20/99 5:00:00 AM -- Message: Database 'ASSMTCTRPROD' (116647 pages) dumped to file <1> on device 'D:DATDUMPASSMTCTRPROD_990520_DATA.DMP'. (Message 4035)
Any help that anyone could give would be greatly appreciated. Thanks
OS- Windows server 2012Â SQL - Windows SQL 2012 R2 Sharepoint 2010 SP2 Â Â
SQL has DB restored from earlier server. DB is quiet large in size because used with sharepoint.
Following steps have been followed on this restored DB -Â
Maintenance Plan Rebuild-Reorganize the indexes Update Statistics
After above steps, query on Sharepoint table found performant. But after some delay/idol time(overnight) on server. Query takes much more(20X) time to execute. On running execution plans observed that some warnings are seen on columns which are primary keys.Â
Columns with no statistics 'AllDocs.tp_DocID'
When Update statistics is executed again in SQL management studio above issue is again seen resolved, but came again after some delay.
Is there any SQL logs where can I find activities performed during overnight with SQL which make this issue to happen? This issue was not there on Win2k8 environment.
The default Engine Thread property of a data flow task is set to 5, is this the best setting? what if I would like to run complex data flow tasks on multi-processor machines, should I increase the engine thread? If so, then what is the recommended Engine Thread number for running complex data flow tasks in a multi processor system?
Even if i am running simple data flow tasks on a multi processor machine, should I change the engine thread?
On our particular database server, we run the Rebuild Index Task (Using classic Maintenance Plan Designer) every night. Running the  script below, I saw that about 77 tables had an avg_fragmentation_in_percentage between 80% and 99% !!
SELECT OBJECT_NAME(ind.OBJECT_ID) AS TableName, ind.name AS IndexName, indexstats.index_type_desc AS IndexType, indexstats.avg_fragmentation_in_percent FROM sys.dm_db_index_physical_stats(DB_ID(), NULL, NULL, NULL, NULL) indexstats INNER JOIN sys.indexes ind ON ind.object_id = indexstats.object_id AND ind.index_id = indexstats.index_id WHERE indexstats.avg_fragmentation_in_percent > 30--You can specify the percent as you want ORDER BY indexstats.avg_fragmentation_in_percent DESC
I dont understand why these tables are highly fragmented after a daily index rebuild! Unless the users are doing heavy inserts/updates/deletes during the day.
OK. I give up and need help. Hopefully it's something minor ...
I have a dataflow which returns email addresses to a recordset.
I pass this recordset into a ForEachLoop configuring the enumerator as (Foreach ADO Enumerator). I also map the email address as a variable with index 0.
I then have a Execute SQL task which receives this email address as a varchar variable (parameter 0) which I then use in my SQL command to limit the rows returned. I have commented out the where clause and returned all rows regardless of email address to try to troubleshoot this problem. In either event, I then use a resultset to store the query result of type object and result name 0.
I then pass this resultset into a script variable to start parsing the sql rows returned as type object. ( I assume this is the correct way to do this from other prior posts ...).
The script appears to throw an exception at the following line. I assume it's because I'm either not passing in the values properly or the query doesn't return anything. However, I am certain the query works as it executes just fine at the command prompt.
My intent is to email the query results to each email address with the following type of data by passing the parsed data from the script to a send mail task. Email works fine and sends out messages but the content is empty. I pass the parsed data as string values to the messagesource and define the messagesourcetype as a variable in the mail task.
part number leadtime
x 5
y 9
....
Does anyone have any idea what I might be doing wrong?
I am getting the following warning message in the event viewer. I have only one schedule in sql. there is no other schedules in task scheduler or database management.
Event Type:Warning Event Source:SQLSERVERAGENT Event Category:Job Engine Event ID:208 Description: SQL Server Scheduled Job 'Transaction Log Backup Job for DB Maintenance Plan 'DB Maintenance Plan1'' (0xD25A4CDDD1C59E48B4D28F61F0A411B8) - Status: Failed - Invoked on: 2004-08-16 02:30:00 - Message: The job failed. The Job was invoked by Schedule 2 (Schedule 1). The last step to run was step 1 (Step 1).
Hey guys....I have a quick question. I am having a problem with a certain app which connects to my SQL 2000 SP3a Database. It only occurs when I have a ^ in the password, both using SQL and NT authentication.
I have checked on MSDN and they do not list that character as one to avoid. I did find in the books online documentation that the ^ performs a bitwise exclusive OR operation....has anyone tested using ^ as part of a password to authenticate to a SQL db?
ALTER TABLE [konnik].[PORTAL_CONFIG] ADD CONSTRAINT [DF_PORTAL_CONFIG_AUTOPM_ON] DEFAULT (1) FOR [AUTOPM_ON]
and i received the following warning
Warning: The table 'PORTAL_CONFIG' has been created but its maximum row size (12068) exceeds the maximum number of bytes per row (8060). INSERT or UPDATE of a row in this table will fail if the resulting row length exceeds 8060 bytes.
Bear in mind that this table come from Access through the Dts. Could anyone translate this for me? How can i overcome this???
i have this error and fail to resolve it :- Warning: Null value is eliminated by an aggregate or other SET operation. what does that means? my code :-
SELECT @BookQtyIN = ISNULL(SUM(x.pickqty), 0), @PickQtyIN = CASE WHEN ISNULL(SUM(x.pickqty), 0) >0 THEN ISNULL(SUM(d.PickQty), 0) - ISNULL(SUM(x.pickqty), 0) ELSE ISNULL(SUM(d.PickQty), 0)END,
FROMtblItemDetail d
LEFT JOIN (SELECT p.PickQty AS PickQty, j.PickComplete AS PickComplete, p.ItemStorageID AS ItemStorageID FROM tblItempicked p LEFT JOIn tblJobDocuments j ON j.Docrefid=p.Docrefid AND j.JobID=p.JobID WHERE j.PickComplete='N' )AS x ON x.ItemStorageID=d.ItemStorageID
WHERE d.ItemID = @ItemID AND d.Ownership = @ClientID AND d.CustomLotNo = @CustomLotNo
If you have a package that uses an environment variable for an indirect configuration and the environment variable is not present when the package runs you get the following warning:
The configuration environment variable was not found. The environment variable was: "Seer.ConnectionManager.CUECommonReference". This occurs when a package specifies an environment variable for a configuration setting but it cannot be found. Check the configurations collection in the package and verify that the specified environment variable is available and valid.
IMO, if the package is expecting to find something and cannot find it then it should cause an error, not a warning. I'm happy for people to disagree tho :)
I'm having a problem with one of my packages due to a truncation warning that I can't get rid of. It's not the end of the world, because the package still works. It's just extremely frustrating.
The problem arises in a derived column item in a data flow task. There is a postcode field in the data flow which has space for 20 characters. I create a derived column from this which simply removes any spaces:
Derived Column Name: Postcode
Derived Column: Replace 'Postcode'
Expression: REPLACE(" ",Postcode,"")
Data Type: string [DT_STR]
Length: 20
Code Page: 1252 (ANSI - Latin I)
However when I use this expression, or anything else which uses the replace function, I end up with the warning message:
Warning 1 Validation warning. Create Staging Tables: Derived Column [20555]: The result string for expression "REPLACE(" ",Postcode,"")" may be truncated if it exceeds the maximum length of 4000 characters. The expression could have a result value that exceeds the maximum size of a DT_WSTR.
I have tried everything I can think of to get rid of the warning. Is there some way I can use the replace function, but not have the system convinced that I'm about to go over the maximum size limit?
Can some one please help. Lately I am getting an error: LAZY WRITER WARNING: LRU List is Empty
Than the system locks! can anyone help. This is driving us crazy. There is no reference in the KB. We are trying to insert records into a table by a SP.
I'm getting a strange sql server agent warning in my server's event log (not in the sql server log). Its occurring after each agent job runs and says:
"Unable to determine if the owner (sa) of job Trans Backup has server access (reason: Unable to connect to server - check SQL Server and SQL Agent errorlogs)."
I have checked the errorlogs - nothing. My job completes OK, but I get this warning.
I'm running SQL Server 2000 SP2. I'm running in a cluster. The problem has only occurred since a reboot yesterday.
I've recently restored one of my development databases from it's production backup. The restore completed successfully, users were able to connect to the database without any problems. However, when I checked the error log I found the following message:
Bypassing recovery for database 'Test' because it is marked IN LOAD.
I could not find anything on the web or in BOL. Any ideas?
Under event viewer->application log i am getting the below error continously. Pls help
Regards,Event Type: Warning Event Source: SQLServerAgent$abc Event Category: Printers Event ID: 208 Date: 2/9/2004 Time: 9:23:03 PM User: N/A Computer: abc1 Description: The description for Event ID ( 208 ) in Source ( SQLServerAgent$abc ) cannot be found. The local computer may not have the necessary registry information or message DLL files to display messages from a remote computer. The following information is part of the event: SQL Check, 0xAF04DFE7917DEA489F33D8A38E8A0793, Failed, 2/9/2004 9:23:00 PM, The job failed. The Job was invoked by Schedule 31 (6.x schedule). The last step to run was step 1 (Step 1)..
I use VB 6.0 with ADO 2.6 to access an SQL-server. I know that the table I wish to create generates this warning.
But from today ( without any changes to the SQL-server ) , this warning is reported as an error to the client instead of continuing. So my error trap routine traps this as an error now.
I have a workaround that I would to avoid and that's reinstalling the SQL-server ( Only the MSSQL ).
Does anybody know a better solution.
PS : Changing the table structure is not possible in this case.
Microsoft Internet Information Services (IIS) is either not installed or is disabled. IIS is required by some SQL Server features. Without IIS, some SQL Server features will not be available for installation. To install all SQL Server features, install IIS from Add or Remove Programs in Control Panel or enable the IIS service through the Control Panel if it is already installed, and then run SQL Server Setup again. For a list of features that depend on IIS, see Features Supported by Editions of SQL Server in Books Online.
how do i get hold of Internet Information Services I installed it on vista premium.
While intalling SQL 2005 on an x64 machine, I got an error message that warned that some functions of SQL would not install unless IIS was installed (and enabled) on the machine. It did not say what would not install, but I gather it is Reporting Services. I don't recall getting these warnings in previous (32 bit) installations of SQL 2005.
On the other hand, when I go into the BI Intelligence Studio on a (32 bit) server that does not have IIS, I seem to be able to create a report anyway (Not really my area of expertise ). Does that mean that Report Services is working without IIS?
Hi,I have a few create table statements which create temporary tablesthat has row definitions which exceeds the 8060 size limit, this causesa warning message being generated (Message 1708).Is there a way I can disable/avoid the warning message from beingprinted ?Thanks,joe
Awhile back there was some discussion about ways of sending multiplerows of data to SQL Server. The ADO .UpdateBatch method was one ofthe possibilities mentioned, and Erland said he thought that probablyonce the data get to SQL Server they are probably actually insertedone row at a time. I just want to say that based on an experience Ihad a couple of days ago I am now sure that is the case. I was using..UpdateBatch for something quick and dirty I had to do just once andwas too lazy to use XML for. An error was generated during the..UpdateBatch process because of invalid data in one row. However, alarge number of the rows in the batch WERE written to the database.Until now I haven't used .UpdateBatch in a production environment onlybecause this implies granting INSERT or UPDATE privileges to the user.But based on this experience I would NEVER use .UpdateBatch in aproduction environment, even for a user like a batch job where thepermissions would not be an issue.FWIW.
I have an SSIS package which uses a third party RMS driver. It is not licensed and so it throws a warning messagebox when the package is run and every time the connection is used. Is there a way to suppress these messages so that the package can be scheduled and run without user interaction?
I don€™t know if this is the case but I need some help. We have itanium server with Microsoft SQL Server 2005 - 9.00.3054.00 (Intel IA-64) Mar 23 2007 18:42:19 Copyright (c) 1988-2005 Microsoft Corporation Standard Edition (64-bit) on Windows NT 5.2 (Build 3790: Service Pack 2).
We has had lot of problem with that the sqlserver.exe process hangs 100% cpu. It€™s totally non responsive I€™m not even able to contact the server with DAC. The general solution has been just kill the process. Not nice.
I am able to "CREATE ASSEMBLY" for the stdole.dll without issue, however, I receive the warning:
Warning: Assembly "stdole" was built using version v1.0.3705 of the .NET Framework. SQL Server currently uses version v2.0.50727. (Microsoft SQL Server, Error: 6215)
I am unable to find a PIA for stdole that is .NET 2.0. Has anyone resolved this warning? Can I create a 2.0 PIA myself without causing other problems?
when using lookup i am geting the following warning.our OLEDB connection is Oracle.how to resolve this will this have any performance impact.
[Lookup [14342]] Warning: Cannot retrieve the column code page info from the OLE DB provider. If the component supports the "DefaultCodePage" property, the code page from that property will be used. Change the value of the property if the current string code page values are incorrect. If the component does not support the property, the code page from the component's locale ID will be used.
I'm running dtexec with the /WarnAsErrors flag, to be sure to detect any configuration warning (configuration errors are actually throwed as warning, which is not sufficient) or any other important warning.
I only get one warning (0x80047076, optimisation warning). Is there a way to disable this specific warning ?
I have insured that I have the x86 version downloaded of both the ver. 2.0 .net foundation. I have insured that I meet the min. requirements of all software requirements. For the Std. edition my hardware reports I exceed the minimum requirements. For the Adv. edition my hardware meets the min. mem. requirements.
I proceed with the Adv. installation, I find a warning during the preamable or the installation support tools regarding insufficient hardware. I then back out of installation, I remove the /program files/ms sql folder and verify registry is cleared.
I proceed to install the std. edition, I find the same warning as above. I suspect I have an error in the hardware profile reporting tool. I then search for a ms analysis tool equivalent to the one used by SQL 2k5 express to no avail.
As this installation is a requirement for classes I find myself a little frustrated; however, expert assistance would be most welcome. Note: In either case of the above installations I can proceed with the installation; however, the client fails. This issue does not appear resolveable with normal user access tools, suggestions for a Remedy in the next few hours will be most welcome. Class starts tomorrow.
Document contains one or more extremely long lines of text. These lines will cause the editor to respond slowly when you open the file. Do you still want to open the file? on every file containing a Script Component? I've got a project with about 21 packages, ~ 3/4 of them having multiple Script Components and every single one of them makes me click through the warning dialouge to open the file. That means simply opening the project takes about three minutes and almost complete attention through the entire process. Builds offer the same headache since the creation of a Deployment utility inexplicably requires every file in the project be open. As far as I can tell, the warning is a result of Script Components being included in the .dtsx file as a pre-compiled binary. I didn't specifically ask for the binary to be included in the file, so why would the standard behaviour raise a warning dialouge by default? Please tell me I'm doing something wrong.
I am using Reporting Services 2005 on SQL Express with Advanced Services...
I am dipping into the SOAP API and trying to add DataSources and Reports dynamically, as is how I would like to be able to deploy them...
My problem is this:
I have a shared datasource and a report that were deployed to the server via the 'Deploy' method in VSTS...I can access that report just fine, and it obviously uses the shared datasource...
Next, I copy the .rdl file of the report I already deployed and give it another name ("NewReport.rdl"). What I want to do is add this report to the server (it's an exact copy of the functioning report with a different report name). This report should use the same shared datasource as the other report.
Here is my code:
ReportingService2005 rs = new ReportingService2005(); rs.Credentials = System.Net.CredentialCache.DefaultCredentials;
Byte [] reportDefinition = /* read in the rdl file and create byte array out of it (didn't want to paste this code, cuz you don't care...trust me, it's correct */
foreach (ReportingTest.localhost.Warning warning in warnings) { Console.WriteLine(warning.Message); }
The report is added to the server just fine, but I receive this 1 warning:
Code: "rsDataSourceReferenceNotPublished" Message: "The data set €˜AdventureWorksTestDataSource€™ refers to the shared data source €˜AdventureWorksTestDataSource€™, which is not published on the report server."
And then when I try to view the report, I get this exception:
"The report server cannot process the report. The data source connection information has been deleted. (rsInvalidDataSourceReference)"
I know that the datasource is there, because the other report still works just fine...and I am adding an exact COPY of that report's RDL to the server, so it SHOULD work, IMO...
Any word on this?...I haven't been able to find much on this error...they say that an 'rsInvalidDataSourceReference' is usually caused when the datasource doesn't exist on the server when the report is deployed...or the datasource is deleted after the report has been deployed, but this is not true in my case...