I have a Job that runs a DTS package. This package in turns call several other packages. All individual packages run fines. When the main package is run manually it also runs fine. The jobs that calls the main package should be runnning for at least a couple of Hours maybe more and updtae several tables. It actually run nightly for only 20 minutes and doesn't update any tables. Yet it is in "successful"? Any thougts?
How do I change the color of the task icon back to green? I have and FEL with tasks in it that occessionally fail. The error is trapped to allow the container to continue processing. I would like to change the color of the icon back to green on the next successful iteration of a task but I haven't found a way to do it.
I got the following error message when the scheduled DTS run for copiying the records from one server to another server.Actually it has copied successfully,but still returned this error?!There is a similar job on a different server doing the same thing returns a success message properly.?!! Thanks for any help.
''...ding... DTSRun: Executing... DTSRun OnStart: Create Table request124 Step DTSRun OnError: Create Table request124 Step, Error = -2147217900 (80040E14) Error string: There is already an object named 'request124' in the database. Error source: Microsoft OLE DB Provider for SQL Server Help file: Help context: 0 Error Detail Records: Error: -2147217900 (80040E14); Provider Error: 2714 (A9A) Error string: There is already an object named 'request124' in the database. Error source: Microsoft OLE DB Provider for SQL Server Help file: Help context: 0 DTSRun OnFinish: Create Table request124 Step DTSRun OnStart: Copy Data from Results to request124 Step DTSRun OnProgress: Copy Data from Results to request124 Step; 10 Rows have been transformed or copied.; PercentComplete = 0; ProgressCount = 10 DTSRun OnFinish: Copy Data from Results to request124 Step DTSRun: Package execution complete. Process Exit Code 1. The step failed.''
I used Query Analyser to De-tach a SQL 7 database and that worked fine. The database became 'invisible' in Enterprise Manager and the entry was removed from the sysdatabases in Master. The mdf and ldf file remain in the file system. However, when I ran the sp to Attach the database QA said successful but it does not show in Enterprise Manager or in sysdatabases in Master. I'm going to run through the procedure again and if problems I'll check the logs but has anyone else had difficulty with this. Also with attaching the db to a SQL2000 install I believe the logins on the SQL7 install are not brought over. In my case, the client is using trusted logins with their database application so their is probably not a lot to manually recreate on the SQL 2000 install but doees anyone have nay sample code/script that can give me an idea of what is required in the area of getting logins out of the Master of SQL7 and into SQL2000.
Hi All,I have a DTS package that runs calling a few stored procedures and anactiveX module. The package does some data cleanup and load. It takesan hour to complete. I have emails sent at the beginning and end ofthe package execution along with error notifications if any of thestep fails.The issue is that, the DTS runs successfully. I ran the package fromthe server and it completes. If I schedule it, it has been running forover a few months without any issues. One fine day hell broke looseand it started to give an attitude!!When the job is invoked automatically on the scheduled time, it runsthe package. Does what its supposed to do including sending out thefinal email. but is not finishing the job. The job says its stillexecuting but the final step has completed. I have tried logging thepackage, and it says completed the last step.. But the JOB Log hasn'twritten anything to the log file.. (Job Log is where I am logging thestep where I am calling the DTS).Has anyone have any idea what this issue is? I am at a dead end andhave no clue why this is happening. I am thinking of changing thewhole process but before that I want to try posting this.Any ideas, suggestions will be much appreciated. Please feel free tolet me know if there is any additional information you need in orderto troubleshoot this issue.Thanks in advance..Aravin.
I have a standard Login ASP.NET 2.0 control on a login Page, a LoginName and LoginStatus controls on the member's page. once the user login successfully I am redirecting the user to Member.aspx page. The following is my machine configuration
Windows XP Pro Service Pack2 IIS 5.1 SQL Server 2000 visual Studio 2005 DISABLE ANONMYOUS ACCESS IN IIS ENABLE WINDOWS AUTHENTICATION
I am Using form authentication, and I am getting following error "Your login attempt was not successful. Please try again".
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
<html xmlns="http://www.w3.org/1999/xhtml" > <head runat="server"> <title>Untitled Page</title> </head> <body> <form id="form1" runat="server"> <div> You are back ........ <asp:LoginName ID="LoginName1" runat="server" /> <asp:LoginStatus ID="LoginStatus1" runat="server" LogoutAction="Redirect" LogoutPageUrl="~/Login.aspx" />
</div> </form> </body> </html>
Member.aspx.cs
using System; using System.Data; using System.Configuration; using System.Collections; using System.Web; using System.Web.Security; using System.Web.UI; using System.Web.UI.WebControls; using System.Web.UI.WebControls.WebParts; using System.Web.UI.HtmlControls;
public partial class Member : System.Web.UI.Page { protected void Page_Load(object sender, EventArgs e) { Response.Write("<br/>Page.User.Identity.Name: " + Page.User.Identity.Name); Response.Write("<br/>Page.User.Identity.IsAuthenticated: " + Page.User.Identity.IsAuthenticated); Response.Write("<br/>Page.User.Identity.AuthenticationType: " + Page.User.Identity.AuthenticationType);
Ok, here is the deal. In T-SQL, an error is handled depending on its severity. For example, a severity of 15 or less (or is it 10 or less? doesn't matter) will only raise a warning that can be caught through the message event. 16 or higher (or whatever) will cause an exception to be thrown on the .NET side, and 20 or higher causes the connection to be closed.So far so good.Now here's my issue: I have a stored procedure that does 2 queries (inside the same SP). Sometimes, the first query will succeed, while the second one will cause an error of severity 16. Again, in .NET that will throw an exception, making me unable to fetch the first resultset (which I require for logging purpose). If the error was, let say, severity 9, I could simply subscribe to the message event to get my error, yet still be able to get the first result set. But unfortunately life isn't perfect, and the error is indeed a severity 16.Anyone know of a way to be able to get the first result set, even if the second query (within the same SP) completly fails?
I'm running a pretty standard insert command on a button (myCommand.ExecuteNonQuery();) I have a few other methods that run, but I only want them to proceed if the above was inserted successfully, how would i check that?
After a successful transaction (after committing), I found that i sent wrong records of one the column (wrongly updated). How do I get back my initial stage? I mean, how do I get back all the previous results? Is there a way to retrieve my previous data which were before committing a transaction?
I've installed 2005 / RS on my testing server several times, no problem.
My production server throws an error when Reporting Services is installed. There are a few threads on this error:Microsoft SQL Server 2005 Reporting Services -- Error 29528. The setup has encountered an unexpected error while Setting reporting service and share point exclusion path. The error is: Fatal error during installation. I've Googled several similar complaints which discuss any of: Ensure IIS default web site is set to "All Unassigned" Ip[/*]Delete the RS Web services pool and retry[/*]Install RS separately[/*] The first two do nothing for the problem. While installing RS separately doesn't generate an error, it also doesn't create the RS databases.
Can the databases be created separately after installation?
I run XP Professional on a home machine. I recently installed VB 2003 with MSDE. When I go to the server explorer window, right click to add a connection, the Data Link properties dialog box pops up as it should. I select the name of home server, indicate integrated security, and select a sample database such as model or master from the list. The test connection is successful and clicking OK causes the connection to appear in the Server Explorer window. Yet when I go to open the connection, there is no data. The folders for Tables, Stored Procedures and so forth appear but they are empty. I have another 2000 database that a friend sent me. If I try to indicate in the Data Link propetries dialog box that I would like to attach this database, I get the same symptom. Test connection successful, yet when I create and open this connection, the folders are empty. I am a newbie to ADO .NET and I am not sure where to begin. Can someone help?
I keep receiving the error message below when I try to connect to SQL Management Studio
"A connection was successfully established with the server, but then an error occurred during the login process. (provider: Shared Memory Provider, error: 0 - No process is on the other end of the pipe.) (Microsoft SQL Server, Error: 233)"
I have tried all of the suggestions in Google, in the forum, etc with not success.
I have enabled the
Named Pipe protocol/ Shared Memory protocol
SQL Server browser is enables
I even tried changing the SQL Server 2005 Shared Surface Configuration with no success.
I did not have this problem until the day light saving time change. I know this may having nothing to do with it but I need some help pronto.
In the error log I have the following error
Could not connect because the maximum number of '2' user connections has already been reached. The system administrator can use sp_configure to increase the maximum value. The connection has been closed. [CLIENT: <local machine>]
Error: 17809, Severity: 20, State: 1.
Not sure what to do or if the problems are related.
hello, I'm new to store procedure in slq 2000. I want to create a sproc in my sql 2000 sevrer,I have a table named "Supplier", now I want to return my identiy field value when I use insert into sql command (insert into "Supplier") successful. So how to create this store procedure ? much thanks
I wish to have a batch file exit with a return code of 3 if it fails. In the batch file I have the single command: Exit 3 On SQL Server the Process Exit Code of a successful command is set to 0.
Yet, when I run the job SQL Server says that it is successful.
Hi all, Im not really sure if this is a PHP or and SQL problem but here goes.
Im using MSSQL and have developed a webpage that enables users to run various PACKAGES manually, however I need to display if the package has been successfully run.
Is there a system table that logs package information or is there a PHP function that I can use.
Thanks
P.s I know there are some system tables with the information for jobs but I do not want to create a job for each package.
More than 10 times, i validated the package can be executed successfully, if i use a SQL job to run this package. All data and function goes well, however the job result reported failure after i saw the job 100% completed. That 's so weird.
Anyone met the same the issue. Could anyone give me some suggestion or hints?
I am using a XML source and an oldedb destination to load an XML file into a database, the XML and corresponding XSD are provided by a different system and I dont have any control on them.
In the data flow I can see all the tables in the schema and I can map the columns, however when I run the package it validates fine, and then completes without going into the execution mode, its all green but doesnt load any data. Its weird and I cannot figure out if it is an XML or XSD problem, any idea?
I have code that has worked just fine for some time, and now all of the sudden I am having an issue. I have a simple INSERT statement built and then make the following call:
RecordID = cmd.ExecuteScalar
I have never had a problem with this before. The RecordID of the newly inserted record is returned into the RecordID Integer varibale. All of the sudden, the varibale has a value of 0 (null I assume is being returned), but yet the INSERT worked just fine. I can check the table in SQL and it is populated with no issues.
No exception is thrown of any type or anything. Does anybody know what may be happening?
I have a job which is scheduled to run once daily. The job is enabled and the schedule is also enabled. After the job runs (successfully), the schedule has its 'Enabled' flag reset (i.e. the 'Enabled' checkbox in the schedule properties has become unchecked) and so is not rescheduled.
I have other jobs configured in a similar way, but they are run and then rescheduled in the normal way without any problems.
We use AlwaysOn availability groups, which has 2 SQL nodes configured (version: 11.0.3373.0), my full and differential backups are working for all my databases, however I am unable to perform any LOG backups.
I have double checked my Availability Group settings, and the backup preferences is set to: 'Prefer Secondary'
I've tried creating a maintenance job as well as using Ola Hallengren's maintenancescript job to back them up, but nothing is written to the drive. All jobs return successful every time, and take less then 3 seconds to run. There are no events being written in the SQL error log or event log.
I recently attempted to change the password of the sa login. There were no issues while updating the password neither did I get an error when I logged in as sa. Everything appeared to connect and I was able to query a database. However, an error was logged (error 18456 severity 14 state 8) as though I did not log in successfully. Even though I can connect successfully, I don't like having these misleading errors in the SQL Server Log.
I have a maintenance job which has many job steps - checkdb, backup, rebuild index, etc...
I was asked to setup a job which will copy all backup files to a central location.
Now I wrote the script and its working good (part of a new job step)
But, I am now thinking to add some more logic in the job step:-
that is, if the previous job (server maintenance) is successfully completed with the backup job (if completed successfully), then only the copy to location job will start, if the other job step is failed, my copy bkp job will not run.
I have a Data Driven Subscription which runs every day for the previous day. If for any reason the report subscription fails, i would like to get the last time it ran successfully. I know in the subscription table we have lastRunTime, but it will show the recent one which failed but i would like to know when it ran successfully last.
Example: Saturday 5:00 AM - Ran Successfully Sunday 5:00 AM - Ran Successfully Monday 5:00 AM - Ran successfully Tuesday 5:00 AM - Failed LastRunTime column will have Tuesday 5:00 AM.
Is there a way to get Monday 5:00 AM which is the last ran successfully ?? How can i get the history of a subscription?
I have created a simple SSIS package using BIDS which does data copy from SQL Server to a excel file
I'm able to execute it using dtexec command locally connecting to the remote db server.
We have a file shared server where we are keeping the dtsx files.
Our DB server is located on another server.
We have another file shared server where Tidal agent is installed and we have created a batch file which has the dtsexec command and this batch is kept here.
Now when I create a job and run it it gives us the following outputCode: 0xC0014019 Source: tidal_jobtest Description: The connection manager "dbserver.database.db_user" will not acquire a connection because the package OfflineMode property is TRUE. When the OfflineMode is TRUE, connections cannot be acquired. End Error Error: 2007-05-04 09:09:58.06 Code: 0xC020801C Source: Data Flow Task OLE DB Source [1] Description: SSIS Error Code DTS_E_CANNOTACQUIRECONNECTIONFROMCONNECTIONMANAGER. The AcquireConnection method call to the connection manager "dbserver.database.db_user" failed with error code 0xC0014019. There may be error messages posted before this with more information on why the AcquireConnection method call failed. End Error Error: 2007-05-04 09:09:58.06 Code: 0xC004701A Source: Data Flow Task DTS.Pipeline Description: component "OLE DB Source" (1) failed the pre-execute phase and returned error code 0xC020801C. End Error
I have checked the SSIS package properties and the offline mode property is set to false but still i get the same error.
Also Can anyone let me know how to check whether we are using 32bit or 64bit drivers in connection or for dtsexec command
I am trying to restore a SQL Server 2005 database from a backup file and experiencing a hanging issue after its "finished"
I am doing this in SQL Server Management Studio, generating the following SQL for Restore:
RESTORE DATABASE [AdventureWorks2] FROM DISK = N'C:Program FilesMicrosoft SQL ServerMSSQL.2MSSQLBackupAdventureWorks.bak' WITH FILE = 1, NORECOVERY, NOUNLOAD, REPLACE, STATS = 10 GO
When I run this on the machine i originated the backup with (creating AdvWorks2) it runs fine in no time.
When I run this command on another SQL Sever 2005 instance on another host. It appears to run fine, and I see progress going up to 100% and it says "Restore Completed Successfully"
BUT, for some reason, the database in object explorer is stuck with a "(Restoring...)" label attached to its tree item and I am unable to perform any activities on that database instance. It claims, it's in the middle of a restore operation! again this is after it had reached 100% on progress and declared successfull completion.
Any ideas what could be causing this?
(Note: Both instances are SQL Server 2005 - Service Pack 2)
I have a Mantenance Plan running on my Production Server. Each job has one step. -Optimization (each week) -Integrity Checks (each week) -T-Log Backups (each hour) -Full Back up (once a day)
The Job is saying the last three are failing but niether the SQL Server Logs or MS Application Log are reporting a success or faluire. "write to windows application log on faluire" is set for all of them. The backups are occuring, as the files are in thier destination.
Our backups by default go to a network location, but I'd like to modify our maintenance plans to backup to an alternative location if the primary location isn't available. I've setup two Backup Database Tasks where the second one runs only if the first one Fails, and if the second one runs (on first one's failure) it then sends a notification to me so I know this occurred.
The Plan is running as expected, when I simulate a bad path in the first Backup Task the second one runs and the notification is sent, but the Job shows failure. I'd like to show the job as Successful when this occurs since I'm handling the issue and notification within the Plan, but I'm unable to find out how. I've set FailParentOnFailure to False on the Plan and I've changed the MaximumErrorCount to 2 with the assumption that this would work, but neither didn't.
Also I'm running into this in both SQL 2008 and SQL 2012.