In my case I have to log the errors raised by any task in a package to either windows event log, text file or SQL server. Also I need to send an email notifications to a group of people telling them about the error.
Now can I use SSIS package logging for logging the errors into the required destinations. I mean right clicking on the package and selecting Logging, then adding the required log providers and enabling the events for logging into those. I think I have to upfront select the log providers to log the error, I will not have the liberty to log the error to the destination, the name of which is passed as a variable to the package. This is okay with me though.
Now what will a custom log provider help me to do in this case. Also can I somehow configure my package to call the send mail task everytime an error is raised.
Also, one more option can be developing a package that only does the error handling. It will take in the paramters or the error codes and descriptions, the destination to write to and a flag to send mail or not for that particular type of error.
I get the below error only when my IDE open. It connects well when it is found closed. [SqlException (0x80131904): Cannot open user default database. Login failed.Login failed for user 'JPASPNET'.] I could solve this by giving the logged in windows user to impersonate under IIS window > WEBSITE > ASP.NET tab > EDIT CONFIG > APPLICATION tab But I wish someone could give me the proper solution. I almost tried all from giving ASPNET user as a administrator to configuring the same in Express management tool. Environment: XP pro, VWD and SQL Express
I got an approach like that: 1) Read something from DB - check the value, if true stop if false go on2) Read the second Value (another SQL Statement) - check the value etc. Now I could open the connection at 1) and if I have to go to 2) I leave the connection open and use the same connection at 2). Is it ok to do that? The other scenario would be opening a connection at 1), immediately close it after I read the value and open a new connection at 2). Thanks for the input!
Post installation of SQL Server 2014 Express edition, I am able to connect to the Database Instance.
But while opening a new query window in SSMS or opening a table getting the error:
Package 'RadLangSvc.Package, RadLangSvc, Version 12.0.0.0, Culture=Neutral, Public Token=89845dcd8080cc91' failed to load
Object reference not set to an instance of an object. (mscorlib)..Have already tried installing the componentsDACProjectSystemSetup_enu.msi, TSql LanguageService_enu.msi, DACFramework_enu.msi from path VS 2010 WCU DAC.
Just a quick question about connection management. My application willnever need more than 1 or 2 connections about at any given time. Also, I donot expect many users to be connected at any given time. For efficiency, Iwould like to keep connections alive throughout the lifetime of the objectsrequiring them, rather than opening a new connection, executing code andthen closing it again. What is the most efficient way of doing this?Should I perform the open/close or just one open when I create the objectand a close when I dispose of it?
the logs on our database keep locking up and return and file size of 0. Does any know of a way to dump the log file. I tried: Dump Transaction log to .... but it did not want to work any ideas.
I'm running a SQL Server 2005 Express installation. Now I notice that the SQL Server logs (under Management in Mgmt Studio Express) are empty - no logs are in there, neither under Current, nor under the "archived". Do I have to switch on logging explicitly?? (I'm logging on as "sa".)
Hi,I'm using SQL Server Management Studio Express and I have made a website that uses a database with stored procedures.When running pages in my webpages it cost around 2 seconds to load each page.Thanks too long!! But where is the problem???I like to see a list of all executed stored procedures with there execute time (for each page).In this way I can check if the problem is here.How can I get a logfile like this?Thanks!
What is the best way to clear the transaction logs. My backup job each night is ending because it says it is running out of disk space and I need to clear up the transaction logs. Any help is appreciated. I see many different options (trucate option, auto shrink, etc.), just need some assistance tosome more specific best approaches. Thanks
I have August22 full backup and August22 and August23 transactions log backups . I want to get the August23 and August23 full backup. So I performed a restore backup with August22 full backup and all the August22 trn backups. Then I backed up giving it August23 name and repeated the process with the August23 trn backups. But, I get an error saying that the transactions haven't been rolled foward far enough. But, I am not missing any trns! I will really be grateful if anyone would help. thanks.
Im having issues truncating my transaction logs. I have logs in excess of 40 gigs. All the info in the BOL is very vague. Any assistance would be apreciated.
Currently, my client is allowing all the developers to log in as sa and most servers have a NULL sa password. Not a good idea. I am recommending them to move to integrated security with NT groups and to move away from using the sa login altogether. Then to secure the sa login, actually give it a password (something cryptic). I also wanted to set up an alert whenever someone logs in using the sa password. How can that be accomplished? A trigger on the sysprocesses table would work but seems a little risky.
I am currently trying to export from SQL to notes via NotesSQL.I have set up the necessary ODBC data sources & DTS job. I am encountering errors when I attempt to execute the DTS package . I.e "job failed" - How do I drill down to aquire more detailed error logging on the package to find out exactly what's wrong with DTS?
I've truncated and shrunk the database and now, it says that the size of the db is 14 mb and 8mb is free. However, on the file system, the mdb and ldb files together adds up to 14 mb. How can I free up the 8 mb for other usage?
In Enterprise Manager, Management and then SQL Server Logs, when I clicked on Current or Archive#1 or 2 etc, nothing happens. This has been going on for the past 3 weeks does any one knows what is causing such problems?
When I'm in EM, there is one log that when I click on it to view it hangs my EM. All other logs (current and others) I can click and view just fine. I'm on my way out to the server to see the log in the file directory but what could cause this?
to take the transaction log back up regularly I should have truncate log on check point false. If I do so then how will I truncate the log. regards, Renu
Is there a way to disable logging ? I'm about to split a 1000000 rows table into three other tables , and would prefere not to log this action ... Does anyone know how to do it in a SQL 7.0 statement ?
I am trying to import records via bcp (about 1,500,000 records) and I keep running out of disk space. Is there any way to limit or do away with the transaction log (and still be able to import)?
Help on Backup of Logs I get the following errors when I am trying to backup log every 15 minutes. It fails only during the peak hours.Does anyone have any ideas why it happens.
Cannot allow BACKUP LOG because file 'XX_data' has been subjected to nonlogged updates and cannot be rolled forward. Perform a full database, or differential database, backup. [SQLSTATE 42000] (Error 4213) Backup or restore operation terminating abnormally. [SQLSTATE 4200
I want to have the Transaction logs truncate on checkpoint what is the best way to do this, I take a complete backup everynight of the database. Wouldnt this set a checkpoint and clear the logs? If so it doesn't do that for me.
Hi, my log has grown out of proportion after performing an indexdefrag in bulk-logged mode! Now I cant shrink the damn thing coz I have forgotton how to shrink it when the logical log files are still in use.
Can somebody enlighten me please. thankyou.
error msg: " Cannot shrink log file 2 (MyDatabase_Log) because all logical log files are in use."
Is it normal for database response times to slow down as the transaction log fills up. I have noticed that once it gets above 65% our clients response time through the web slows down? Any suggestions on how to fix this?