Easiest Way To Start And Read A Trace From Within A SSIS Package
Jun 19, 2006
I'm trying to gather information from within a SSIS package for benchmarking, reconciliation, and reporting purposes in regards to cube processing, which I'm initiating using the AS processing task.
What is the easiest way to capture this information?
The only way I've been able to come up with is to use a profiler trace. If this is really the only way, what is the easiest way to execute and read the trace from within SSIS?
Also, if a script task has to be used, does anyone have a code sample?
I have a SSIS package which copies data from Excel file to the database. As soon as the file is copied to a specific location on the file system I insert an entry in the database table, which should kick off the above mentioned SSIS package.
I tried to read on WMI Event Watcher Task to do the above thing. I am not sure that I can do something like that.
I have a Stored Procedure preparing data, which then are exported (flat files) using 2 SSIS packages. What is the best way to execute those 2 SSIS packages (sp,job, other) ?
This is a repeat listing - third time - of this problem.
Here's the deal: If I turn on logging on an SSIS package in Development Studio, when the package executes it will log all the events I choose to the sysdtslog90 table in the MSDB database - INCLUDING the PACKAGESTART and PACKAGEEND events.
When I create my own custom logging, however, those two events ARE NOT being logged, even though I explicitly state in my script I want those two logged. Everything else in the script (OnWarning, OnPreExecute, OnPostExecute, etc.) is being logged.
In my reading, it states that the PACKAGESTART and PACKAGEEND events are defaults and are always logged and cannot be excluded.
If this is the case, can someone explain why they aren't getting logged?
I've seen other people have run across the same issue...
We have an extensive set of SSIS packages that run daily to build our datawarehouse. We have run into a bit off a maintenance issue which I am hoping someone can help me with. The SSIS packages execute stored procedures. What I would like to know is, is there a quick way to determine what SSIS package is running a certain stored procedure?
My first approach was to load the packages into a table in a database using various xml functions. Basically, the table would contain the package name, and then the stored procedures that package executes. I have run into some difficulties with this though. Has anyone tried anything similar or perhaps know of a tool which could provide similar functionality?
I want to create a package which start and stop the SQL server's services... i know i can achive this via NET COMMAND.... but i coudnt find in which task (SSIS) I can place that command?..
I also came across that I can achieve this using Execute Process task but for this I have to define executable file.... actually i dont want ne thing outside from my SSIS package
I would like to have some key-value pairs set up in the form of a XML file and read the same from within a SSIS package. What I mean is something like business information needs to be configured using XML or INI or something similar. Could anyone help me with a similar sample solutions or give me some links which will lead me to the solution.
Is it possible in any ways to Configure a SSIS package in such a way that based on the User Input the package runs. For e.g if there is a table which has say 10 distinct groups. Normal SSIS package would ideally pick all the data from the source to the Destination
I want to know how to configure in Such a way that I should be able to say Group X as the input and data related to GroupX alone should be copied.
Select * from SomeTable where GroupName = @CongigVar.
Can't an SSIS package run "in the background", so to speak, without having either the cmd.exe or dtexecui windows open while executing? I'd obviously rather not have to have a window open when the thing is running right?
We are trying to import data from a .csv file which sits on shared location. This package runs fine when we run it from designer. but we are having problem when we do it at run time (accessing it through a service). Same package runs fine if that file is on same server.
Is any one gone through this issue before? i appreciate any help in resolving this issue.
I am trying to develop a SSIS package which will read the records from the flat file and insert them into a destination table. I have some validations written in script component. I have declared two Read Write variables with package level scope. when i try to assign a value to the variable in the script component and run the package, the package throws me an error "The collection of variables locked for read and write access is not available outside of PostExecute".
What should be done to over come the problem please help me on this regard
Could anyone assist me with this issue, nothing has changed on the server except trying to restore a db from another sql server to the one with the error (SQL01 to SQL02)
SQL02 Server does not startup, i get this error:
Cannot start C2 audit trace. SQL Server is shutting down. Error = 0x80070003(The system cannot find the path specified.)
Then I try to apply the -f parameter on service start up, and i get this error:
Could not create tempdb. You may not have enough disk space available. Free additional disk space by deleting other files on the tempdb drive and then restart SQL Server. Check for additional errors in the event log that may indicate why the tempdb files could not be initialized.
I have a SSIS package which reads an excel file and loads data into a table using script component(C#) as a source. The package runs without any errors when I manually run it on my machine and on the server. But the package fails when run as a SQL Server Agent job.
I tried all the possible fixes I found on the web but still can't get it to work.
I am attempting to create a new trace but I get the following error message: "failed to start a new trace".
I have been doing some digging and as I understand it, I had to find the directory Profiler uses for temporary files. So, I typed the following in the command window "SET TMP" and I received the following reply:
C:UsersRossAppDataLocalTemp
Now, according to the forum: [URL] ...
I am supposed to check that the system folder pointed to by the TMP environment variable exists and is not crammed with files.
Well, when I went to the directory C:UsersRossAppDataLocalTemp, it is indeed full of both files and directories. The size is 16.3 MB and has 133 files and 63 folders.
When I had a look at the Environment Variables window and chose TMP the value is "%USERPROFILE%AppDataLocalTemp" which according to my limited understanding is the equivalent to C:UsersRossAppDataLocalTemp.
So, what I am wondering is am I supposed to totally clear out this directory? I am not too keen on doing this because I don't want to stuff my PC up.
"SSIS 2012 Catalog doesn't have option to give read access to SSIS Catalog to view package run reports" ... Any luck allowing power developers / operators access to READ the SQL 2012 SSIS Execution Reports without granting them SSIS_Admin or Sysadmin?
According to this link posted back in 2011 (w/ Microsoft's feedback in Nov 2011: "We’re closing this issue as “Won’t Fix.” At this point the bug does not meet our bar for resolving prior to SQL Server 2012 RTM. As we approach the SQL Server 2012 release the bar for making code changes gets progressively higher." URL....Regarding Permissions to SSIS Catalog, here are the findings. We can give access in three ways:
1. READ Access – We can provide a user db_datareader access. With this the user can see the objects within the SSIS catalog database, but cannot see the reports.
2. SSIS_ADMIN – Add the user to this database role in SSISDB. With this the user can view the reports. But it also provides them privileges to modify catalog information which is not expected. We can add it using below script EXEC sp_addrolemember 'ssis_admin' , 'REDMONDPAIntelAnalyst'
3. SYSADMIN - Add the user to this server role. This will make the user an admin on the SQL server. This is not intended. Is there any method available which will have provision to give read only access to see SSIS Catalog package execution reports and not having modify Catalog access.
Im new to SSIS. Ive started reading about the xml datatype in sql 2005 and want to use that as a column type in a table im building. In a nutshell, I need a job of sorts that will do a nightly extract of specific records from a table, including grabbing the xml data in one column, parsing it to build a flat or csv file, then emailing this file to a user.
I am scheduling the package to run on the nightly basis everyday using windows scheduled task. How do i check whether the package ran successfully or it has given any error.
Can anyone please suggest me if I can log the error in some table of SQL server.
Running Package disappears without a trace. It appears to be failing silently. It does not produce an OnError or OnTaskFailed message in the sysdtslog90 table.
The events I am logging are: OnError OnExecStatusChanged OnPipelinePostEndOfRowset OnPostExecute OnPreExecute OnProgress OnQueryCancel OnTaskFailed
The package gets executed from a web application which launches it in a new thread like this:
Package pkg; ... <Sets up pkg object>
private void RunPackage() { ThreadStart threadStart = new ThreadStart(PackageExecute); Thread thread = new Thread(threadStart);
I see the PackageStart in the log and it sucessfully executes the first few tasks and then the last entry in the log is: OnPreExecute (source=SQL Delete task, datacode=0)
Any ideas how I can troubleshoot this? I checked the event viewer on the server and it showed the Package started but that is all.
I also checked the running packages node of the Integration Services server and there was no running package.
Hey, I've a few jobs which call SSIS packages. If I run the SSIS package, it runs fine but if I try to run the job which calls this package, it fails. Can someone help me troubleshoot this issue? None of my jobs that call an SSIS package work. All of them fail.
I'm baffled by a problem I noticed yesterday, I have a 3 NODE SQL 2005 x64 ENT cluster which was setup with SSIS and Notification services upon setup. I have gone through patching SQL 2005 with SP1 (initial setup) SP2 and Hotfix 3052 through the last two months and it in the NT applicaiton log that SSIS was patched and started up to 6/15. Yesterday I received an email stating SSIS was not installed by one of our developers, upon logging in a confirmed it's no longer there! I re-installed through ADD/RM Programs SQL 2005 and ran through the SP2 and HF 3052 setup.. but upon scanning the Application logs I can't find any record of SSIS or Notification services being uninstalled. The MSinstaller shows the initial install of both packages, and then the re-install today but there's no log of uninstalling it. Does anyone know where else I can look or has can explain this odd occurance?
I truncate some tables before refreshing the data and that is one of the last steps shown in the package execution progress window when watching the package run.
Both in Visual Studio and when I use the Execute package utility once I have put the package on the server?
Is there a problem with having multiple truncate statements in one execute T-SQL statement task?
I expect to create a trigger to post updated data from GoldMine hosted in MS-SQL to my migration MS-SQL database in the appropriate tables mirroring the destination PICK data tables.
Then, start an ActiveX DTS package to migrate the data via a PICK DSN to data tables in a PICK database.
Currently the dba has been able to use VB6.0 with ADO to push data into PICK. He also was able to do similar using MS-Access.
However, PICK (RainingData) is of the opinion that he must script a PICK server side Basic (RealBasic) insert script to receive the data from a VB6.0 application triggered by MS-SQL.
I think that I could skip the Basic script and go direct with ADO in DTS as he has before with VB6.0 with a user form.
Can I have the trigger start the DTS or should I just schedule it to run as often as necessary to update the PICK database?
FYI, this is a one-way data flow into PICK.
TIA
Anyone within the L.A. CA area that has experience with PICK and MS-SQL can get some well paid consulting hours. I'm just the GoldMine GMT whose been enlisted to get the job done, but would appreciate an expert with PICK to join the project.
I create a local package in DTS a few months ago. And set a schedule to run daily. Now i need to modify. How to read the content of DTS local package? i try to right click the package in job, and edit, but i didn't see the content. How can i see it and modify. Why mine can't be read? Thanks.
And there is a task (Execute SSIS package) in First package that calls the execution of second package.
I m continuously receiving an error "Failed to decrypt protected XML node "PackagePassword" with error 0x8009000B "Key not valid for use in specified state.". You may not be authorized to access this information. This error occurs when there is a cryptographic error. Verify that the correct key is available."
As we are running first package by job, job runs successfully logging above error
The protection level of second package is set to "EncryptSensitiveWithUserKey"
Newbie here.In my database I'm needing to automate some data imports. I have theimport set up as a DTS package and it works wonderfully. But I'mhaving trouble kicking it off as a stored procedure, or even from theQuery Analyzer. I used dtsrunui to get a proper connection string, butwhen I enterEXEC xp_cmdshell 'dtsrun /S "(local)" /N "MyPackage" /A"KeyNum":"19"="19687627" /W "0" /E'I get an error that says "Could not find stored procedure'xp_cmdshell'". xp_cmdshell is indeed there, under Master, ExtendedProcedures. I tried calling it dbo.xp_cmdshell, but that didn't help.I'm guessing that I need to point the command to the location of theSP, but I have no idea how to do that. Anyone willing to shed a littlelight would get my eternal gratitude. :)Thanks in Advance, maddman
I'm trying to create a package that copyes file from one folder to another. I have created a package configuration for the destination file connection manager and specified that i set the connection string with it. Now when i deploy the package into sql server it uses the package configuration file from this location : C:Program FilesMicrosoft SQL Server90DTSPackages . This is not the location of the package configuration file i told the package to use.
when i change the destination folder for the package in the "c:program files ...." the sql server agent picks up the changes, but when i specify the change in the package configuration file that i specified for the package to use, it gets ignored ???
Previously this worked allways... i dont know what i could have possibly done wrong. Except when i deploy the package it asks for the location of the package dependencies which points to "C:Program FilesMicrosoft SQL Server90DTSPackages". I been to course about ssis and there never was any discussion that package dependencies should be changed, nor ahve i encoutered anywhere in the net that this property should be changed ?
Am i wrong to assume, that when i create the package configuration for the package, that the place where i tell it to be is not in fact the place where the sql server agent integration services job looks for it ?
Update
If i delete the package configuration file from "C:Program FilesMicrosoft SQL Server90DTSPackages" it still doesnt use the package configuration file that i have specified in the package configurations when i created the package ???
Update no 2
The package looks for the corrent package configuration file when i test it in the bids, but when i deploy it into sql server then the confguration is read from the "C:Program Files"....
In short thank you Microsoft for making a product that actually works the way the users wants it to, is simple to use and is simple to debug, like i can totally read from the logs, event manager or just someplace else that i have yet to discover the reason for the package for reading the freaking config file from the wrong location... not. I have only spend like 6 hours today trying to make it work but it simply doesnt want to co-operate...
There is and SQL Agent job that starts a package (from a file system using cmd command). Usually job takes 8-10 minutes. But sometimes it get stuck for a long time (1+ hour).
DTexec process can be found with procmon, but it seems it just not doing anything (And package is not logging to file Start of the execution) After long wait it just runs a package quickly.
I've moved a package to SSIS catalog to try to get more detailed logging, but with no luck.
Job starts at 1 PM, package execution starts at 1:49 PM. Without any messages about the execution in SSISDB log.
First I've thought it might be long validation problem, but when package executes validation messages are there and they perform quick.