Integeration Service (SISS) - SQL Execute Task (destination Table)??
Oct 8, 2007
I had created an Script Task to put data into the few variables then I created an SQL Execute Task to use the SQL query to collect data from a database. Now I want to put that data into a different database.
I noticed it can not be done because the SQL Execute Task is one object and the other object is Data Flow Task. (The Destination database is inside that object).
So, if it is possible then how? If not then what should be done instead?
OK. I give up and need help. Hopefully it's something minor ...
I have a dataflow which returns email addresses to a recordset.
I pass this recordset into a ForEachLoop configuring the enumerator as (Foreach ADO Enumerator). I also map the email address as a variable with index 0.
I then have a Execute SQL task which receives this email address as a varchar variable (parameter 0) which I then use in my SQL command to limit the rows returned. I have commented out the where clause and returned all rows regardless of email address to try to troubleshoot this problem. In either event, I then use a resultset to store the query result of type object and result name 0.
I then pass this resultset into a script variable to start parsing the sql rows returned as type object. ( I assume this is the correct way to do this from other prior posts ...).
The script appears to throw an exception at the following line. I assume it's because I'm either not passing in the values properly or the query doesn't return anything. However, I am certain the query works as it executes just fine at the command prompt.
My intent is to email the query results to each email address with the following type of data by passing the parsed data from the script to a send mail task. Email works fine and sends out messages but the content is empty. I pass the parsed data as string values to the messagesource and define the messagesourcetype as a variable in the mail task.
part number leadtime
x 5
y 9
....
Does anyone have any idea what I might be doing wrong?
I have a simple Control Flow setup that checks to see if a particular table exists. If the table does not exists, the table is created in an alternate path, if it does exist, the table is truncated before moving to a file import Data Flow that uses an OLE DB Destination to output the imported data.
My problem is, that I get OLE DB package errors if the table the OLE DB Destination Container references does not exist when I load the package.
How can I over come this issue? I need to be able to dynamically create the table in an earlier step, then use that table to import data into in a later step in the workflow.
Is there a switch I can use to turn off checking in the OLE DB Destination Container so that it will allow me to hook up the table creation step?
Seems like this would be a common task...
Steps:
1. Execute SQL Task to see if the required table exists 2. Use expresions to test a variable to check the results of step 1 3. If table exists, truncate the table and reload it from file in Data Flow using OLE DB Destination 4. If table does not exist, 1st create it, then follow the normal Data Flow
Hi all, I got this scope: - SSIS and SQL server are on the same machine - myPackage run perfectly on debug mode and also via DTSexec choosing from stored package directory
My problem is when try to run the REAL job. Always got an error. ----- The import process has failed: Unable to complete the phase pre-esecuzione for component SQL Server Destination "(97). Error code returned: 0xC0202071. ----
This is a generic error and don't give me any feedback...one time is on data task A another to diffrent. SSIS pkg are compose by 34 data task...they insert into SQL server destination a RAW file source. SQL destination is the same machine of SQL server services and files are store to another machine...is a share like \BEWSSIS2007-12-18
I have try to run job with different agent (sql agent) and proxy user (same of the log-in sql server machine) where I 'm on. But always have error.
Now, I'm death point, 'cause PKG run ok into Debug and DTSexec...job is failure and all logs that I write ara vague. Some data task have big numbr of row like 95000, 625000, 53000.... But I have no one buffer error...anyway I have increase the defaultNumberOfMaxRow on the proprety for "my personal security" but no way out....
Can be a problem of buffer?? A memory problem, maybe I must use anyway DBole destination??
---------------------------- this error message on windows log under history SSIS...'cause I got an handler on PKG --------------------------- Messaggio
Executed as a user: LQS1 sql. ... Ersion 9.00.3042.00 systems to 32-bit Copyright (C) Microsoft Corp 1984-2005. All rights reserved. Start: 15.52.26 Error: 2007-12-19 15:53:26.52 Code: 0xC002F210 Origin: Important Team Activities Run SQL Description: Running query "EXEC msdb.dbo.sp_send_dbmail @ profile_name = 'ePrice Alerts', @ recipients = 'myemail @ importance =' High ', @ subject =' Import site hours 19/12/2007 15.53.03 is Failure in task: SSIS: error sku task to define, @ body = 'The import process is failed: Unable to prepare activity Added mass SSIS for data entry. ", @ body_format = 'TEXT' failed with the error:" Incorrect syntax near activity. ". Possible causes: problems in the query, setting incorrect property "ResultSet" parameters that do not correct ... Implementation of the package ... not rius Step failed.
------------------------- SQLserver LOG from SSIS execution ---------------------- Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Enforcement query "EXEC msdb.dbo.sp_send_dbmail @ profile_name = 'ePrice Alerts', @ = recipients'myEMAIL', @ importance = 'High', @ subject = 'Import site hours 19/12/2007 15.52.50 is FALLITA in task: SSIS: error in task boxes Home products', @ body = 'import process failed: Unable to prepare activity Added mass SSIS for data entry. ", @ body_format = 'TEXT' failed with the error: "Incorrect syntax near activity.". Possible causes: problems in the query, setting incorrect property "ResultSet" parameters set correctly or not problems activation of the connection. Enforcement query "EXEC msdb.dbo.sp_send_dbmail @ profile_name = 'ePrice Alerts', @ = recipients' myEmail', @ importance = 'High', @ subject = 'Import site hours 19/12/2007 15.52.50 is FALLITA in task: SSIS: error in task boxes Home products', @ body = 'import process failed: Unable to prepare activity Added mass SSIS for data entry. ", @ body_format = 'TEXT' failed with the error: "Incorrect syntax near activity.". Possible causes: problems in the query, setting incorrect property "ResultSet" parameters set correctly or not problems activation of the connection. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(128). Error code returned: 0xC0202071. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(128). Error code returned: 0xC0202071. Unable to prepare activity Added mass SSIS for data entry. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(128). Error code returned: 0xC0202071. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Enforcement query "EXEC msdb.dbo.sp_send_dbmail @ profile_name = 'ePrice Alerts', @ = recipients' myEmail', @ importance = 'High', @ subject = 'Import site hours 19/12/2007 15.52.50 is FALLITA in task: SSIS: error task Payment methods tasks', @ body = 'import process failed: Unable to prepare activity Added mass SSIS for data entry. ", @ body_format = 'TEXT' failed with the error: "Incorrect syntax near activity.". Possible causes: problems in the query, setting incorrect property "ResultSet" parameters set correctly or not problems activation of the connection. Enforcement query "EXEC msdb.dbo.sp_send_dbmail @ profile_name = 'ePrice Alerts', @ = recipients' myEMAIL', @ importance = 'High', @ subject = 'Import site hours 19/12/2007 15.52.50 is FALLITA in task: SSIS: error task Payment methods tasks', @ body = 'import process failed: Unable to prepare activity Added mass SSIS for data entry. ", @ body_format = 'TEXT' failed with the error: "Incorrect syntax near activity.". Possible causes: problems in the query, setting incorrect property "ResultSet" parameters set correctly or not problems activation of the connection. Unable to prepare activity Added mass SSIS for data entry. Unable to prepare activity Added mass SSIS for data entry. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(169). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(169). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(169). Error code returned: 0xC0202071. Enforcement query "EXEC msdb.dbo.sp_send_dbmail @ profile_name = 'ePrice Alerts', @ = recipients' myEMAIL', @ importance = 'High', @ subject = 'Import site hours 19/12/2007 15.52.50 is FALLITA in task: SSIS: error task Payment methods tasks', @ body = 'import process failed: Unable to prepare activity Added mass SSIS for data entry. ", @ body_format = 'TEXT' failed with the error: "Incorrect syntax near activity.". Possible causes: problems in the query, setting incorrect property "ResultSet" parameters set correctly or not problems activation of the connection. Enforcement query "EXEC msdb.dbo.sp_send_dbmail @ profile_name = 'ePrice Alerts', @ = recipients' myEMAIL', @ importance = 'High', @ subject = 'Import site hours 19/12/2007 15.52.50 is FALLITA in task: SSIS: error task Payment methods tasks', @ body = 'import process failed: Unable to prepare activity Added mass SSIS for data entry. ", @ body_format = 'TEXT' failed with the error: "Incorrect syntax near activity.". Possible causes: problems in the query, setting incorrect property "ResultSet" parameters set correctly or not problems activation of the connection. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(212). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(212). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(3409). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(3409). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(212). Error code returned: 0xC0202071. Unable to prepare activity Added mass SSIS for data entry. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(3409). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(188). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(97). Error code returned: 0xC0202071. Unable to prepare activity Added mass SSIS for data entry. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(188). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(150). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(97). Error code returned: 0xC0202071. Unable to prepare activity Added mass SSIS for data entry. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(188). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(150). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(97). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(150). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(509). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(509). Error code returned: 0xC0202071. Unable to complete the phase pre-esecuzione for component SQL Server Destination "(509). Error code returned: 0xC0202071.
IS that possible to get teh output of a execute sql task to excel destination.I have query which will comapre the data difference between two databses. It will comapre all tables in both databses and list out the difference in data by each table. I need to run this query using SSIS and need to get the output to a excel sheet...I have used the data flow task to run this query but my query is giving some error when used with data flow task. So i have used excecute sql task and need to write teh out put to a excel sheet.
I am using SSIS integration between two database. Both databases are sql server 2008. Â using many integration but getting problem in two only only two integration giving problem, both are executing perfectly and out put also not showing any error.
but destination table not inserted/updated anything.
first issue integration is using data flow task with oledb source and destination. second one is using execute task with for-eachloop container.
Is it not possible to have table name as a parameter? For example have the SQL something like:
Delete From ? Where ID = ?
.. I get error:
[Execute SQL Task] Error: Executing the query "Delete From ? Where ID = ?" failed with the following error: "Must declare the table variable "@P1".". Possible failure reasons: Problems with the query, "ResultSet" property not set correctly, parameters not set correctly, or connection not established correctly.
I am having a Data flow task in For each loop which will gets 100 sourcetable names and 100 target table names...
am having a simpleData flow task which trasferes from OLEDBSource to OLEDBDestination. I am repeating the Dataflow task which transfers from sourcetablename extracted from for loop to a destination table var.
The problem am gettting is for the first table it is able to transfer correcly because I did mapping for those tables at design time...but for the next coming sourcetable-desttable (which r having different no of cols,datatypes) its giving Validation failed...and...needs to refresh metadata....
is there any way to refresh the metadata of Data flow task (I set the property of OLEDBSource validate external meta to false then also same error is coming)
Can the DTS execute SQL task recognize a table in an Access database that is linked from another Access database?
I'm attempting to use a DTS Execute SQL task connected to Access1.mdb to make a new table in Access1.mdb that is using a Subquery from a linked table in Access2.mdb . (I know it sounds convoluded..) Its giving me an error in my FROM clause as if it can't recognize the linked table from Access2.mdb.
The reason why I'm doing this is that I'm not sure how to run an Access make table query from DTS that will overwrite the existing table in Access. I get an error that the table already exists.
Have an SSIS package running great in 2008R2. It generates several flat files based on inline database queries. The first step of the package inserts a record into a log stats table and the last step of the package updates this record with the package name, run time and execution status. Now I need to add the records counts for each flat file to the log table.Â
Is there a way I can update one field for run counts with each of the counts for each file. So the [run counts] table column would look something like:
file1: 43522 file2: 645367 file3: 7883
Is it possible to store the record counts and flat file names in variables then concat them at the end when updating this record?
Or, is a better way to just insert/update a new record for each flat file step and log the counts for that file for its own record?
In either case, how I can capture the file count and pass that to the update statement.
I had got the below error when I execute a DELETE SQL query in SSIS Execute SQL Task :
Error: 0xC002F210 at DelAFKO, Execute SQL Task: Executing the query "DELETE FROM [CQMS_SAP].[dbo].[AFKO]" failed with the following error: "The transaction log for database 'CQMS_SAP' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases". Possible failure reasons: Problems with the query, "ResultSet" property not set correctly, parameters not set correctly, or connection not established correctly.
But my disk has large as more than 6 GB space, and I query the log_reuse_wait_desc column in sys.databases which return value as "NOTHING".
So this confused me, any one has any experience on this?
I'm looking for a way to refer to a package variable within any Transact-SQL code included in either an Execute SQL or Execute T-SQL task. If this can be done, I need to know the technique to use - whether it's something similar to a parameter placeholder question mark or something else.
FYI - I've been able to successfully execute Transact-SQL statements within the Execute SQL task, so I don't think the Execute T-SQL task is even necessary for this purpose.
I'm using SSIS in Visual Studio 2012. My Execute SQL Task calls a Stored Procedure where I have a TRY-CATCH. Last week there was a problem and the CATCH was executed and logged an error to my error table, but for some reason the Execute SQL Task didn't fail. Is there a setting to make the Execute SQL Task fail when an SP encounters a failure?
I am using SQL 2005 SSIS. I am joining several large tables and then the move result into another table in the same database.
I would like know which method is faster:
Use Execute SQL Task to insert the result set to the target table
Use the Data Flow Task to insert the result set to the target table. (Use OLE DB source to execute SQL command and then use the SQL destination) Could you tell me why then other is slower?
I have a SQL Task that calls a stored procedure and returns an output parameter. The task fails with error "Value does not fall within the expected range." The Stored Procedure is defined as follows: Create Procedure [dbo].[TestOutputParms] @InParm INT , @OutParm INT OUTPUT as Set @OutParm = @InParm + 5 The task uses an OLEDB connection and has a source type of Direct Input. The SQL Statement is Exec TestOutputParms 7, ? output The parameter mapping is: Variable Name Direction Data Type Parameter Name User::OutParm Output LONG @OutParm
In the Control flow tab, I have an Execute SQL Task that outputs full Result set into a variable of an object type. Now how can I write the contents of the Full Result Set into a text file using Script Task. I also want to format the following way while I output into a file:
Column Name 1 : Column Value
Column Name 2: Column Value and so on
I tried writing the contents of the Object Variable into a file, but the file had an output of single word: System.__ComObject.
Code for Writing the Full Result Set into a Text File
Dim RSsqloutput as String = Dts.Variables("objVariable").Value.ToString
Dim strVal as String = "File completed on " & Now() & vbCrLf & "------------------------------------------------------" & vbCrLf
I have an application like fetching records from the DataBase(MS Access 2000) and results i have to use in Script Task. At present i have used the record fetching query,connection string in Script itself. I would like to use in Independently. Is there any Tools like (Control Flow Tools like Execute SQL Task) are there to fetch the result set from Acccess and can use the fetching results in Script Task....
I have a SSIS package contains an "Execute SQL Task". The SQL will raise error or succeed. However, it sounds the package won't pick up the raised error?
Or is it possible to conditional run other control flow items according the the status of SQL task execution?
For first time I'm testing this task and surprisingly, when I try "Edit Package" option:
1)The DTS host failed to load or save the package properly 2)The selected package cannot be opened 3)Error HRESULT E_FAIL has been returned from a call to a COM component
But after these messages you can see all the tasks but they haven't name!!
It seem as if RCW mechanism has failed between managed and unmanaged coded-partially.
I don't dare to follow doing more stuff, I don't know if that package is well-loaded or not from there. ?¿
is is posible to execute a stored proc (with parameters) from an OLE DB Destination ??
reason we are trying this is cos
our current setup is an OLE DB Command doing the first database update and then passing over to an OLE DB Destination that does the second update. There is error handling coming off the OLE DB Command to a Script component that passes to an OLE DB Destination.
we are having a problem getting the error reporting working from the OLE DB Command - the updates work fine - but not getting any updates of the error database when there is an error.
have got the error reporting fine on the OLE DB Destination.
if we can execute the stored proc from the OLE DB Destination, we will then do both updates via one stored procedure executed by the OLE DB Destination.
thx
m
n.b. think the updating of the error logs from the OLE DB Command used to work - but cant get it to work now ??!!!!??
With the new improvments to the web service task, we can now use variables as arguments in web service calls. I am trying to setup a call to the amazon web service ECS. I am trying to do a simple sellerlookup. I have played with the settings and gotten nowhere. I get one of two error when I try to execute. I can always use a scripting task or write my own task, but I would like to use the built in task if it is possible. Has anyone used AWS with SSIS?
[Web Service Task] Error: An error occurred with the following error message: "Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebserviceTaskException: Could not execute the Web method. The error is: Object reference not set to an instance of an object.. at Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebMethodInvokerProxy.InvokeMethod(DTSWebMethodInfo methodInfo, String serviceName, Object connection) at Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebServiceTaskUtil.Invoke(DTSWebMethodInfo methodInfo, String serviceName, Object connection, VariableDispenser taskVariableDispenser) at Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebServiceTask.executeThread()".
or
[Web Service Task] Error: An error occurred with the following error message: "Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebserviceTaskException: Could not execute the Web method. The error is: Method 'ProxyNamespace.AWSECommerceService.SellerLookup' not found.. at Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebMethodInvokerProxy.InvokeMethod(DTSWebMethodInfo methodInfo, String serviceName, Object connection) at Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebServiceTaskUtil.Invoke(DTSWebMethodInfo methodInfo, String serviceName, Object connection, VariableDispenser taskVariableDispenser) at Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebServiceTask.executeThread()".
Copy a full directory from source to destination (Done)
then for each file on the destination directory,it must process that file and insert rows on the table.
So I created a foreach loop, and created a varriable aclled CURRENTFILENAME, and assigned it into the foreachloop to index 0.
Inside the foreachloop I created a dataflow task, in the dataflow task I dragged a flat file source and an oledb destination, but I noticed that the flat file source requires flat file manager, and the flat file manager requires a unique FILE NAME. I cant put this c:copia*.txt.
I took a loot at flat file source properties and it has associated the flat file manager, but I can not assign the filename to the variable from the foreach.
Has anyone had any success sending or receiving file(s) from either Script or FTP task? I've Google and found examples and no luck for me. The main idea is to send a file from local PC/server to mainframe.
I've used this workaround, SSIS Script Task but no good.
SQL Server Feedback Workarounds 281893. SSIS FTP Task - Mainframe When you try to connect to a mainframe (os / 390) to ftp receive a file you get an error message stating that the path does not begin with a "/". Active feedback entered 6/7/2007 by EWisdahl
Entered by EWisdahl on 6/7/2007
Add a script task (as follows) to download the desired files... ' Microsoft SQL Server Integration Services Script Task ' Write scripts using Microsoft Visual Basic ' The ScriptMain class is the entry point of the Script Task. Imports System Imports System.Data Imports System.Math Imports Microsoft.SqlServer.Dts.Runtime
Public Class ScriptMain ' The execution engine calls this method when the task executes. ' To access the object model, use the Dts object. Connections, variables, events, ' and logging features are available as static members of the Dts class. ' Before returning from this method, set the value of Dts.TaskResult to indicate success or failure. ' ' To open Code and Text Editor Help, press F1. ' To open Object Browser, press Ctrl+Alt+J.
Public Sub Main() Try 'Create the connection to the ftp server Dim cm As ConnectionManager = Dts.Connections.Add("FTP") 'Set the properties like username & password cm.Properties("ServerName").SetValue(cm, "myServer") cm.Properties("ServerUserName").SetValue(cm, "myUserName") cm.Properties("ServerPassword").SetValue(cm, "myPassword") cm.Properties("ServerPort").SetValue(cm, "21") cm.Properties("Timeout").SetValue(cm, "0") 'The 0 setting will make it not timeout cm.Properties("ChunkSize").SetValue(cm, "1000") '1000 kb cm.Properties("Retries").SetValue(cm, "1") 'create the FTP object that sends the files and pass it the connection created above.
Dim ftp As FtpClientConnection = New FtpClientConnection(cm.AcquireConnection(Nothing)) 'Connects to the ftp server ftp.Connect() ftp.SetWorkingDirectory("MyFolder.MySubFolder.MySubSubFolder")
Dim files(0) As String files(0) = "MyfileName" ftp.ReceiveFiles(files, "C: emp", True, True)
' Close the ftp connection ftp.Close()
'Set the filename you retreive for use in data flow Dts.Variables.Item("FILENAME").Value = maxname Catch ex As Exception Dts.TaskResult = Dts.Results.Failure
I have a strange problem, which I never encountered earlier, I have a data flow which has one source and multiple destinations, I am also using multicast and derived column transformations too, however when I run the package (either from cmd line or designer) the data flow task hangs when it comes to inserting data into the target tables. and it keeps waiting forever. Initialy I thought it was my data set , however if with very few records the problem persists. I have also tried removing the table lock from the destination , increasing the rows per batch to 10000 and commit size to 10000 but to no change. I am using the fast load option within my destination. When I run sp_who2 on my Management Studio I see that the status of all the processes is sleeping and the command is "awaiting command" , any help appreciated.
I have an SSIS package which calls a web service and returns a Dataset object in the form of an xml file (it does this successfully - and stores it successfully).
The Data Flow then picks up the file using an xml source task (in which use inline schema is specified) , and passes it to an SQL Server Destination task - for bulk load.
I've checked the mappings and everything seems to be ok, the package gives no warnings or errors when run.
The destination table is created on the SQL Server without problem, but at that stage the task finishes "successfully", without actually loading any table rows into the destination table.
Also in the SQL Server Destination task - when I select "preview" in the connection manager section - I can see the column definitions, but again - no rows of data.
Thanks to Simon Sabin for pointing me here - I did post on the managed msdn newsgroups in sql server programming but have had no replies as yet.
Below is the beginning of the xml Dataset object which is output as an xml file for information. If you need any more information just let me know.
Not quite sure where i'm going wrong here - thanks in advance
I have a SSIS package in which at the end i have a excel destination task which imports a table data in excel files and then send it as an attachment using email task but i am facing a issue which is mentioned below -
1. The SSIS package runs as a SQL server agent job once in a day now every time the job runs i need to creat a new excel ( or delete the older one create again ) import data in that excel using Excel destination and then send it as an attachment using email task , How can i dynamically change the excel destination to point to a template file or new file
2. WIth the email task i am able to send only one attachment , i tried seperating the file names using semi colon but it is giving error "You dont have permission to access the file or file does not exists"
I am writing a custom task to import data from delimited files into SQL tables. I use the standard Flat File Source adapter, a custom transformation to add a URN column and a filename column to the data and the standard OLEDB Destination Adapter.
Most of my test data files work fine except for ones with a lot of columns (around 350 columns). I get an error when I call the ReinitializeMetaData() method for the destination adapter.
Q1) Is there a restriction on the number of columns (or data row size) that can be imported into an OLEDB Destination Adapter?
Q2) The reason I use this adapter rather than the SQL Server Destination Adapter is that I need to set the destination table name using a variable. I don't believe I can do this with the SQL Server Destination Adapter. Is this the case?
Q3) Anyone know of a better/alternative way of acheiving the above? One way I have thought of is to create a custom destination adapter using the SQL Server Destination adapter as the base but I'm not sure whether this is a) possible and b) worth the hassle.
I want to export to a Text File (destination) from a SQL query using a DTS DataPump Task. My query has 28 columns, some how when I try to define the destination columns for my text file the Microsoft Managment Console Crashes completely.
I tough may be my query has two many columns only to find out that it has one to many. If I ramove a column form my query, any column. I get no error at all.
¿Is there a limit to a Text File destination connection?