Why You Don't Call COM Objects Fro Script Component Task?
Jun 13, 2006hi,
I was just trying to add COM reference but I don't see how.
Let me know any info about this.
TIA
hi,
I was just trying to add COM reference but I don't see how.
Let me know any info about this.
TIA
I am trying to cleanly shutdown a dataflow task, which contains a script component, when RunningPackage.Stop() is called from the SSIS runtime.
I've been going in ever decreasing circles with no success - it looks like the cleanest way to find out whether RunningPackage.Stop() has been called is to call FireQueryCancel(). But I can't find any reference to anything useful in a dataflow task script component that gives me something that implements IDTSComponentEvents. The nearest thing seems to be Me.ComponentMetaData which gives a reference to IDTSComponentMetaData90, but this only has methods for calling FireError, FireInformation, FireProgress, FireWarning, and FireCustomEvent. But no FireQueryCancel.
Is there a way in a script component that I can find out the state of QueryCancel?
Any help would be apprecieated.
In a Data Flow, I have the necessity to use a SSIS variable of type €œObject€? inside Script Component and assign to it the content of 'n' variables of string type.
On exiting from the script the variable of type object should contain something like in the following lines:
AAAAAAAAAAAAAAAAAAAAAAAAAAAAA
BBBBBBBBBBBBBBBBBBBBBBBBBBBBB
CCCCCCCCCCCCCCCCCCCCCCCCCCCCC
DDDDDDDDDDDDDDDDDDDDDDDDDDDDD
€¦€¦€¦€¦€¦€¦€¦.
€¦€¦€¦€¦€¦€¦€¦.
On exiting from the data flow I will use the variable of type Object in a Script Task, by reading each element in a cyclic fashion.
Is there anyone who have experienced something like this? Could anyone provide any example of that?
Thanks in advance!
I am using the "Transfer SQL Server Objects Task" to copy some tables from database A to database B including data.
The tables, primary key constraints, Foreign key, data and all transfers nicely except for "DEFAULT CONSTRAINTS" on the tables.
I have failed to find any option in the "Transfer SQL Server Objects Task" task to explicitly say "copy default constraints". So I guess logically it should happen automatically but it doesn't. I hope it is not a bug :-)
Any option anyone knows will help.
Thanks.
In short, does the €œTransfer SQL Server Objects Task€? support distributed transactions?
In trying to use a €œTransfer SQL Server Objects Task€? in a container using a transaction on the container. The task is set to support the transaction. It is setup to copy table data from several tables from a non-domain server (sql server 2000) to a domain-based server (sql server 2005). I get an error stating, €œThis task can not participate in a transaction€?.
I am wondering if it means exactly what it says €“ this task in SSIS can€™t participate at all. Or does it mean that it won€™t in this scenario for some reason. I attempted a simple copy of data from mssql 2005 to mssql 2005 (same server) and the task still failed). MSDTC appears to be running properly on my machine and such (I can do a simple distributed transaction across linked server to the 2000 server in Query Analyzer (QA)). Also, MSDTC appears to be working on both servers with distributed transaction query tests in QA.
Here€™s the error info€¦
SSIS package "Development BusinessContacts and Products Migration.dtsx" starting.
Information: 0x4001100A at Copy BusinessContacts Data: Starting distributed transaction for this container.
Error: 0xC002F319 at Copy BusinessContacts database table data 1, Transfer SQL Server Objects Task: This task can not participate in a transaction.
Task failed: Copy BusinessContacts database table data 1
Information: 0x4001100C at Copy BusinessContacts database table data 1: Aborting the current distributed transaction.
Information: 0x4001100C at Copy BusinessContacts Data: Aborting the current distributed transaction.
SSIS package "Development BusinessContacts and Products Migration.dtsx" finished: Failure.
The program '[4700] Development BusinessContacts and Products Migration.dtsx: DTS' has exited with code 0 (0x0).
I have a custom component that has no editor (i.e. NoEditor=TRUE in DtsPipelineComponent attribute).
This component derives all of its metadata based on the input (hence no editor is needed). What this means is that if the input changes in any way (e.g. the datatype of a column changes) my component will fail validation. That's not a problem, the problem is my Validate() method returns DTSValidationStatus.VS_NEEDSNEWMETADATA but ReinitializeMetaData() is not being called.
I get the familiar warning from the component:
Warning 1 Validation warning. Data Flow Task: Normaliser: The component has inconsistent metadata. TestHarness.dtsx 0 0
and when I double-click on the component I get the familiar dialog box:
TITLE: Editing Component
------------------------------
The component is not in a valid state. Do you want the component to fix itself automatically?
------------------------------
BUTTONS:
&Yes
&No
Cancel
------------------------------
but when I click on 'Yes', nothing happens. RMD never gets called.
The workaround is to detach and reattach the input (which results in a call to ReinitializeMetadata() ) or set NoEditor=FALSE in DtsPipelineComponent attribute but I'd rather not have to do that because strictly speaking the user has no need to edit the component through advanced properties.
Is there another way around this problem? Is there a way to call RMD on a component that has no editor?
Thanks
Jamie
[Microsoft follow-up]
Appreciate if anyone can show me the code to call a web service from Script Component ?
I cannot use the Web Service Task. Because parameters to the webservice are from rows of data inside Data Flow Task.
Thanks !!
Just wondering if it's possible to call a package from within a script component. I'd think so, but not quite sure how to.
Thanks,
Jeff Tolman
E&M Electric
Hi Guys,
I am new to SSIS. Heree is the transformation I need to do. In database one, I have tables:
A
B
C
Where B is a detailed table of A, and C is a "derived" table of B. There is a one to one relationship between B and C, but there may be more than one record in B for each record in A.
In database two, we have the table structure:
AA
CC
And there is a one to one relationship between AA and CC. And I need to design a transformation to migrate data from database one to database two.
Table A(->AA) and B(->BB) will be easy, just one to one migration.
The mapping rule for table C(->CC) they decided was: I need to concate each record in B and C and for a record in CC. For example, suppose we have:
A1
B1(C1)
B2(C2)
A2
B3(C3)
B4(C4)
B5(C5)
Then we will have the following records in database two:
A_1
(B1+C1+B2+C2)
A2
(B3+C3 + B4+C4 + B5+C5)
I looked through all the stock data flow components, and it seems to me that none of them can perform this task, so I am thinking to design a Script transform component to do the task.
I have written the script:
in each ProcessInputRow() sub, I check record's foreign key to A, and if they are the same I concate the records, then I put them into a VB.NET collection. Once all the records are processed, in the PostExecute() function, I count the number of new rows, (in the above example 2), then I call AddRow to add the rows, by:
OutputBuffer.AddRow()
But this does not work, I got "Object reference not set to an instance of an object". It seems that in PostExecute, the OutputBuffer is not longer valid?
Help please! :)
Is there a better way of doing what I am trying to do?
Thanks!
Wenbiao
We have schedule process server, calling SSIS package via command line (see below) to physical SSIS server. Get message "insufficient for component" and package call bombs.
Facts:
1. schedule process server has Workstation tools / Clients / Connectivity for SSIS loaded
2. SSIS is 2005, SP1
What are we missing?
c:>dtexec /DTS "File SystemSalesDWgyp_dm_carrier" /SERVER BPATLQDDW /CONFIGFILE "\bpatlqddwd$SSISSalesDWgypdm.dtsconfig" /MAXCONCURRENT " -1 " /CHECKPOINTING OFF /REPORTING V
Error: 2006-11-09 10:43:34.94
Code: 0xC00470FE
Source: DTF - Xfer DW to DM DIM_CARRIER DTS.Pipeline
Description: The product level is insufficient for component "Slowly Changing Dimension" (289).
End Error
Error: 2006-11-09 10:43:34.94
Code: 0xC00470FE
Source: DTF - Xfer DW to DM DIM_CARRIER DTS.Pipeline
Description: The product level is insufficient for component "OLE DB Command" (775).
End Error
Warning: 2006-11-09 10:43:34.94
Code: 0x80019002
Source: gyp_dm_carrier
Description: The Execution method succeeded, but the number of errors raised (2) reached the maximum allowed (1); resulting in failure. This occurs when the number of errors reaches the number specified in MaximumErrorCount. Change the MaximumErrorCount or fix the errors.
End Warning
DTExec: The package execution returned DTSER_FAILURE (1).
Started: 10:43:32 AM
Finished: 10:43:34 AM
Elapsed: 2.383 seconds
I tried to access a dot net assembly (.dll) file in ssis script component using following steps.
Create new Script Task in Data Flow Task
Edit Design Script button-> Loads script Project in MS VSA
Locate Object Browser
Select Custom Component Set from the dropdown and hit browse button
Browse and place the custom component dll (This Custom component dll has to be in GAC - Global assembly cache before browsing)
Select the namespace from the Component list of the object browser and click on €œAdd to references to Selected project in the solution explorer€? button
Write Imports <namespace> in the script code to invoke class methods from the .NET custom component
The following steps worked properly with June CTP version of yukon.In september CTP version of Yukon in SSIS when i browse the dot net assembly with the same above steps i get a error stating " The file could not be browsed ".Can anybody help me in the same.
Prashant Utekar
Error code: 0xc0012024
Using "Integration Services Project" template in Business Intelligence Studio. Using platforms Visual Studio 2005 along with SQL Server 2005.
Getting the error while trying to execute package after loading it programmaticaly.
I've just one task "Transfer SQL Server Objects Task" on my Integration Services package. But when I try to execute it from VS 2005 project programmaticaly, it gives the above mentioned error.
The commands I use:
Package pkg = new Package();
pkg = a.LoadPackage(@"C:Documents and SettingsabcMy DocumentsVisual Studio 2005ProjectslSSISSSISPackage.dtsx", null, true);
DTSExecResult dResult = pkg.Execute();
The the error comes like: error: 0xc0012024 The task Transfer SQL Server Objects Task cannot run on this edition of Integration Services. It requires higher level edition.
Please help me.
Thanks in advance,
Bhupesh
Hi
I have created a SSIS package with an FTP call. I have it set to EncryptSensitiveWithPassword.
I can manually run the package from my IDE. Once I build it, create a job, it fails.
If you try to run the package directly from the Stored Packages in Integrations Services it fails with the following error:
Error: The task "FTP Call" cannot run on this edition of Integration Services. It requires a higher level edition.
This is running on our production with a full version.
Does anyone have any suggestions?ideas?
Any incite would be greatly appreciated.
PR
My goal is need to synch couple of tables in QA and Prod. I used "Copy SQL Server Objects Task" in DTS package with copy option "replace existing data". It is trying to truncate and insert new data. But BOL is says "Overwrite existing data in the destination objects with the new data from the specified source".
Let me know your thought about this.
Hello,
I am trying to copy a subset of tables, stored procedures, and views from one database to another database on a named instance of SQL Server 2005. I am attempting to use the Transfer SQL Server Objects task in Integration Services.
I am able to create both Source and Destination connections, and specify the objects I want to transfer. When I run the package, the task turns yellow while it's processing, then green when it completes.
At this point, no objects have been transferred to my destination database. No tables, no procs, no views, nothing. No error has been generated. No information is written to an SSIS log file, other than the fact that objects are being transferred, and then the package is finished.
Does this have something to do with the fact that I'm using a named instance of SQL Server 2005?
Thanks,
Ken
Hi all,
I am using SSIS Transfer SQL Server Objects Task to copy all the objects of a database to another without the data. The properties that i have set for this task is as:
DropObjectsFirst = True
IncludeExtendedProperties = True
CopyData = false
CopySchema = True
UseCollation =True
IncludeDependentObjects = True
CopyAllObject = False
CopyAllTables = True
CopyAllViews = True
CopyAllStoredProcedures = True
CopyAllUserDefinedFunctions = True
CopyAllDefaults = True
CopyAllUserDefinedDataTypes = True
CopyAllPartitionFunctions = True
CopyAllPartitionSchemes = True
CopyAllSchemas = True
CopyAllSqlAssemblies = True
CopyAllUserDefinedAggregates = True
CopyAllUserDefinedTypes = True
CopyAllXMLSchemaCollections = True
CopyIndexes = True
CopyTriggers = True
CopyFullTextIndexes = True
CopyPrimaryKeys = True
CopyForeignKeys = True
All the security options are set to false.
When I execute this task, it fails (with the errorCode=-1073548784) at a point where it tries to create a table which is inside a schema, because the script for creating the table is executed before creating the schema. So, if have a table [Person].[Employee], the script "Create Table [Person].[Employee]" gives an error
The specified schema name "Person" either does not exist or you do not have permission to use it.
It cannot be a permission related issue as am using the sa account to connect to both the source and the destination.
Hence, I conclude that the SSIS task here tries to copy the table first without creating the schema and so the excution fails.
Is this a bug in SSIS? or am missing anything? and what can be the workaround this?
Hi,
When i'm trying to transfer SQL Server Objects from a SQL Server 2000 Database to another i got the following error: [Transfer SQL Server Objects Task] Error: Table "bank" does not exist at the source.
This just appens in some spcific situations:
- When i select the tables i want to transfer, using the option "CopyAllTables" it works fine;
- When i use specific instance as source. Using the Development machine as source it works fine, when i use the machine from Pre-Production (the one i whant to use) it doesn't .
Can anyone tell me why does this appens!?
Thanks,
VÃtor Ferreira
Hello all,
A little background... I have a 25GB database (called DevDB) that my co-workers use for SQL development. The data in this database isn't important and all I really need are the SQL objects. So instead of doing a database backup, which includes the data as well, I was planning on just copying the database objects to another database called DevObjects (on the same server) and backing it up instead. This is SQL 2005 SP2.
I've created an IS package and have 3 items...
1. Check for existance and drop DevObjects database. This is to eliminate the need for dropping the objects first. Successful.
2. Create a new database DevObjects. Successful.
3. Transfer SQL Server Objects Task. Fails.
My transfer objects task is setup like this:
The connections are to the same server, source db: DevDB, destination db: DevObjects.
DropObjectsFirst: False
IncludeExtendedProperties: True
CopyData: False
CopySchema: True
UseCollation: True
IncludeDependendObjects: True
CopyAllObjects: True
CopyDatabaseUsers: True
CopyDatabaseRoles: True
CopySqlServerLogins: False
CopyObjectLevelPermissions: True
CopyIndexes: True
CopyTriggers: True
CopyFullTextIndexes: True
CopyPrimaryKeys: True
CopyForeignKeys: True
GenerateScriptsInUnicode: True
I get an error trying to create a login that's not even in the database I am trying to copy the objects from. Since this is the same server, the login is already there.
Is there a way to not copy the server logins (so that CopyAllObjects works)?
Anyway, I got past that error by changing CopyAllObjects to False and setting all the subgroups to True. Now, I get an error creating one of the FK's, it seems that the PK it references isn't created first.
Is there a way to force creating the PK's before the FK's?
If I set CopyForeignKeys to False, then it completes successfully, but I need the FK's to be transferred as well. Any ideas?
Does anyone have a better way to accomplish what I'm trying to do?
Thanks in advance.
Jarret
I have the following problem:
All Sp's and UDF's in one base database have to be compiled into five identical databases located in the same SQL Server in order to keep all those five db's stored procedures synchronized.
I made a simple SSIS package which content one "Transfer SQL Server Objects Task" component.
Hardcoded connection/Source db/Destination db and setup properties as
"Drop object first" = True
"Copy All Stored procedures" = True
"Copy All UDF" = True
Then I executed the Package right from Visual Studio environment and after the substantial thinking finally it show me a green light.
But when I verified the result no Sp's in the destination database were re-compiled/copied from the source database.
Is it kind of bug or I did something wrong ?
Any advice will be appreciated.
I have a SQL2000 database that I need to copy tables from to my SQL2005 database. The table in the 2000 database are owned by a login named tsreader. This login is also in the dbo database role. So, when I access tables I have to prefix tha table name with tsreader. Currently in a SQL2000 database I have DTS packages that pull data over from this database everynight and they work fine. However when I try to do this from SSIS, I get the error "Object does not exist". My assumption is that it is trying to access the database tables as a "select * from tablea" as opposed to "select * from tsreader.tablea".
Any thoughts on how to make this work? I know I have the login correct because it is the same one I use for the SQL 2000 packages. Also, in SSIS it allows me to chose the tables I want yet I still get "Table does not exist at source" when I try and execute it.
Any ideas?
Thanks,
fjk
What is the security requirement for running this task? I have no problem running this taks as a system administrator but all my users who are in the db_dtsltduser cannot run the same task successfully. They are DBOs in both databases involded in the task. Thank you in advance for your help.
View 2 Replies View RelatedHello all :)
First, sorry for my english but I don't practice everday
So,
I'm a new user of Microsoft SQL 2005 and I'm testing some functionalities in SSIS.
I used the Transfer Objects Task to copy one database to another.
The both db's are on the same instance.
I chose these options for the transfer
-SourceConnection : My_Instance
-SourceDatabase : My_DB
-DestinationConnection : My_Instance
-DestinationDatabase : CopyOfMy_DB
-DropObjectsFirst = True
-IncludeExtendedProperties = Fasle
-CopyData = True
-ExistingData = Replace
-CopySchema = True
-UseCollation = True
IncludeDependentObjects =False
CopyAllObjects = True
The others options are set on FASLE
When I execute the package, I got an error message, the logins are dropped. I cannot connect to the database engine, even with my admin account.
I have to restore the master database !
So my question is : why this SSIS task drop all the logins in sql server in place of just copy one database??????
What's the solution for this problem?
Thanks for your comments !
Matt
I have written an SSIS package with a Transfer SQLServer Objects task which I want to use to copy database objects from a SQL Server 2000 database to SQL Server 2005.
When I run this task, I find the following error:
[Transfer SQL Server Objects Task] Error: Execution failed with the following error: "Version80 database compatibility level is not supported
Is there a way around this aside from changing the compatibility level of my SQL Server 2005 database?
Any help is much appreciated.
https://connect.microsoft.com/SQLServer/feedback/ViewFeedback.aspx?FeedbackID=127100
I don't believe that the fix to the above issue has made its way into an SP yet. Can anyone confirm/refute this, and/or state when the fix will become publicly available &/or give a workaround?
Thanks,
Tamim.
Hi everyone,
I'm currently having some trouble with the "Transfer SQL Server Objects Task".
All I wanna do is to copy a simple database from one server (SQL2000) to another one (SQL2005).
A small excerpt from my settings:
DropObjectsFirst: true
ExistingData: Replace
All table options are set to true
.
No matter what I do the package always appends the data from the source to the destination database.
Is this a bug or am I missing something ?
Thanks in advance,
Kevin
I am struggling to copy a 2000 DB to 2005 using transfer sql server objects task.
I can get it to work, but without the foreign keys, which I also need.
When also copying the foreign keys, I get the following message:"
[Transfer SQL Server Objects Task] Error: Execution failed with the following error:
"ERROR : errorCode=0 description=There are no primary or candidate keys in the referenced
table 'SVS' that match the referencing column list in the foreign key 'FK_WRM_SVS_WRM_SVS'.
helpFile= helpContext=0 idofInterfaceWithError={8BDFE893-E9D8-4D23-9739-DA807BCDC2AC}".
Apparently, SQL tries to create a foreign key on the WRM table, while the primary key on SVS is not there yet.
Since it used to work in 2000, I am pretty sure I am doing something wrong. I can harldy believe this is by design.
(By the way, the CopyAllSchemas is set to True).
Can anyone please help?
Regards,
Pipo
Hello,
I am using the Transfer SQL Server Objects Task to copy Stored Procedures from one DB to another. I quite painfully discovered the problems with using the "DropObjectsFirst" flag in that if you set it to true and the object does not exist in the destination DB, SSIS throws a "Object does not exist" error. If you set it to false and the object exists at the destination DB, SSIS throws a "Object already exists" error. Sort of a catch 22 problem, so I decided to use a Data Flow task to build a Recordset of SP names that were common between the Source and Destination DB's and then feed that to a For Each task to drop the existing SP's from the destination DB before running the "Transfer SQL Server Objects" task.
So the problem is that I am getting 3 SP's copied over to the destination DB that do not exist on the source DB. If I delete all SP's on the destination DB and run the package it works fine the first time, when I run the package again, I get an "Object already exists" error at the destination DB for one of the SP's that don't exist on the source DB.
To explain the details of how I am building the list of SP names to be dropped, my Data Flow Task uses 2 DataReaders with a query to sys.objects to get a list of SP's from both DB's. This is then fed to the required Sort transforms and into a "Merge Join" Transform (Joint Type = Inner Join). The Merge Join Transform output is then passed to a RecordSet Destination variable.
When I query the source DB sys.objects, I can see all the SP's that were copied to the destination DB EXCEPT these 3 SP's. All the other SP's are successfully dropped on subsequent executions of the package except these 3. One of the 3 SP's (...Yardrequest...) does exist on the Source server (sort of) but when it gets copied over to the destination server, the Case of the name gets changed (only 1 letter in the middle of the name) to "...YardRequest...". I can't believe SQL is smart enough to capitalize each word in a SP name let alone what reasoning there could possibly be for doing such a thing.
Any ideas on where these 3 SP's are coming from?
My gut is telling me the DB got hosed somehow and therefore the Transfer SQL Objects Task is somehow seeing these addtional SP's that a simple query on the sys.objects view does not see. Not knowing how the "Transfer SQL Server Objects" task works internally (what queries it uses to identify what objects will be transferred from the source DB), I am not able to figure out where these SP's are located in the DB tables or views. Any pointers would be helpful. Thanks
I was wondering if there could be a race condition when using the Transfer SQL Server Objects Task to copy all tables from an online database.
I'm thinking of a scenario that while the copy task is executing, 2 different tables are updated in the online copy, with one new row each, and the new rows are linked by a key. Could it happen that when these values are inserted, one table has already been copied so the new value doesn't make it over, but the second table hasn't yet been copied so its new value does make into the destination (thereby breaking the relationship)? Or does this task take care of the details automatically?
Thanks!
Hi everyone,
I'm currently trying to copy a database from one server to another (both SQL2005) using Business Intelligence Development Studio.
I've created an SSIS package. The following parameters are defined:
DropObjectsFirst true
IncludeExtendedProperties true
CopyData true
ExistingData Replace
CopySchema false
UseCollation false
IncludeDependentObjects true
CopyPrimaryKeys true
CopyForeignKeys true
The package fails with the following error:
Violation of PRIMARY KEY constraint 'PK_tblCallStatus' Cannot insert duplicate key in object 'dbo.tblCallStatus'
I know what that error means but I don't understand why I get it.
Isn't the package supposed to completely overwrite the destination database ?
It obviously does not. When I manually delete all records from 'tblCallStatus' in the destination database it works fine. I can't remember I had to do that in a SQL2000 environment using DTS.
Hope anyone can help since this is almost driving me nuts ;-)
Thanks in advance,
Kevin
Hi :Can i call 2 procs within one task?I have sp_proc1 ? (and have declared one global variable as inputparameter)now i have another sp_proc2 which uses same input parameterbut if i write two statements like this within one task, i get anerrorexec sp_proc1 ?exec sp_proc2 ?I can solve the problem by writing them in 2 separate tasks, but wouldlike one task.Please help..thanksRashika
View 1 Replies View RelatedI am trying to call a web service using the web service task and passing it variables, but even when I set variables to be Strings/any other data type, they always get passed out as objects. The call is then rejected based on the fact that there is no web service that accepts those parameter types. If there is a workaround to this problem, that would be great, otherwise I think I'm going to have to call the web services in a script task. Would someone please give me an example of the steps necessary to call a web service from the script task?
Thank you,
James Mac William
Hi,
I m trying to use the web services task, but when i try to run it i got the following error
[Web Service Task] Error: An error occurred with the following error message: "Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebserviceTaskException: Could not execute the Web method. The error is: Object reference not set to an instance of an object.. at Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebMethodInvokerProxy.InvokeMethod(DTSWebMethodInfo methodInfo, String serviceName, Object connection) at Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebServiceTaskUtil.Invoke(DTSWebMethodInfo methodInfo, String serviceName, Object connection, VariableDispenser taskVariableDispenser) at Microsoft.SqlServer.Dts.Tasks.WebServiceTask.WebServiceTask.executeThread()".
This is probably because i never worked with this task, even so, does anyone knows what might be wrong?
I got a connection a wsdl, i can configure the service, method and variables(i use fixed values), and i configure the output to a variable with the type object(i have tryed string and int)
I even can download the wsdl file, but the error seems to be in the connection
Thanks
I'm perplexed and in need of a sanity check ...
Instance of SQL2K5-SP1 on Server2003R2. Created two databases Test1 and Test2, create single 2-column table (tblTest)in Test1 with one row of data. Create SSIS package to transfer the table and its data to database Test2 using "Transfer Objects Task". Fails with error code 1073548784.
I have tested for security issues and don't beleive these to be the cause: I have monster privelidges and can successfully use SSIS packages to execute CREATE TABLE etc.
Have also tried most combinations of Task options (eg COPY ALL OBjects etc etc)
Have tried this on three different servers: same results.
This simple test is the result of distilling down problems I'm having with SSIS manipulating some 60G within some 300 tables, so I'd really like to use the power this task promises!
My head is getting a little sore from scratching - any advice gratefully received.
Thanks