We have deployed an SSIS package successfully to production. We needed to apply SP1 to fix a different issue and now have encountered a new problem. We have numerous Data Reader Sources in different Data Flow Tasks that connect to a IBM iSeries (DB2) source. Pretty simple extracts that have worked fine in the past. They pump the data into staging tables on the SQL2K5 instance running the package (64-bit).
After we applied SP1 however, all of the Data Reader tasks fail AFTER they successfully copy the records with the following error.
[iSeries Invoice Details [1]] Error: System.NullReferenceException: Object reference not set to an instance of an object. at Microsoft.SqlServer.Dts.Pipeline.DataReaderSourceAdapter.PrimeOutput(Int32 outputs, Int32[] outputIDs, PipelineBuffer[] buffers) at Microsoft.SqlServer.Dts.Pipeline.ManagedComponentHost.HostPrimeOutput(IDTSManagedComponentWrapper90 wrapper, Int32 outputs, Int32[] outputIDs, IDTSBuffer90[] buffers, IntPtr ppBufferWirePacket)
If I delete the source and destination and recreate identical transforms, they work fine, but I don't feel like rebuilding all of the extracts. Any ideas! The problem occurs in all environments that we've tried.
TIA,
Michael Shugarman
P.S. I just tried the SP2 CTP, but that doesn't fix the problem.
Hi I have created a simple SSIS project on my client that carries out 4 Data Flow tasks, each one copying a few hundred rows from an Oracle 10.0.2 database. This works OK and will also run in debug mode fine.
I have copied the package to the file system on our development server and get the following error when in debug mode:-
[DTS.Pipeline] Information: Validation phase is beginning. Progress: Validating - 0 percent complete [OLE DB Source [1]] Error: SSIS Error Code DTS_E_CANNOTACQUIRECONNECTIONFROMCONNECTIONMANAGER. The AcquireConnection method call to the connection manager "Server.user" failed with error code 0xC0202009. There may be error messages posted before this with more information on why the AcquireConnection method call failed. [DTS.Pipeline] Error: component "OLE DB Source" (1) failed validation and returned error code 0xC020801C. Progress: Validating - 50 percent complete [DTS.Pipeline] Error: One or more component failed validation. Error: There were errors during task validation. Validation is completed [Connection manager "Server.user"] Error: SSIS Error Code DTS_E_OLEDBERROR. An OLE DB error has occurred. Error code: 0x80004005. An OLE DB record is available. Source: "Microsoft OLE DB Provider for Oracle" Hresult: 0x80004005 Description: "Error while trying to retrieve text for error ORA-01019 ". Validation is completed
If you go to the source of each flow task and select preview you can retreive the data.
I cannot get a simple package to execute a data pump to an access database from SQL2005. I have tried it in both SSIS and by running the Export data function. I have been able to write to this database in the past using dtp in SQL2000 but I am not able to write to it using SQL2005. What is the deal with the new SSIS? Does anybody have any ideas I can try to get my export to work. I have many more to do and I have to migrate over all of my SQL 2000 DTS packages to SQL2005 and some export to MSAccess.
This is the only error message I can find: [DTS.Pipeline] Information: "component "OLE DB Destination 1" (2196)" wrote 0 rows.
Edit: I found more errors in the debug section and a post here that discussed the problem as they had run into it. I was able to use part of that and some more research in order to tackle my problem.
I would still be interested in finding out why I suddenly had this problem arise after I upgraded to SQL2005. This is going to be a real pain as apparently SQL2005 treats NULL as Zero Length and now all of my databsaes that had that set in access will have to be modified to deal with this in the export.
The scenario is as follows: I have a source with many rows. Each row has a column called max_qty_value. I need to perform a calculation using another column called qty. This calculation is something similar to dividing qty/(ceiling) max_qty_value. Once I have that number I need to write an additional duplicate row for each value from the prior calculation performed. For example, 15/4 = 4. I need to write 4 rows to the same target table as in line information for a purchase order.
The multicast transform appears to only support fixed and/or predetermined outputs. How do I design this logic in SSIS to write out dynamic number of rows to a target table.
I can't find 'SQL Server: SSIS Pipeline' performance object in performance monitor on a 64-bit SQL Server. I see it on a 32-bit. Does anybody know why?
I am trying to use SSIS package to trasfer data and hitting road block for
[SQL Server Destination [37]] Error: SSIS Error Code DTS_E_OLEDBERROR. An OLE DB error has occurred. Error code: 0x80040E14. An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Could not bulk load because SSIS file mapping object 'GlobalDTSQLIMPORT' could not be opened. Operating system error code 2(The system cannot find the file specified.). Make sure you are accessing a local server via Windows security.".
I have been getting a recurring error while running the folowing script from an SSIS package. I have bolded the parts that I think may be of use. I didn't know if this would be a T-SQL or SSIS question, but thanks in advance for help.
Error: 0xC002F210 at Execute SQL Task 1, Execute SQL Task: Executing the query "declare @dbname varchar(200)declare @mSql1 varchar(8000)
DECLARE DBName_Cursor CURSOR FOR select name from master.dbo.sysdatabases where name not in ('mssecurity','tempdb') Order by name
OPEN DBName_Cursor
FETCH NEXT FROM DBName_Cursor INTO @dbname
WHILE @@FETCH_STATUS = 0 BEGIN Set @mSQL1 = ' Insert into [tempdb].[dbo].[DBROLES] ( DBName, UserName, db_owner, db_accessadmin, db_securityadmin, db_ddladmin, db_datareader, db_datawriter, db_denydatareader, db_denydatawriter ) SELECT '+''''+@dbName +''''+ ' as DBName ,UserName, '+char(13)+ ' Max(CASE RoleName WHEN ''db_owner'' THEN ''Yes'' ELSE ''No'' END) AS db_owner, Max(CASE RoleName WHEN ''db_accessadmin '' THEN ''Yes'' ELSE ''No'' END) AS db_accessadmin , Max(CASE RoleName WHEN ''db_securityadmin'' THEN ''Yes'' ELSE ''No'' END) AS db_securityadmin, Max(CASE RoleName WHEN ''db_ddladmin'' THEN ''Yes'' ELSE ''No'' END) AS db_ddladmin, Max(CASE RoleName WHEN ''db_datareader'' THEN ''Yes'' ELSE ''No'' END) AS db_datareader, Max(CASE RoleName WHEN ''db_datawriter'' THEN ''Yes'' ELSE ''No'' END) AS db_datawriter, Max(CASE RoleName WHEN ''db_denydatareader'' THEN ''Yes'' ELSE ''No'' END) AS db_denydatareader, Max(CASE RoleName WHEN ''db_denydatawriter'' THEN ''Yes'' ELSE ''No'' END) AS db_denydatawriter from ( select b.name as USERName, c.name as RoleName from ' + @dbName+'.dbo.sysmembers a '+char(13)+ ' join '+ @dbName+'.dbo.sysusers b '+char(13)+ ' on a.memberuid = b.uid join '+@dbName +'.dbo.sysusers c on a.groupuid = c.uid )s Group by USERName order by UserName'
--Print @mSql1 Execute (@mSql1)
FETCH NEXT FROM DBName_Cursor INTO @dbname END
CLOSE DBName_Cursor DEALLOCATE DBName_Cursor " failed with the following error: "Line 15: Incorrect syntax near '2003'.". Possible failure reasons: Problems with the query, "ResultSet" property not set correctly, parameters not set correctly, or connection not established correctly.Task failed: Execute SQL Task 1
When I have an alternet Data Flow in an event handler, caused by a record failing to be inserted due to a unique-key constraint violation, does this increment the number of errors, counting towards the MaximumErrorCount? How can I NOT count it as an error?
The thing is, I need to insert 300,000+ records each day, and some may be duplicates from data already in the table. So I set a unique key constraint on the table, and if during the load, it fails, it will trigger an alernate data flow to load the error records into another table. But if someone tries to load a file that already has been loaded, for example, all the records would be duplicates, which would be equivelant to 300,000+ errors, and I don't want to keep setting the MaximumErrorCount property higher and higher.
Is there any way to treat the error as "being handled" in the dataflow, so therefore doesn't treat it as an error? Or conversely, can I set the MaximumErrorCount property to 0 or -1 to accept all errors, no matter how many?
ID,Operator Emailed,Operator Net sent,Operator Paged,Retries Attempted 09/24/2007 15:00:00,Hourly Extract From OReSA,Error,0,INHSCTSTTOMVM82,Hourly Extract From OReSA,(Job outcome),,The
job failed. The Job was invoked by Schedule 7 (Hourly). The last step to run was step 1 (OReSA
Extract).,00:00:59,0,0,,,,0 09/24/2007 15:00:01,Hourly Extract From OReSA,Error,1,INHSCTSTTOMVM82,Hourly Extract From OReSA,OReSA
Extract,,Executed as user: INENVts_hia. hod call failed. End Error Error: 2007-09-24 15:00:58.93 Code:
I have been getting a recurring error while running the folowing script from an SSIS package. I have bolded the parts that I think may be of use. I didn't know if this would be a T-SQL or SSIS question, but thanks in advance for help.
Error: 0xC002F210 at Execute SQL Task 1, Execute SQL Task: Executing the query "declare @dbname varchar(200)declare @mSql1 varchar(8000)
DECLARE DBName_Cursor CURSOR FOR select name from master.dbo.sysdatabases where name not in ('mssecurity','tempdb') Order by name
OPEN DBName_Cursor
FETCH NEXT FROM DBName_Cursor INTO @dbname
WHILE @@FETCH_STATUS = 0 BEGIN Set @mSQL1 = ' Insert into [tempdb].[dbo].[DBROLES] ( DBName, UserName, db_owner, db_accessadmin, db_securityadmin, db_ddladmin, db_datareader, db_datawriter, db_denydatareader, db_denydatawriter ) SELECT '+''''+@dbName +''''+ ' as DBName ,UserName, '+char(13)+ ' Max(CASE RoleName WHEN ''db_owner'' THEN ''Yes'' ELSE ''No'' END) AS db_owner, Max(CASE RoleName WHEN ''db_accessadmin '' THEN ''Yes'' ELSE ''No'' END) AS db_accessadmin , Max(CASE RoleName WHEN ''db_securityadmin'' THEN ''Yes'' ELSE ''No'' END) AS db_securityadmin, Max(CASE RoleName WHEN ''db_ddladmin'' THEN ''Yes'' ELSE ''No'' END) AS db_ddladmin, Max(CASE RoleName WHEN ''db_datareader'' THEN ''Yes'' ELSE ''No'' END) AS db_datareader, Max(CASE RoleName WHEN ''db_datawriter'' THEN ''Yes'' ELSE ''No'' END) AS db_datawriter, Max(CASE RoleName WHEN ''db_denydatareader'' THEN ''Yes'' ELSE ''No'' END) AS db_denydatareader, Max(CASE RoleName WHEN ''db_denydatawriter'' THEN ''Yes'' ELSE ''No'' END) AS db_denydatawriter from ( select b.name as USERName, c.name as RoleName from ' + @dbName+'.dbo.sysmembers a '+char(13)+ ' join '+ @dbName+'.dbo.sysusers b '+char(13)+ ' on a.memberuid = b.uid join '+@dbName +'.dbo.sysusers c on a.groupuid = c.uid )s Group by USERName order by UserName'
--Print @mSql1 Execute (@mSql1)
FETCH NEXT FROM DBName_Cursor INTO @dbname END
CLOSE DBName_Cursor DEALLOCATE DBName_Cursor " failed with the following error: "Line 15: Incorrect syntax near '2003'.". Possible failure reasons: Problems with the query, "ResultSet" property not set correctly, parameters not set correctly, or connection not established correctly.Task failed: Execute SQL Task 1
at System.Data.Common.DbConnectionOptions..ctor(String connectionString, Hashtable synonyms, Boolean useOdbcRules)
at System.Data.SqlClient.SqlConnectionString..ctor(String connectionString)
at System.Data.SqlClient.SqlConnectionFactory.CreateConnectionOptions(String connectionString, DbConnectionOptions previous)
at System.Data.ProviderBase.DbConnectionFactory.GetConnectionPoolGroup(String connectionString, DbConnectionPoolGroupOptions poolOptions, DbConnectionOptions& userConnectionOptions)
at System.Data.SqlClient.SqlConnection.ConnectionString_Set(String value)
at System.Data.SqlClient.SqlConnection.set_ConnectionString(String value)
at System.Data.SqlClient.SqlConnection..ctor(String connectionString)
at ScriptTask_dbe60b31acf14201b5f012a4f1c608c2.ScriptMain.GetAppParameterCMD(String strParamaterName) in dts://Scripts/ScriptTask_dbe60b31acf14201b5f012a4f1c608c2/ScriptMain:line 49
at ScriptTask_dbe60b31acf14201b5f012a4f1c608c2.ScriptMain.Main() in dts://Scripts/ScriptTask_dbe60b31acf14201b5f012a4f1c608c2/ScriptMain:line 27
I am having a couple of problems with an SSIS package. The first is whenever I add a Data Viewer to one of my data flow tasks it crashes Visual Studio. This happens every single time. I have tried rebooting, but that didn't help. The second error I am getting is whenever I call ComponentMetaData.GetErrorDescription(Row.ErrorCode) from within a data flow script task it generates a "Catastrophic Failure" and indicates that it occurred when calling the GetErrorDescription method. I am not sure if something is corrupted or what the issue is. Does anyone have any insight?
that first "truncate the product table" then populate the table with new rows....if there is an error then i will invoke the send email task. However in the "truncate the product table" task, the sql i put "truncate dbo.product" instead of "truncate table dbo.product" but this error is not captured in the text file that I m going to send an email to the appropriate personnel.
Under logging, and when i go into "truncate product table" tasks , "Details", i saw "OnError", "OnWarning" and other error handlers...which should i tick in order for the text file to ONLY show errors I encounteered during the running of the SSIS package?
II am using file system for the SSIS Packages. I have several packages in a project. Most of these packages use configuration files. Infact there are several packages which use the same configuration files. Now when I try to run the build utility, it errors out saying that the name.dtsconfig file already exists. Seems like this is a bug. I would really like to use this utility.I know if the config file is not used in any other pacakge probably the build will not fail. But it is not good practice to have one config file for each package, it is redundancy, does not make sense to that. Is there any way I can use the build and deployment utility without this error.
I have a simple SSIS package I built to migrate data from a SQL db on one server to another SQL db on another server. All it does is: Execute SQL task to disable some triggers on the target table, Data Flow Task to pull the data from the source (a simple sql query) into the destination table (using OLEDB), then another Execute SQL task to re-enable the triggers.
This package ran fine several times during my testing. However, this morning it's not working anymore. When I run it and check its progress, the "Validation has started" is the only thing runs, then nothing else happens... no error messages, no timeouts, nothing in the output window besides "DTS.Pipeline: Validation phase is beginning".
The only thing that has changed is one change to a date in the where clause of my source query, so I'm baffled on why it no longer runs.
I'm pretty new to SSIS, so I could easily be missing something obvious here.
Anyone have any ideas on what's up, or who can point me in the right direction? I'm currently reading up on how to debug and add more comprehensive error handling, in hopes of finding a better description of the problem.
Hi guys, I got these errors when writing a scripting component. Anyone encounteer these errors before?
Warning 1 The dependency 'EnvDTE' could not be found. Warning 2 The dependency 'Microsoft.SqlServer.VSAHosting' could not be found. Warning 3 The dependency 'Microsoft.SqlServer.DtsMsg' could not be found. Warning 4 The dependency 'Microsoft.SqlServer.VSAHostingDT' could not be found.
Hi All, we have a set of packages which has to to be implemented across differnet environments. the packages invariable uses OLEDB Source/Destination Components. In these two components the Code Page Property is not Configurable. so if the package has to be Deployed in a different environment than,where it is Developed it gives a Validation Error. Is there a Workaround for this problem. if anybody have faced the problem earlier,please post it. thanks in advance. cheers srikanth
I continue to receive intermittent timeout errors from the SendMail task in my SSIS packages. I know that the SMTP server I am connecting to is often slow to respond to due other business needs, so my question is what property is available to increase the waiting period before a tiemout occurs with the SendMail task. I have not found a property in the task that allows me to say, wait 120 seconds for the SMTP server before throwing a timeout error.
I have a SSIS package that I have scheduled to run (under SQL Server Agent jobs) and when I check its history, it shows it being invoked regularly, and with no errors. The problem is, nothing is actually happening. I can load up the job in MS Visual Studio, and run it, and it runs fine, and takes care of the data transfer. The data is not getting transferred when it runs as a job under SQL Server Agent though the logs show that job was run, and completed with no errors.
I have another package that is similarly configured and does run correctly, and I have been checking the two packages to see what the difference in configuration is, but I don't see anything. All the package in question is, is a single foreach container that gets all the files in a directory, adds them as flatfiles to a database, and then moves them to an archiving directory.
I hope someone can help becuase this problem is issue us several headaches.
We are currently trying to deploy an SSIS package to a production server. The deployment goes fine, the package runs ok when executed manually. The issues start when we try and execute it under the SQL agent.
Having gone back to the drawing board and spent much of the day reading various articles and applying the various options (especially those within the MS KB article 918760), we are still no closer to a resolution.
The SSIS package was created under an Administrator, and the SQL agent runs under a different Domain Admin account.
When we set up the Schedule to read from SQL Server or the SSIS Store the standard "Executed as user: DOMAINUSERNAME. The package execution failed. The step failed" in the history.
We tried to create the package as a file access and now get "Package could not be found" even though you can browse to i in the schedule list. The Domain account as full access to the folder where the package resides.
Has anyone else come across this issue, or have a workable solution?
I tried to execute SSIS package that created by a sql developer contractor.
There are several errors that i am not familiar with:
-Error at Get results to faw file [DTS Pipeline]: input column "" (221) has lineage ID 32 that was not previously used in the Data Flow task.
- Error at Get results to raw file [DTS.Pipeline]: "component "Raw File Destination" (208) failed validation and returned validation status "VS_NEEDSNEWMETADATA".
-Error at Get result to raw file (DTSipeline): One or more components failed validation.
(Microsoft Data TransformationServices VsIntegration).
Can someone guide me since I am consider myself a newbie to SSIS.
We just upgraded our accounting packages database from Pervasive 8.1 to V9.11. MY SSIS extract packages were working just fine. Now I am getting several erros stating that some columns cannot be found:
For example, in the [sysdtslog90] table, I have these errors:
Column "CreditLimitAmt" cannot be found at the datasource. or
Column "RsrvRptky" cannot be found at the datasource. (from a different pkg)
The table schemas did not change, the columns still exist. The SQL I use also runs just fine in the client interface.
Has anyone experienced this? The bottom line is that now the extract packages are broken.
I have an SSIS package that fires an OnTaskFailed event whenever any of my tasks fail. I would like to put any SSIS task failure message into a user defined variable. Any idea how to do this? Any help appreciated. Regards, Paul.
I get errors when I check out an SSIS package from source control (both Source Safe and TFS) relating to the connection objects where I was not the original developer who checked it in. Is there a solution to this other than altering the connection loginpassword for every connection object in the package before deploying?
I am in the process of creating an Integration Services package to automate the process of training mining models and getting predictions. Until recently, I have been processing the models directly from Business Intelligence Studio without any problems. However, when I try to use the exact same training set as an input to the Data Mining Model Training destination, I get several errors. Here is the output:
[Mining Models [1]] Error: Parser: An error occurred during pipeline processing. [Mining Models [1]] Error: Errors in the OLAP storage engine: The process operation ended because the number of errors encountered during processing reached the defined limit of allowable errors for the operation. [Mining Models [1]] Error: Errors in the OLAP storage engine: An error occurred while the 'CPT MODIFIER' attribute of the 'BCCA DMS ~MC-CLAIM LIN~5' dimension from the 'BCCA LRG DMS TEST' database was being processed. [Mining Models [1]] Error: File system error: The record ID is incorrect. Physical file: . Logical file: . [Mining Models [1]] Error: Errors in the OLAP storage engine: The process operation ended because the number of errors encountered during processing reached the defined limit of allowable errors for the operation. [Mining Models [1]] Error: Errors in the OLAP storage engine: An error occurred while the 'BILL TYPE' attribute of the 'BCCA DMS ~MC-CLAIM LIN~5' dimension from the 'BCCA LRG DMS TEST' database was being processed. [Mining Models [1]] Error: File system error: The record ID is incorrect. Physical file: . Logical file: . [DTS.Pipeline] Error: The ProcessInput method on component "Mining Models" (1) failed with error code 0x80004005. The identified component returned an error from the ProcessInput method. The error is specific to the component, but the error is fatal and will cause the Data Flow task to stop running.
I have not been able to find an answer as to why this is happening. I found a post regarding a similar problem with processing an OLAP cube in SSIS, but it seems that the author of that post never found an answer. Has anyone else here seen similar errors when processing mining models from Integration Services?
Also, if I process the mining models manually then try to run only predictions in SSIS, I get many of the same errors. I'll keep looking into the problem myself, but I would be very grateful if someone in this forum could shed some light on this issue.
I am trying to handle errors occurred in SSIS package.
I have created flat file destination task in Data flow to handle the error records,configured the output columns with redirect row option.I am getting below error when I execute the package.
Error: 0xC0202009 at DTSTask_DTSDataPumpTask_2, OLE DB Source [1]: SSIS Error Code DTS_E_OLEDBERROR. An OLE DB error has occurred. Error code: 0x80004005.An OLE DB record is available. Source: "Microsoft OLE DB Provider for SQL Server" Hresult: 0x80004005 Description: "The statement has been terminated.".An OLE DB record is available. Source: "Microsoft OLE DB Provider for SQL Server" Hresult: 0x80004005 Description: "Subquery returned more than 1 value. This is not permitted when the subquery follows =, !=, <, <= , >, >= or when the subquery is used as an expression.".
Package execution is stopping and none of the error records are logged in the error file task.
I want to incorporate this code but I dont know how to import Microsoft.SqlServer.Dts.Pipeline in an Integration Services Project template. I was thinking of putting this code in the script task but still, I cant import Pipeline. Add reference list does not have it as well. Please let me know how to incorporate this code. Thanks!
Code: if (ComponentMetaData.RuntimeConnectionCollection["SourceFileConnection"].ConnectionManager != null) { cm = DtsConvert.ToConnectionManager(ComponentMetaData.RuntimeConnectionCollection["SourceFileConnection"].ConnectionManager);
I am re-writing old DTS packages(from sql 2005) to convert them to SSIS packages(sql 2014) and in one of the script task, the old activex script does not run.
In my SSIS package, I have a backup database task. When I run the package with DestinationAutoFolderPath set to a folder ("Network Service" account has full permission on this folder) and DestinationCreationType set to Auto, the task works just fine creating a backup with its own name. (similar to database_date<count>).
But what I want is in my front-end I am allowing the user to specify the name of the backup file. So I want the task to create the backup file in the name I supply. I set the DestinationCreationType to manual and in the application code added the DestinationManualList with the path from the UI.
Now the pacakge runs fine but does not take any backup. There is no errors as well. If I set the FailPackageOnFailure and FailParentOnFailure to true, then I am getting the DTSExecResult.Failure but I am not getting the actual error from the backup database task.
I have created ssis package. It has been successfully running at UI level.
But when i am trying to execute it from the command prompt by using dtexec utility it is showing the following error messages.
Error: 2005-12-23 17:01:57.67 Code: 0xC00470FE Source: Data Flow Task DTS.Pipeline Description: The product level is insufficient for component "Flat File Source" (1). End Error Error: 2005-12-23 17:01:57.67 Code: 0xC00470FE Source: Data Flow Task DTS.Pipeline Description: The product level is insufficient for component "Script Component" (9). End Error
I have added an email task to the ON Error Event of my SSIS package, so that I will always know when there are errors. However I would like the SQL Server job executing the package to succeed even if the package fails. What setting do I change in the SSIS packageto achieve this? MaximumErrorCount?
I'm currently trying to pull data from a ProvideX database and replicate it in a collection of SQL Server tables. However, I'm having a heck of a time trying to convert some strange decimals stored by the ProvideX database. As an example of the data I'm trying to retrieve, I'll see something like [. 1] or [. 1] ([]'s are to show the bounds of the field). After analyzing the data, it seems the decimal in the field represents a 1,000 placeholder. Thus [. 1] really means 1, and [. 1] really means 10. Something like .100 would be 100. 6.500 would be 6500.
As you can imagine, the spaces are causing errors when trying to pull the data, and I can't for the life of me figure out to just pull it as a string, run a script to convert it to a correct number, and then save the transformed data into SQL Server. When running the import wizard, it seems I'm being forced to pull these columns as decimals. Currently I'm trying to just pull the data out "as is" and throw it in a raw file, to be processed out of SSIS. Obviously doing it all within SSIS would be ideal, but if that can't be done, I'll do whatever it takes. I should also say I'm new to SSIS packages, but not necessarily new to SQL Server or SQL in general.
1) How can I pull these columns as strings? If I try to change the Export columns in the source query data flow step, it gives me an error saying that I can't do that.
2) If I have to pull as decimals, how can I capture the row on error, process it, and send it back to the export? So far, when I get an error, I lose all information in the row to the right of and including the error field.
I appreciate any responses, as I'm kind of going in circles at this point. If this sort of thing has been discussed here prior, I apologize...I didn't find it in any searches I did. Please just point me in the right direction if you've dealt with this sort of problem before. It seems to me that it should be an easy thing to do. I'm just not finding any tutorials on it.