Wether I'm trying to access a datapump task in an imported package in SSIS, or in 2000 enterprise manager
they crash once created. I can not alter or redo mappings or anything else.
After enterprise manager churns for 5-10 min it crashes.
SSIS partially crashes: Attempted to read or write protected memory. This is often an indication that other memory is corrupt. After this I can no longer click on anything and have to end task on Visual studio.
I can do a disconnected edit on the package in enterprise manager and alter the statement, but as far as I know you can't really change much as far as mappings back in there.
This also means that my 2000 packages have been limited to 1 teradata pump per package.
This happened on my laptop, 2 coworker's desktops, and now that i've migrated to a new high end developer box it's happening on it also.
I've spent quite a few hours on this one... I've successfully gotten tables to datapump between source and destination datbases, however... I get the following message when I hit a table with a blob field:
Step 2 failed, error: x80040E21, x80040000 + 3617 General error -2147217887 (80040E21). (Microsoft Data Transformation Services (DTS) Data Pump (80040e21): Cannot instantiate Storage Object for source column 6 ('description') Blob data. Your provider may require that all Blob columns be rightmost in the source result set.) (Microsoft OLE DB Provider for ODBC Drivers (80040e21): Errors occurred)
I've written a VB com dll that does a simple datapump... but am stuck on this "Storage Object" instantiation thingy...
Any recommendations as to what I can put in my datapump object to allow just a simple xfer of blob types?
I get this error when I execute a job calling a dts pkg. The package itself is fine and executes without any error.It's transferring around 400000 records. Shceduling the pkg gives the following error. Can any body led a hand please?
DTSStep_DTSDataPumpTask_1 DTSRun OnProgress: DTSStep_DTSDataPumpTask_1; 1000 Rows have been transformed or copied.; PercentComplete = 0; ProgressCount = 1000 DTSRun OnProgress: DTSStep_DTSDataPumpTask_1; 2000 Rows have been transformed or copied.; PercentComplete = 0; ProgressCount = 2000 DTSRun OnProgress: DTSStep_DTSDataPumpTask_1; 3000 Rows have been transformed or copied.; PercentComplete = 0; ProgressCount = 3000 DTSRun OnProgre... Process Exit Code 1. The step failed.
On my MS SQL Server 2000, I am trying to create a generic way to load tables into my datawarehouse.
I have as input to the process a large number of table definition(s) stored individually as files on my server. And, ascii delimited data files in various locations but mostly accessible via NFS mounts.
I created two DTS package in MSSQL2K that in theory represents what I want to do:
package1 ... invoke package2 with global variables to load a system of related tables
package2 ... check for a trigger file ... set the "Execute SQL Task" statement to my first file ... run the "Execute SQL Task" which drop/add's a table ... set a "Connection" to a data source file that I want to use ... run the transformation and, with that my package starts to fall apart ... set the "Execute SQL Task" statement to the next file, and ...... goback and execute it
I can't figure out how to set the table in the transformation section to the table I want to use. And, I assume next to have the transformations links between the source and new table relinked.
The source files contain in the first row the column names as found in the tables I just created.
I'm using an SSIS under SQL 2005; i transfer data with a DataFlow Component:
DataSource = OLEDB Source;
Destination = OLEDB Destination
The source is a View returning about 100000 records; if i execute it in query analyser it takes few seconds. But if i use it as DataPump Source it takes HOURS to execute it...
I've noticed that If i modify the SELECT query inserting a "TOP 10000000000", it takes few seconds..
I am scripting a DTS Package using VB. The problem I am having is that I get the following error when I execute the package from Visual Basic:
Step Error Source: Microsoft Data Transformation Services Flat File Rowset Provider Step Error Description:Incomplete file format information - file cannot be opened.
In the DTS Designer when I open the DataPump Transformation and click on the Destination tab I am prompted With the Define Columns dialog. I define the columns and click execute. The package works properly after doing this when executed from Enterprise Manager. Does anyone know how to script the file definition creation??? I have searched high and low through BOL and other resources and am not finding anything. I doesn't seem like it should be this hard to figure out so, maybe I am missing something.
I want to export to a Text File (destination) from a SQL query using a DTS DataPump Task. My query has 28 columns, some how when I try to define the destination columns for my text file the Microsoft Managment Console Crashes completely.
I tough may be my query has two many columns only to find out that it has one to many. If I ramove a column form my query, any column. I get no error at all.
¿Is there a limit to a Text File destination connection?
I am trying to query Teradata tables from SQL Server 2005 using SSIS that will populate a table in SS 2005. I have been able to test the connection successfully. I choose the Query option b.c the tables in Teradata are over 100 million records and I don't want to copy those tables locally.
I am able to enter my query and parse it. I can even preview the query and see the results. A very simple query:
select distinct acc_yr from d4574fdp.clm_dim
But when I try and execute the query, I get the following error message:
quote: TITLE: SQL Server Import and Export Wizard ------------------------------
Could not connect source component.
Error 0xc0202009: Source - Query [1]: An OLE DB error has occurred. Error code: 0x80040E21. An OLE DB record is available. Source: "OLE DB Provider for Teradata" Hresult: 0x80040E21 Description: "Multiple-step OLE DB operation generated errors. Check each OLE DB status value, if available. No work was done.".
I'm trying to convert old DTS packages over to SSIS and also create new SSIS packages that use teradata as a warehouse source for many different departments.
So far i've tried what I can find here on the forums:
using special syntax and wrapping the sql command in an alias/variable name
Methods i've tried: using an ODBC connection string that works just fine in VB for pulling data. - test connection succeeds but when I select data reader source I get:
Error at Data Flow Task [DataReader Source[185]]: Cannot acquire a managed connection from the run-time connection manager
Using an OLE DB connection where I get the standard error many have gotten and I tried the listed use special syntax and wrapping the sql command in an alias/variable name from : MSDN Error at Data Flow Task [Source - Testing]: An OLE DB error has occurred. Error code: 0x80040E21. An OLE DB record is available. Source: "OLE DB Provider for Teradata" Hresult: 0x80040E21 Description: "Multiple-step OLE DB operation generated errors. Check each OLE DB status value, if available. No work was done.". Currently i'm testing/trying this on a very simple view that produces 2 columns and 173 rows of string data. I have not tried the .net provider and am reluctant to accept that as the only solution if it works as it would involve permissions issues and red tape here at the bank and is a less user-friendly solution for my dev team.
Any ideas?
Driver version on my machine and likely the other 5 that would need this to work: 3.03.00.01 12/17/2003
I have a 32 bit development/playbox SQL 2005 SSIS that can see and run 32 bit OLE DB drivers from Teradata. Granted I have the same issues as others where a table or a view would work and not a query or a variable query. The alternative is to use Linked Server and OPENQUERY.
But our real development box is a 64 bit IA 64 using 64 bit SQL 2005. Non of the Teradata.com drivers would work/show up in the dropdown box of the Connection Manager since they are 32 bit. One option is to use the 64 bit Teradata ODBC drivers and then use the .Net ODBC connection instead. Linked server doesn't work either because the SQL 2005 is 64 bit and it is looking for 64 bit drivers.
Has anyone have any experience with 64 bit SQL 2005 and Teradata?
Looked at other 3rd party vendors such as ETI but their documentation excludes IA 64 (Itanium).
Planning to try again using the 64 bit ODBC and .Net Adapter but I'm interested to hear if someone (MSFT & NCR) would have a real solution for 64 bit.
I want to add a step to an SSIS package to passthrough native Teradata sql statements so that they can be executed on Teradata. Does anyone know if I can use an "Execute SQL Task" to accomplish this? It should be very similar to passing native Oracle SQL to an Oracle server.
I've been informed that our data feed systems are going to be shut down and that our data will be residing in Teradata V2R5 databases. I was wondering if anyone knows if you can get connectivity to a Teradata server via Linked Server in the Security folder of SQL Server Enterprise Manager? If no then; Is it possible through a DTS package? Thanks for the info...
I am extracting data from an SQLServer database to load into Teradata using DTS. The performance is abysmal. The same data in a text file loads quickly via multiload. I can move the data to other DBMSs via DTS quickly as well. Is there some way for me to improve the elapsed time/performance while using DTS? If not, what is the best way to move data from SQLServer into Teradata?
Still having a problem with my SQL WHERE clause…
I have three variables on an Excel form that connects to a Teradata database. The first variable is a date format (DateworkedF and DateworkedT) the other two are text fields. (StatusX and ErrorTypeX)
I want to be able to search on any or all of these fields. (If the field is blank return all values)
I am trying to set up the Linked 2005 Server to Teradata v2r5, and have got most of the way there. My problem is that using both methods (OPENQUERY and SELECT with 4-part naming string) seems to submit only the SELECT * part to Teradata with the rest of the SQL waiting until the data is all pulled back to SQL Server. This is a problem when working with big tables! It puts a huge strain on the network, and also runs me out of temp space in Teradata. Is there a way to get ALL the SQL passed to Teradata? I am looking to create a Report Model for Report Builder or Pro-Clarity. I have a dummy SQL Server 2005 database set up with views of the Teradata tables through the Linked Server. Any help would be greatly appreciated!
I recently installed the Teradata .NET data provider and I am trying to use named parameters and multi-value parameters in my SQL. I am finding information on this topic hard to come by.
In order to get parameters to work do I need to write my own Teradata data extension from scratch, expand an existing data extension to work with Teradata, or can I simple edit my .config files to point to a generic extension wrapper already existing?
I'm using SQL Server 2005 Enterprise on a Win2003 x64. I want to connect to a Teradata 2.6, but there are no x64 ODBC- or OLE-DB-drivers for Teradata yet. So, only OLE-DB can't be used as the 32-Bit OLE-DB-driver for Teradata is not visible e.g. if you want to create a linked-server-connection. One workaround would be to use "OLE-DB for ODBC", but this OLE-DB is also not available on my system.
Does anybody know a way to connect SQL 2005 x64 to Teradata or to get "OLE-DB for ODBC" up and running???
There is an entry to a similar issue (http://forums.microsoft.com/MSDN/ShowPost.aspx?PostID=699116&SiteID=1) where the workaround "use 64Bit-ODBC with .net ODBC" (?) puzzles me!
Does anyone have any experience setting up paramters when using a Teradata data source? I am specifically looking for how to set up dates and multi-select drop down boxes as parameters. Using a fixed query against Teradata using SSRS is straight forward and is working well. Parameters, not so much.
Please let me know whether we can use Teradata as Source in SSIS (Target is SQL Server 2005). ie. Do we have oledb driver for NCR Teradata to connect to it. Our SSIS will be hosted in 64 bit SQL, but for development we use 32 bit.
I've created a data source that uses the OLE DB Provider for Teradata to connect to an existing Teradata database. Works fine.
I then use this data source within my project to create a Data Source View. Again, all works well.
However, the final step - creating a Report Model, bombs out in the wizard at the Completing the Wizard step with the following error: "[NCR][Teradata Database] Syntax error, expected something like a name between the 'SET' keyword and the 'TRANSACTION' keyword." (I elected for the wizard to only create entities for all tables, and create attributes).
If I try to create a Model manually, I get the following error when trying to build: "The Entity '<EntityName>' does not have any IdentifyingAttributes. Entity must have at least one IdentifyingAttribute."
Still fairly new to SSRS, so I'm hoping the issue is just a loose nut behind the keyboard Does anyone know what I'm doing wrong in trying to set this up?
I am fetching large amount of data from teradata to sql server using linked server. I am facing below query:
A transport-level error has occurred when receiving results from the server. (provider: TCP Provider, error: 0 - The semaphore timeout period has expired.)
Hi I have a problem to connect to TeraData from SSIS. I have the TeraData client (Query Man) that I connect to the TeraData using ODBC connection. When I connect with Query Man I can see the objects (views) of the correct scheme. However, when I try to connect with SSIS using the same ODBC connection €“ I don€™t see the objects of the specific scheme. The way that I connect to TeraData using SSIS is creating DS connection (with odbc) and then I create DSV €“ where this is the point that I can see lots of other schemes objects, but not the scheme€™s objects that I can see in the Query Man. Can anyone assist?
I am fetching large amount of data from teradata to sql server using linked server. I am facing below query:
A transport-level error has occurred when receiving results from the server. (provider: TCP Provider, error: 0 - The semaphore timeout period has expired.)
I have an SSIS package that successfully runs a Teradata connection query. However, when I schedule the job to run in the SQL Server Agent I get the following error:
Executed as user: SERVER_NAMEAdministrator. Microsoft (R) SQL Server Execute Package Utility Version 11.0.2100.60 for 64-bit Copyright (C) Microsoft Corporation. All rights reserved.   Started: 7:53:26 AM Error: 2015-06-08 07:53:26.78    Code: 0x00000216    Source: Package1 Connection manager "Teradata"
[Code] ....
Why would a connection work in SSIS but not in a scheduled SQL Server Agent job?
This webcast from our partner, ETI, may be of interest to readers on the forum - we see questions quite often about high-performance interaction with other databases.
I've build a few reports in visual studio, they work and they don't work, in a minute I can run them next minute I can't and I get this errors in event viewer every time they can't be run:
Report Server (MSSQLSERVER) cannot load the SQLPDW extension.
Report Server (MSSQLSERVER) cannot load the TERADATA extension.
SP2 for SQL Server 2005 x64 has been less than happy for me.
The initial problem, mentioned in another thread, concerned SSIS client which stopped liking Data Flow components. It was solved by a removal and reinstallation of the client tools.
However, there now seems to be a problem with making use of the Teradata oledb provider within SSIS. Prior to the installation of SP2 there was no problem.
I create a new SSIS package Add a Data Flow component Within the data flow add an oledb source Create a new connection - using the Teradata oledb provider Test the connection (it works!) Use SQL something like "select * from some_teradata_table_on_the_TD_server" Click on columns, wait a second, a list of columns is returned Add a SQL Server destination (with a new SQL Server connection) Connect the two components Save the package
Click go The TD oledb source component turns bright red and the error is "class not registered"
I have uninstalled the TD oledb provider (and underlying TD components) and re-installed (32-bit and 64-bit) - no change
I'm new to SSIS, using Terdata Attunity connector for integrating data flow between Terdata (source) to SQL Server (Target).
SSIS package is getting failed because of length mismatch between source and Target for Unicode character datatype columns. Reason is Teradata TPT always occupies 3 times more length of actual defined in DB.Â
Even I tried by increasing length of attribute in Source but it didn't work.
I know by converting datatype from unicode -> Latin would work, but i don't want to do conversion since loosing some characters.
################################################## Error is [Teradata Source [263]] Error: TPT Export error encountered during Initiate phase. TPTAPI_INFRA: API306: Error: Conflicting data length for column(5) -  Source column's data length is (200) Target column's data length is (300). ##################################################