Problem In Loading Roman8 Characters (ä, ü )into SQL Server.
Jul 20, 2005
We are trying to load text tab delimited files into SQL server using
informatica 5.1 version. The text files are FTPed to a local server
and then files are moved to applicable folder before loading.
The problem is when we open the files after FTP the characters are
already converted to something else. EX: ü got converted as Ï.
Now when we load them into SQL server we want Ï to transfer as ü
again.
In informatica we use relational database connection and specify SQL
server as server type, there is no place where you can specify the
character conversion parameter.
Can anybody help solving this problem ?
Thanks in advance.
Ram.
View 4 Replies
Sep 8, 2006
I'm using the For Each loop container to load multiple XML data files into SQL Server, and noticing some peculiar behavior and need some advice.
The pattern I'm trying to accomplish is this: Iterate over a collection of XML files in a specific folder, loading each in turn into SQL Server. If the file has already been loaded, delete the records first before the load. After the load succeeds, move the file into an Archive folder.
To accomplish this, I've set up a For Eac Loop container using the For Each File enumerator, and retrieve just the file name and extension into a variable. The first task in this is an Execute SQL task that uses a SQL DML statement to delete records based on a field in the table containing the file name (DELETE FROM table WHERE PROG_NAME = ?), and map a variable to the parameter. The next task is the data flow task that uses an XML source using the variable as the file name, and SQL Server destination. I use a derived column task in between to plug the variable holding the file name into the PROG_NAME field. So far, so good. This works.
But now comes the peculiar part. I initially had the XSD files in the same folder as the XML files, but wanted to put them in their own directory, so moved them, and made the change to the XML source adapter for the new path to the XSD file. The next time I ran my package, it failed. For some reason, as the For Each Loop tried to iterate over the directory, it was using the XSD path assigned in the XML source instead of the path for the XML files. Unusual...
My question is, why when choosing the File name & Extension retrieval type (as opposed to the fully qualified name) will the task try to use the XSD location to find the files? Is my variable getting reassigned somewhere?
View 2 Replies
View Related
May 27, 2008
I have requirement to load the EDI.TXT format data in SQl server using the SSIS.The edi file data looks like that
@pRecType="A",@pA010="XC",@pA020="270",@pA110="CDC LOCAL"
@pRecType="C",@pC010="M",@pC015="H",@pC050="20080408",@pC060="B",@pC070="E40245",@pC080="P",@pC110="95000",@pC112="000000000",@pC120="P",@pC121="N",@pC122="",@pC124="100000002274",@pC125="166759",@pC210="Y",@pC301="000061131006",@pC320="20080211",@pC321="20080211",@pC511="",@pC512="",@pC900=" ABY"
@pRecType="E",@pE100="1",@pE101="5789",@pE110="",@pE111="",@pE120="",@pE121="",@pE130="",@pE131="",@pE140="",@pE141="",@pE150="",@pE151="",@pE160="",@pE161="",@pE170="",@pE171="",@pE180="",@pE181="",@pE190="",@pE191=""
@pRecType="H",@pH110="8",@pH120="3",@pH130="1",@pH210="",@pH220="20080211",@pH230="20080211",@pH235="01",@pH240="20080211",@pH250="20080211",@pH310="",@pH320="",@pH710="",@pH711="",@pH712="",@pH510="",@pH511="",@pH512="",@pH520="",@pH521="",@pH522="",@pH530="",@pH531="",@pH532="",@pH540="",@pH541="",@pH542="",@pH550="",@pH551="",@pH552="",@pH560="",@pH561="",@pH562="",@pH570="",@pH571="",@pH572="",@pH580="",@pH581="",@pH582="",@pH590="",@pH591="",@pH592="",@pH600="",@pH601="",@pH602="",@pH610="",@pH620="",@pH630="",@pH640="",@pH650="",@pH660="",@pH670="",@pH680="",@pH690="",@pH713=""
@pRecType="M",@pM250="170"
@pRecType="P",@pP010="SE",@pP110="E",@pP130="",@pP150="721153832",@pP210="",@pP310="",@pP340="",@pP350="",@pP360="",@pP370=""
@pRecType="S",@pS010="",@pS020="200278911",@pS110="BABIN",@pS111="JENNIFER",@pS112="L",@pS120="",@pS121="",@pS125="",@pS126="",@pS127="",@pS211="BABIN",@pS212="JENNIFER",@pS213="L",@pS221="F",@pS222="19850919",@pS231="M",@pS310="1605 SOUTH SHIRLEY",@pS311="",@pS315="GONZALES",@pS316="LA",@pS317="70737",@pS410="N"
@pRecType="V",@pV130="AA",@pV140="000000100",@pV150=""
@pRecType="X",@pX010="20080211",@pX020="20080211",@pX131="45378SG",@pX133="5789",@pX134="",@pX141="000095000",@pX142="000000000",@pX147="001",@pX210="0490"
can you help me how I solve this problem.I will appreaite for that id you help me this.
Thanks
HK
View 4 Replies
View Related
Apr 24, 2006
I am trying to load 14+ million rows from a text file into local Sql Server. I tried using Sql Server destination because it seemed to be faster, but after about 5 million rows it would always fail. See various errors below which I received while trying different variations of FirstRow/LastRow, Timeout, Table lock etc. After spending two days trying to get it to work, I switched to OLE DB Destination and it worked fine. I would like to get the Sql Server Destination working because it seems much faster, but the error messages aren't much help. Any ideas on how to fix?
Also, when I wanted to try just loading a small sample by specifying first row/last row, it would get to the upper limit and then picked up speed and looked like it kept on reading rows of the source file until it failed. I expected it to just reach the limit I set and then stop processing.
[SS_DST tlkpDNBGlobal [41234]] Error: An OLE DB error has occurred. Error code: 0x80040E14. An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Cannot fetch a row from OLE DB provider "BULK" for linked server "(null)".". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "The OLE DB provider "BULK" for linked server "(null)" reported an error. The provider did not give any information about the error.". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Reading from DTS buffer timed out.".
--------------------------------
[SS_DST tlkpDNBGlobal [41234]] Error: The attempt to send a row to SQL Server failed with error code 0x80004005.
[DTS.Pipeline] Error: The ProcessInput method on component "SS_DST tlkpDNBGlobal" (41234) failed with error code 0xC02020C7. The identified component returned an error from the ProcessInput method. The error is specific to the component, but the error is fatal and will cause the Data Flow task to stop running.
...
[FF_SRC DNBGlobal [6899]] Error: The attempt to add a row to the Data Flow task buffer failed with error code 0xC0047020.
[DTS.Pipeline] Error: The PrimeOutput method on component "FF_SRC DNBGlobal" (6899) returned error code 0xC02020C4. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.
[DTS.Pipeline] Error: Thread "WorkThread1" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.
-------
After first row/last row (from 1 to 1000000) limit is reached:
[SS_DST tlkpDNBGlobal [41234]] Error: An OLE DB error has occurred. Error code: 0x80040E14. An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Cannot fetch a row from OLE DB provider "BULK" for linked server "(null)".". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "The OLE DB provider "BULK" for linked server "(null)" reported an error. The provider did not give any information about the error.". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Reading from DTS buffer timed out.".
---------------
When trying to do a MaximumCommit = 1000000. Runs up to 1000000 OK then slows down and then error.
[SS_DST tlkpDNBGlobal [41234]] Error: Unable to prepare the SSIS bulk insert for data insertion.
[DTS.Pipeline] Error: The PrimeOutput method on component "FF_SRC DNBGlobal" (6899) returned error code 0xC02020C4. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.
----
When attempting all in a single batch:
[OLE_DST tlkpDNBGlobal [57133]] Error: An OLE DB error has occurred. Error code: 0x80004005. An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80004005 Description: "The transaction log for database 'tempdb' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80004005 Description: "Could not allocate space for object 'dbo.SORT temporary run storage: 156362715561984' in database 'tempdb' because the 'PRIMARY' filegroup is full. Create disk space by deleting unneeded files, dropping objects in the filegroup, adding additional files to the filegroup, or setting autogrowth on for existing files in the filegroup.".
View 11 Replies
View Related