Loading Flat File With Embedded Column, Text Delimiters And Newline
Jun 14, 2007
I have the misfortune of converting a DTS package to SSIS that loads a flat file that has a text fields that can contain embedded text delimiters ("), column delimiters (,) and even new lines (CR+LF i.e.,hex 0D 0A) in it. A sample line from the file is posted here, remember this is just one line though it shows as three lines, since the third field has embedded new line in it:
4,"Sam","EVP; MARKETING PRODUCT MANAGER ""Level I"",
Internet Sales / HELP
8005551212",100
If you open in excel it handles it perfectly showing four fields, as below, and this is what I want ( I cannot get it aligned right in the posting, just save the above line in *.csv and open to see what it should be):
4 Sam "EVP; MARKETING PRODUCT MANAGER ""Level I"", 100
Internet Sales / HELP
8005551212"
Now, SSIS errors on the embedded text delimiters and breaks into two or three lines based on which option I chose. I have tried few options based on postings in the forum:
a). Using undouble and undoubleout: Does not work when there are embedded column delimiters (,) in the text field
b). Modified undouble script posted by lvovg at http://forums.microsoft.com/MSDN/ShowPost.aspx?PostID=1718225&SiteID=1. Handles the embedded column delimiters (,) perfectly, but the embedded new lines (CR+LF i.e.,hex 0D 0A) are breaking it.
Since, I am using the ragged right format to read from the file then use transform script on the line by lvovg, the line is already broken by the ragged right format at the embedded new lines, hence does not work.
Right now I am stuck. Can someone please help (anyone from MS) ? I am already baffled at the amount of coding required to convert a very basic ( and working ) flat file load DTS package to SSIS. I am willing to persist bit longer to convert this to SSIS, before I give up and stick with DTS and wait for a fix / workaround.
Thanks.
View 7 Replies
ADVERTISEMENT
Aug 31, 2007
Hey Everyone,
I have a Data-Flow task embedded in a Sequence Container (does not fail component on error) on the Control Flow panel of the SSIS designer. This data flow task contains a connection to a Flat File Source -> A Data Transformation -> Into an OLE Db Destination.
The problem is that the Flat File isn't always delimited properly -> the client cannot be relied on to do this.
My question is when the delimiters are messed up, how can I capture the offending error row(s) from the Flat File Source?
What I've tried:
1) Set every column in the source flat file on error to: Redirect Row
2) Added a Script Transformation to pull the description and the record id out of the offending row
3) Added an Error file flat file destination to the end of the flow.
The package always fails on the Flat File Source and never Redirects the offending Row to the error output - I never see my onError Script Transformation go Green, Red, or Yellow - SSIS doesn't let it get there.
I'm really new to SSIS so sorry if this is a super basic question.
Here is the Error Text:
[Source - InventTable_csv [1]] Error: The column delimiter for column "RECID" was not found.
[Source - InventTable_csv [1]] Error: An error occurred while processing file "C:------InventTable.csv" on data row 15228.
[DTS.Pipeline] Error: The PrimeOutput method on component "Source - InventTable_csv" (1) returned error code 0xC0202092. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.
[DTS.Pipeline] Error: Thread "SourceThread0" has exited with error code 0xC0047038.
I do have the MaxErrorCount set to 1 on the Data Flow Task but still think I should see my script task execute and a log entry be generated.
Any Ideas?
Matt
View 15 Replies
View Related
Aug 28, 2007
We're having issues exporting a set of data from SQL to a fixed width flat text file by just doing a right click on the DB, then choosing Tasks > Export Data. You can not specify a row delimiter when you choose a Fixed Width format. The only way around this that we've found is to specificy char(13) and char(10) at the end of the SQL select statement. Without row delimiters you end up with 1 giant record rather than 20,000 regular sized records. Is there any other way around this that we're missing?
Using Ragged Right is not an option either since the record lengths will be inconsistent if the last field doesn't contain a consistent length to the data.
Thanks,
Mike
View 16 Replies
View Related
Mar 20, 2008
Hi,
I have a flatfile which is of the following format. I want to import this data into the sql server 2005 database.
The alphabets at the start of each line represents the type of record. The data are position specific.
XYZQ20214646001V0608
A1249220AUGPSEA D11I0000
BCA030552050000C1234567841234412000
CFO093441050000C0895632175223652000
D100C5D0AA00C5D366
D200C5D92E00C5DB9F
E002720AUGSEA 0NRT 008D00000001 00C5D0E600C5D348N0000000000000
0000000 00000000CA0000FO0000
F002622AUGNRT 0SEA 008F00000002 00C5D96A00C5DB81N0000000000000
0000000 00000000CA0000FO0000
for ex in Record A, the following are the values that are based on column length
1 - (col length is 1)
249 (col length is 3)
20AUG (col length is 5) etc..
P
SEA
D
1
110000
View 2 Replies
View Related
Mar 15, 2006
I have a flat file that has fields delimited by tilde {~} and rows delimited by {CR}{LF}. The problem is, my source is spitting out the {NULL} (&H00) at the end of the row.
The data displays correctly in the pre-viewer (with a non-print symbol at the end of the last field), but gives an error when I try to run the package.
Example:
€œdata€?~€?data€?~€?data€?{NULL}{CR}{LF}
I can process the row in c# using:
€œ
€?
as the row delimiter. What is the equivalent notation for SSIS?
Thanks,
View 6 Replies
View Related
May 6, 2008
Hi,
I have a package that reads a fixed width flat file data into a single CHAR column in an sql server 2005 table.
In the "Columns" tab of the Flat File Connection Manager, I set the RowDelimiter value to {LF} and nothing for the ColumnDelimiter (since I read the entire row from the flat file into a single column in my sql table).
However, in the "Advanced" tab, the ColumnDelimiter Misc property shows {LF}. This was working fine for me.
The problem I was facing was with some files which were recently identified to have rows containing a special character (probably ASCII ZERO) in the middle of the row. So, now if the record was having 400 characters and the 200 th character had this special character, the package was writing the first 200 characters into the sql table and ignoring the rest of the characters.
I am sure that the special character was ASCII ZERO - I wrote a script to read each character in the line and find the ascii code for it.
Has anyone faced this problem ever. If so, pls let me know your solution or any ideas that can help sort this problem. Your help would be much appreciated.
Thanks!!
View 5 Replies
View Related
Sep 6, 2007
Hi everyone,
There is a small problem encountered while creating a package in sql
server 2005.
Actually i am using a flat file which has 820 rows and 2 columns which
are seperated by line feed(for ROW) and tab(for COLUMN).after
importing i found that ther are only 800 rows imported into the table.
Ather verifying the input file i found out that there are some null
values in the second column so there is no line feed for those
values.
Can anyone please help me how to give multiple delimiters for the same
input flat file.
View 9 Replies
View Related
Jun 5, 2007
I have a text file that I'd like to import into a SQL 2005 table. The file is tab delimited, which is easy enough to import, but I'd like the final field broken into multiple fields as well. The final field is space delimited. I've had no luck at being able to get this done. Has anyone done this?
View 5 Replies
View Related
Jul 23, 2005
I want to use DTS to load from a flat file. But I can't be sure whenthe flat file will arrive. I would like the job to check every 10minutes or so for a couple of hours. If the file is present, then loadthe file and rename it so that it doesn't get overwritten. Anysuggestions?TIA,amorphous999
View 3 Replies
View Related
Dec 27, 2006
Hi,
I have a situation where a tab limited text file is used to populate a sql server table.
The tab limited text file comes from a third party vendor. There are fixed number of columns we need to export to the sql server table. However the third party may add colums in the text file. Whenenver the text file has an added column (which we dont need to import) the build fails since the flat file connection manager does not create the metadata for it again. The problem goes away where I press the button "Reset Columns" since it builds the metadata then. Since we need to build the tables everyday we cannot automate it using SSIS because the metadata does not change automatically. Is there a way out in SSIS?
View 5 Replies
View Related
May 11, 2006
I am transferring data from an OLEDB source to a Flat File Destination and I want the column width for all of the output columns to 30 (max width amongst the columns selected), but that is not refected in the Fixed Width Flat File that got created. The outputcolumnwidth seems to be the same as the inputcolumnwidth. Is there any other setting that I am possibly missing or is this a possible defect?
Any inputs will be appreciated.
M.Shah
View 3 Replies
View Related
Aug 29, 2007
While loading the flat file into sql server2005 using SSIS , i am getting error like:
[Source - 20070801PensionPayments_dat [1]] Error: The "output column "Status" (246)" failed because truncation occurred, and the truncation
row disposition on "output column "Status" (246)" specifies failure on truncation. A truncation error occurred on the specified object of
the specified component.
In the EM import wizard, i can see last column [][](blank space similar box type). because of that i unable to load data into table. The delimiter is fixed width.
Please help me - Thanks in Advance.
View 3 Replies
View Related
Mar 14, 2007
I have number of csv files in a folder, all of them with same columns, need to be merged into one table and imported to sql server.
-The first row of the csv file is a header.
-The csv files are updated everyday
-The destination table is replace by new table with new info in the csv.
-The new csv files can be created and old csv files may no longer exist, but we are only interested in information contain in current csv files in the folder.
-I need SSIS to combine all the csv files in the folder and merge into one table.
-Other issue is that the field names may change in csv, so can the SSIS package recognize the change in field name and made necessary change in destination table as well.
Any insight on this issues will be greatly appreciated?
View 1 Replies
View Related
Jan 4, 2006
Here's what I want to do -
Dynamically load a flat file from a dynamic source table-
The source table metadata is known via the SYSOBJECTS and SYSCOLUMNS tables- I can pull the column names, type and lengths from these tables based on the table name. (my goal is pretty simple- pull data from a table in our database, and save it down to a flat file)
Would this be enough to dynamically create the destination flat file? If so, how do I do it?
Thanks
-rob
View 3 Replies
View Related
May 2, 2007
I have a text file that loads just fine with DTS 2000, but in SSIS it does not seems to recognize the row delimiter. For example, most rows in the text file have 10 columns, but some have 8 columns. For those with 8 columns, the SSIS is adding the data from next row, not padding the columns with nulls. Please help...
Thanks in advance.
View 16 Replies
View Related
Apr 2, 2007
Hi Mates,
I am unable to fetch data from Source which is CSV file whose sample contents are given below:
-----------------------------------------------------------------------------------------------------------
"Fiscal year";"Posting period";"Company Code";"Functional area";"Profit Center";"Business area";"Group Account";"Posting Level";"Document Type";"Accounting Principle";"Cost Pool";"Financial Owner";"Period Value GC"
"2006";"12";"AR00";"A107";"P1131";"7200";"3410";"10";"18";"GAAP";"#";"#";"$ 0.00"
"2006";"12";"AR00";"A107";"P1131";"9600";"3410";"10";"18";"GAAP";"#";"#";"$ 0.00"
--------------------------------------------------------------------------------------------------------------
The error message that comes while I preview the source data:
Error:
"The Preview sample contains embedded text qualifiers. The flat file parser doesnot support embedding text qualifiers in data. Parsing columns that contain data with text qualifiers will fail at runtime"
Is there any alternative to remove these text qualifiers from the file. Do we have any utility that can convert these files into text qualifier free CSV file?
View 3 Replies
View Related
Mar 27, 2015
Designing a solution for loading data into SQL destination from a single 5/10 GB flat file? If yes, what kind of performance measures you have taken while designing the solution ?
View 3 Replies
View Related
Nov 17, 2007
I tried to load a fixed width flat file with around 300,000 rows. However, only the first 8xxxx rows were loaded to the destineation table and the rest row were loading blank records. There was no error message showing during package execution. I've tried to split the file in half and the result was the same. So it wasn't the data file problem.
Would there be any buffering issue I need to cater for inside the package? Thanks!
View 10 Replies
View Related
Oct 30, 2007
Iam trying to load a RR Flat File(Ragged Right) into SQL Server DB Destination.
The file feeds the data into two tables and the logic looks like this..
If Position 30 = ''
Record Set goes into Table 1
Else
Record Set goes into Table 2
I have no clue where to start with my limited SSIS Exposure.. Any Ideas as to what i shud be starting with (Transformations)??
Thanks in Advance.
View 1 Replies
View Related
Apr 24, 2007
It seems I can do the type conversion in two place:
1. The advanced editor for flat file source => Input and output roperties. And I can set the output columns data type properties there.
2. Using a Data conversion box in Data flow transformation.
Any difference between these two? Which one should be used?
View 1 Replies
View Related
Feb 12, 2001
I am familiar with the MySQL Load Data command to load an external ascii file into a database table, but am having trouble finding a T-Sql command that is equivalent without creating an executable...any help would be appreciated...
View 1 Replies
View Related
Jan 31, 2008
What is the easiest way to accomplish this task with SSIS?
Basically I have a stored procedure that unions multiple queries between databases. I need to be able to export this to a text file on a daily basis and add a total records: row to the end of the text file.
Thanks in advance for any help.
View 7 Replies
View Related
May 13, 2008
Hello Experts,
I am createing one task (user control) in SSIS. I have property grid in my GUI and 2 buttons (OK & Cancle).
PropertyGrid has Properties like SourceConnection, OutputConnection etc....right now I am able to populate Connections in list box next to Source and Output Property.
Now my question to you guys is depending on Source Connection it should read that text file associated with connection manager. After validation it should pick header (first line of text file bases on record type) and write it into new file when task is executed. I have following code for your reference. Please let me know I am going in right direction or not..
What should go here ?
->Under Class A
public override DTSExecResult Execute(Connections connections, VariableDispenser variableDispenser, IDTSComponentEvents componentEvents, IDTSLogging log, object transaction)
{
//Some code to read file and write it into new file
return DTSExecResult.Success;
}
public const string Property_Task = "CustomErrorControl";
public const string Property_SourceConnection = "SourceConnection";
public void LoadFromXML(XmlElement node, IDTSInfoEvents infoEvents)
{
if (node.Name != Property_Task)
{
throw new Exception(String.Format("Invalid task element '{0}' in LoadFromXML.", node.Name));
}
else
{
try
{
_sourceConnectionId = node.Attributes.GetNamedItem(Property_SourceConnection).Value;
}
catch (Exception ex)
{
infoEvents.FireError(0, "LoadFromXML", ex.Message, "", 0);
}
}
}
public void SaveToXML(XmlDocument doc, IDTSInfoEvents infoEvents)
{
try
{
// // Create Task Element
XmlElement taskElement = doc.CreateElement("", Property_Task, "");
doc.AppendChild(taskElement);
// // Save source FileConnection
XmlAttribute sourcefileAttribute = doc.CreateAttribute(Property_SourceConnection);
sourcefileAttribute.Value = _sourceConnectionId;
taskElement.Attributes.Append(sourcefileAttribute);
}
catch (Exception ex)
{
infoEvents.FireError(0, "SaveXML", ex.Message, "", 0);
}
}
In UI Class there is OK Click event.
private void btnOK_Click(object sender, EventArgs e)
{
try
{
_taskHost.Properties[CustomErrorControl.Property_SourceConnection].SetValue(_taskHost, propertyGrid1.Text);
btnOK.DialogResult = DialogResult.OK;
}
catch (Exception ex)
{
Console.WriteLine(ex);
}
#endregion
}
View 10 Replies
View Related
Apr 24, 2006
I am trying to load 14+ million rows from a text file into local Sql Server. I tried using Sql Server destination because it seemed to be faster, but after about 5 million rows it would always fail. See various errors below which I received while trying different variations of FirstRow/LastRow, Timeout, Table lock etc. After spending two days trying to get it to work, I switched to OLE DB Destination and it worked fine. I would like to get the Sql Server Destination working because it seems much faster, but the error messages aren't much help. Any ideas on how to fix?
Also, when I wanted to try just loading a small sample by specifying first row/last row, it would get to the upper limit and then picked up speed and looked like it kept on reading rows of the source file until it failed. I expected it to just reach the limit I set and then stop processing.
[SS_DST tlkpDNBGlobal [41234]] Error: An OLE DB error has occurred. Error code: 0x80040E14. An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Cannot fetch a row from OLE DB provider "BULK" for linked server "(null)".". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "The OLE DB provider "BULK" for linked server "(null)" reported an error. The provider did not give any information about the error.". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Reading from DTS buffer timed out.".
--------------------------------
[SS_DST tlkpDNBGlobal [41234]] Error: The attempt to send a row to SQL Server failed with error code 0x80004005.
[DTS.Pipeline] Error: The ProcessInput method on component "SS_DST tlkpDNBGlobal" (41234) failed with error code 0xC02020C7. The identified component returned an error from the ProcessInput method. The error is specific to the component, but the error is fatal and will cause the Data Flow task to stop running.
...
[FF_SRC DNBGlobal [6899]] Error: The attempt to add a row to the Data Flow task buffer failed with error code 0xC0047020.
[DTS.Pipeline] Error: The PrimeOutput method on component "FF_SRC DNBGlobal" (6899) returned error code 0xC02020C4. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.
[DTS.Pipeline] Error: Thread "WorkThread1" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.
-------
After first row/last row (from 1 to 1000000) limit is reached:
[SS_DST tlkpDNBGlobal [41234]] Error: An OLE DB error has occurred. Error code: 0x80040E14. An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Cannot fetch a row from OLE DB provider "BULK" for linked server "(null)".". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "The OLE DB provider "BULK" for linked server "(null)" reported an error. The provider did not give any information about the error.". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Reading from DTS buffer timed out.".
---------------
When trying to do a MaximumCommit = 1000000. Runs up to 1000000 OK then slows down and then error.
[SS_DST tlkpDNBGlobal [41234]] Error: Unable to prepare the SSIS bulk insert for data insertion.
[DTS.Pipeline] Error: The PrimeOutput method on component "FF_SRC DNBGlobal" (6899) returned error code 0xC02020C4. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.
----
When attempting all in a single batch:
[OLE_DST tlkpDNBGlobal [57133]] Error: An OLE DB error has occurred. Error code: 0x80004005. An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80004005 Description: "The transaction log for database 'tempdb' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80004005 Description: "Could not allocate space for object 'dbo.SORT temporary run storage: 156362715561984' in database 'tempdb' because the 'PRIMARY' filegroup is full. Create disk space by deleting unneeded files, dropping objects in the filegroup, adding additional files to the filegroup, or setting autogrowth on for existing files in the filegroup.".
View 11 Replies
View Related
Jul 11, 2006
W2k3 server, SQL 2005.
@@version = Microsoft SQL Server 2005 - 9.00.1399.06 (Intel X86)
Standard Edition on Windows NT 5.2
(Build 3790: Service Pack 1)
I'm trying to set up my first SSIS package to import a flat file to a
database table. The connection manager will not allow a text
qualifier.
My data is comma delimited, with doublequotes around each field. If I
enter " as the text qualifier, then I get the error "The preview sample
contains embedded text qualifiers ("). The flat file parser does not
support embedding text qualifiers in data. Parsing columns that
contain data with text qualifiers will fail at run time." The data in
Columns view shows as just one column rather than multiple columns.
If I leave the text qualifier as None, then the data in Columns view
shows as just one column again.
I've tried with more than one datafile, including a short sample file
that I set up manually. All files get the same error.
Has anyone else seen this problem?
Thanks much for any assistance!
View 14 Replies
View Related
Apr 8, 2008
Each month we process 100+ text files into our ERP system. Occasionaly the Voucher Date in the text file does not contain a valid date. I would like to check to see if it is a valid Date, if it isn't replace it with the current date.
I thought I would use a Derived Column transformation, but I don't know how to check a field to see if it is a valid date.
Can anyone help?
Thanks,
Jeff
View 3 Replies
View Related
May 29, 2015
I am looking for a way to convert the following format into a sql table. The format it is Bib Tex.
Essentially a new row in the table would be for each entry, denoted by an @ logo and each column is denoted by an =, as you can see from the example data no one contains all the possible columns and some fields can be over two lines long.
To load this I was considering loading it into a table as each line being a row. Adding a row number, then a column counting the @ signs in order and essentially grouping each record, then for each group running through and looking for the column keywords 'author' , 'title' etc then splitting the data out into those constituent parts using substring and charindex.
@Book{hicks2001,
author = "von Hicks, III, Michael",
title = "Design of a Carbon Fiber Composite Grid Structure for the GLAST
Spacecraft Using a Novel Manufacturing Technique",
publisher = "Stanford Press",
year = 2001,
[Code] ....
View 9 Replies
View Related
Nov 5, 2014
I'm trying to parse out a line of data that is separated by the text "atc1.", "atc2." etc.
For example,
[atc1.123/atc2.456/atc3.789/atc4.xyz/]
If I only want the data after atc2., then I could search the string for "atc2." and collect all the characters afterwards. But how can I make sure to trim off all the data after "atc3." to make sure I'm only collecting "456" from the example above?
View 2 Replies
View Related
Mar 25, 2008
Hi, I need to export data from my sql server to a flat text file.
using select I only have the options of delimmited or column aligned, in either case the output file is unicode encoded. I need an ASNI encoded file.
I can create a non Unicode file if I use Import/Export but then the file does not have CR/LF or enf of record mark which I need.
Is there a way to export a flat text file without unicodes and with CR/LF or enf of record mark ?
I was looking into bcp, but I can't seem to find a good sample for this.
thank you
Alan
View 11 Replies
View Related
Mar 24, 2008
I sql server 2005, I need to export to a fixed width text file.
I can do this using import/export wizard, but my output file does not a CR/LF or end of record marker. There is no option to add this. my main reason for using the wizard is because I need to export the file without unicodes, and I can't seam to find any other way. any options?
thanks!
Alan
View 6 Replies
View Related
Jul 11, 2007
I have a Foreach loop which scans a table, and gets names of a bunch of procedures, and then back in the foreach loop, they get executed. Im trying to figure out how I can create a sort of log file to say the name of the procedure that is getting executed currently and the current date time stamp onto a flat file. I havent been able to figure this out yet..anyone know how to do this? I grab the names of the storedprocedures from the table and store it in a variable and use the name from the variable to actually execute the stored procedure.
I guess in essence, the question is how do i directly write lines of 'text' (from say a variable) into a flat file.
View 6 Replies
View Related
Dec 26, 2006
How do I tranfer a flat file (text) to SQL Server using DTS? Please point me in the right direction.
View 19 Replies
View Related
Jul 31, 2007
Hi everyone,
I am using SSIS, and I got the folowing error, I am loading several CSV files in a OLE DB, Becasuse the file is finishing and the tak dont realize of the anormal termination, making an overflow.
So basically what i want is to control the anormal ending of the csv file.
please can anyone help me ???
I am getting the following error after replacing the '""' with '|'.
The replacng is done becasue some text sting contains "" wherein the DFT was throwing an error as " The column delimiter could not foun".
[Flat File Source [8885]] Error: The column data for column "CountryId" overflowed the disk I/O buffer.
[Flat File Source [8885]] Error: An error occurred while skipping data rows.
[DTS.Pipeline] Error: The PrimeOutput method on component "Flat File Source" (8885) returned error code 0xC0202091. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.
[DTS.Pipeline] Error: Thread "SourceThread0" has exited with error code 0xC0047038.
[DTS.Pipeline] Error: Thread "WorkThread0" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.
[DTS.Pipeline] Error: Thread "WorkThread0" has exited with error code 0xC0047039.
[DTS.Pipeline] Information: Post Execute phase is beginning.
apprecite for immediate response.
Thanks in advance,
Anand
View 1 Replies
View Related