Importing Data On Regular Schedule

Oct 14, 2005

Help needed please from the SQL team,

I need to import data from one table on to another on a regular basis (twice a day). Sorta like...

From:
Server1
Db1
Table1

To:
Server2
Db1
Table1

I used DTS to get the data initially from Server1 but how I can do it on a automated basis?

Thanks.

View 1 Replies


ADVERTISEMENT

Importing A Csv File On A Schedule.

May 6, 2004

I am trying to import some data from a csv file on a regular basis. This file, however, is stored on a website. I can't seem to find a way, either using DTS or a stored procedure to make this task simple. If the file is at http://www.mysite.com/import.csv is there anyway I can get DTS to download that file and import it into a new table?

Thanks,
Scott

View 3 Replies View Related

How To Convert To Regular Text, Data Stored In Image Data Type Field ????

Jul 20, 2005

Hi,This is driving me nuts, I have a table that stores notes regarding anoperation in an IMAGE data type field in MS SQL Server 2000.I can read and write no problem using Access using the StrConv function andI can Update the field correctly in T-SQL using:DECLARE @ptrval varbinary(16)SELECT @ptrval = TEXTPTR(BITS_data)FROM mytable_BINARY WHERE ID = 'RB215'WRITETEXT OPERATION_BINARY.BITS @ptrval 'My notes for this operation'However, I just can not seem to be able to convert back to text theinformation once it is stored using T-SQL.My selects keep returning bin data.How to do this! Thanks for your help.SD

View 1 Replies View Related

How To Get Data From A Excel File To A Sql Database On A Regular Basis

May 12, 2008

Hello All,i have 2 excel files which i want to upload to sql database. i know how to upload to the folder. my question how to i use these excel files and convert them into  tables in sql database.i have to do this on a regular basis. is there anyway i can do this. the 2 files when converted to tables, i need to create relationships etc. any suggestions or references please. i appreciate it. Thanks 

View 2 Replies View Related

Transact SQL :: Using Regular Expression To Find And Replace Data

Sep 9, 2015

I am writing an SQL query to find an replace data in a column. I have a table that is filled with

C:usersXXXXappdata
C:usersYYYYappda
C:usersZZZZZZappdata

I would like to replace the c:usersXXXXX part with %userprofile%

The end result would be %userprofile%appdata

I know how to dot the replacement in powershell. it's quite easy

-replace "c:users[^]+","%userprofile%"  

Basically how would transfer this into SQL...

View 10 Replies View Related

Reporting Services :: Data Present When Run On Report Server URL But Not On Regular URL

Nov 10, 2015

Was reported to me that reports were running but data was not being displayed. After checknig the Config manager I found that if I went to the reportserver url then clicked on a report it would diplay he correct data. But when trying it on the normal website it would not display anything.  

sql server 2008 r2
ver: 10.50.2550.0

View 5 Replies View Related

Using A Regular Expression In MSSQL 2005 To Find Patterns Buried In A Field Full Of Misc Data

Feb 29, 2008

Thanks in advanace for taking the time to read this post.
I am using MSSQL 2005 and have created a function that allows me to use regular expressions in my SQL queries.
My question is I have a pattern buried in a field of misc data that I need to pull out just that pattern and discard the rest of the data.  Here is the Regular Expression I am using
select field1 from table1
where dbo.RegExMatch (field1,'[a-zA-Z]{4}[0-9]{6}[a-zA-Z]{2,4}')=1
This returns all values in the field that match the expression.  What I want to do now is remove all data from the field on the left and right of the expression that does not match the expression.  How would I accomplish this without reading through the 200k+ records and writing rules for every exception I run across?
so I could have Gar  b/a ge 'THE GOOD DATA' m/or1   ba4d da....ta.  All I want to do is return 'THE GOOD DATA'

View 4 Replies View Related

Attempt FASTLOAD If Error Perform Regular Data Flow Transfer And Redirect Errors To Text File

Feb 15, 2008



I'm having problems designing a package to attempt to execute a fast load data transfer but failback to regular speed with error redirection in the event of an error.

The way I designed this was to add one data flow task to my package called "DFT FASTLOAD". The data flow copies a table SRC to another table DEST in the same SQL Server database. In the error handler for the data flow task I copied the original data flow task and changed the name to "DFT REGULARLOAD with Error redirection". In this data flow task I did not use fast load and addtionally redirected errors to a text file.

In the Data Flow Task "DFT FASTLOAD". I am copying from a varchar source field(with non-date strings) to a datetime destination field to force errors. However the Data Flow Task "DFT REGULARLOAD with Error redirection" never seems to start transferring data from source to destination. The data Flow Task "DFT REGULARLOAD with Error redirection" turns yellow (after the error occurs in "DFT FASTLOAD"), but no data is being transferred). It seems like it hangs.

Do I need to increase the MaximumError Count or something? The data flow task "DFT FASTLOAD" does not turn red when the error occurs it just remains yellow, so i assume I'm on the right track since it seems the error is caught.

I have added screenshots ... hopefully these screenshots will clarify my problem.

DESIGN:
http://i256.photobucket.com/albums/hh179/abzbank/DESIGN_FASTLOAD1.jpg
http://i256.photobucket.com/albums/hh179/abzbank/DESIGN_FASTLOAD2.jpg
http://i256.photobucket.com/albums/hh179/abzbank/DESIGN_FASTLOAD3.jpg
http://i256.photobucket.com/albums/hh179/abzbank/DESIGN_FASTLOAD4.jpg
http://i256.photobucket.com/albums/hh179/abzbank/DESIGN_FASTLOAD5.jpg
http://i256.photobucket.com/albums/hh179/abzbank/DESIGN_FASTLOAD6.jpg


RUNTIME:
http://i256.photobucket.com/albums/hh179/abzbank/RUN_FASTLOAD7.jpg
http://i256.photobucket.com/albums/hh179/abzbank/RUN_FASTLOAD8.jpg
http://i256.photobucket.com/albums/hh179/abzbank/RUN_FASTLOAD9.jpg
http://i256.photobucket.com/albums/hh179/abzbank/RUN_FASTLOAD10.jpg
http://i256.photobucket.com/albums/hh179/abzbank/RUN_FASTLOAD11.jpg


I can provide more details if needed... but really this is just a basic test.
Any assistance would be appreciated!

View 9 Replies View Related

MSSQL 2005 Inserting Data On Tables Created By Regular Users With Ddl_admin Role, Using Access Or Other Front End Apps

Feb 14, 2008

Writing to tables created by regular users on MSSQL2005

I have users creating tables through an application, I gave them ddl_admin, datareader, datawriter. They can create tables but cannot insert/update data (to their own tables), I cannot insert data either using Access or any other application to those tables created by them (under dbo schema) Is there something I am missing with permissions? Thank you very much

View 3 Replies View Related

How To Schedule An Export Of Data To Xml

Feb 8, 2006

Hi gurus,
I need to be able to export a table to an xml file on the 1st of every month. How can I accomplish this using SQL.
 
Thanks!

View 1 Replies View Related

Schedule Backup Of Data/log Files

Sep 5, 2006

Is there an equivelant to the SQL Server Agent in SQL server express edition? If not, how can I schedule the backup of the data/log files?

Thanks,

View 3 Replies View Related

Data From Temp Table Into Regular Table.

Feb 21, 2007

Hi everyone, I'm fairly new to sql and right now I am struggling with a script. I am trying to extract data from a normal table into a temporary table, update it in the temporary table, then put it back into the normal table. I'll display my code, let me know what you think, any suggestions are appreciated. Thanks a lot.


Create table scripts (
UserID int,
UserName char(50),
ScrRan char(50),
StartTime datetime default getdate(),
EndTime datetime);

Create table errors (
ID int,
UserName char(50),
UserLogin char(50),
ErrorNumber int,
Message char(100),
TimeOfError datetime default getdate());

declare @error int
declare @msg varchar(100)
declare @startTime datetime
declare @endTime datetime

select @startTime = getDate()

SELECT *
INTO #Temp
FROM Publisher
WHERE pub_Name = 'Scene Publishing'

UPDATE #Temp
SET pub_Name = UPPER(pub_Name)


SELECT *
INTO Publisher
FROM #Temp


--Begins Error Checking Routine
select @error = @@error

IF @error <> 0
BEGIN
select @msg ='error: ' + convert(varchar(7), @error) +
''
insert into errors values (@@SPID, USER, USER_NAME(), @error,
@msg, getDate())
END
ELSE
BEGIN
select @endTime = getDate()

insert into scripts values (@@SPID, SYSTEM_USER, @startTime, @endTime)
END
select * from errors
select * from scripts




lost and loaded.

View 2 Replies View Related

Integration Services :: SSIS - Managing Data Integrity When Importing Sharepoint Data

Sep 28, 2015

I setup this package to import data from a Sharepoint list to a SQL Server data table. The primary key of my SQL table is mapped to the Title column of my Sharepoint list. There is a possibility that duplicate values will be entered in the Title field of the Sharepoint list. So when importing data into my table via SSIS, my package always error-out when there it comes across duplicate values. how you others have managed data integrity when importing from a Sharepoint list with the Title column being mapped to the primary key of a table.

View 4 Replies View Related

Memo Data Type Import Error While Importing Data From Access File Into SQl Server 2005

Sep 10, 2007

I have one column in SQL Server 2005 of data type VARCHAR(4000).

I have imported sql Server 2005 database data into one mdb file.After importing a data into the mdb file, above column
data type converted into the memo type in the Access database.

now when I am trying to import a data from this MS Access File(db1.mdb) into the another SQL Server 2005 database, got the error of Unicode Converting a memo data type conversion in Export/Import data wizard.

Could you please let me know what is the reason?

I know that memo data type does not supported into the SQl Server 2005.

I am with SQL Server 2005 Standard Edition with SP2.

Please help me to understans this issue correctly?

View 4 Replies View Related

Data Access :: Importing Huge Data From One Database To Another Daily

Jul 7, 2015

We have a daily process, which copies millions of rows of data from one DB to another over Linked Server. Just checking on the best practise, are there more efficient ways than the Linked server to copy millions of rows of data from one DB to another? I checked bulk insert but that transfers the data from the file to DB not DB to DB. 

View 6 Replies View Related

Importing Data From Oracle To Sql Loosing Data After The Decimal Point

Jun 18, 2007

I have created a simple package that uses a sql command to pull data from an oracle database and inserts the data into a sql 2005 table. Some of the data fields that i am pulling from contain two digits after the decimal point, however this data is lost when it gets into sql. I have even tried putting the data into a flat file, and still the data is lost.

In the package I have a ole db source connection which is the oracle database and when i do the preview i see all the data I need. I am very confused and tried a number of things to get the data into sql, but none work. Any ideas would be very helpful.

thanks

View 6 Replies View Related

Data Format Issue While Importing Data From Excel To SQL

Jul 17, 2007

hi



when i m importing data from excel to Sql using DTS the column which has text content was not imported as same in excel sheet. whereas a special character is appearing in between the lines. the text field contains multiple lines but the conetent is imported in single line .

ex:









ARIZONA
ALABAMA
STATE


but i m getting imported

as :
ARIZONA ALABAMA STATE

How to Format a single column while importing?



Regards

Raj

View 1 Replies View Related

Importing Data From Various Data Sources With Non Standard Formats

Mar 19, 2007

Hi all :)

I'm wondering if SSIS will be the solution to the problem I'm working on.

Some of our customers give us an Excel sheet with data they want to insert or update in the database.

I've created a package that will take an Excel sheet, do some data conversion so the data types match up and after that I use a Slowly Changing Data component to create the insert/update commands.

This works great. If a customer adds a new row to the Excel sheet or updates an existing row changes are nicely reflected in the database.

But now I€™ve got the following problem. The column names and the order of the columns in the Excel sheet are not standard and in the future it could happen a customer doesn't even use an Excel sheet but something totally different.

Can I use SSIS for this? Is it possible to let the user set the mappings trough some sort of user interface? I€™ve looked at programmatically creating the package but I€™ve got to say that€™s quit hard to do€¦ It would be easier to write the whole thing myself than to create the package trough code ;)

If not I thought about transforming the data in code before I pass it on to the SSIS package in something like XML. That way I can use standard column names and data types.

So how should I solve this problem? Use SSIS or not?

Thnx :)

Wouter de Kort

View 6 Replies View Related

Power Pivot :: Schedule Cannot Be Enabled For A Workbook With No External Data Sources

Apr 22, 2015

I have uploaded excel report with data model ,  but when i click on ,anage data refresh link it is showing shedular disable. I have check logs then I am getting analytics server connection time out error.

Right now architecture of system  having different analytics server and sharepoint server is also different. Any better solution due to which i can able to schedule my reports.

View 2 Replies View Related

Power Point :: Schedule Cannot Be Enabled For A Workbook With No External Data Sources

Jul 1, 2015

I work almost 1 month to try to fix this issue, but still failed.(In my test server, it's working). seem like I have 2 problem.

1. My SQL server analysis service(PowerPivot) service mode is "Multidimensional", in my test server, this is SharePoint mode.

2. After upload PowerPivot worksheet, I check the analysis service, No this sheet DB save in. in "Analysis Server Properties", I add "sp_admin" name that I use this name to install PowerPivot add-in.

View 3 Replies View Related

How Can I Specify The Data Type When Importing Excel Data Via DTS?

Jun 11, 2006

I'm new to SQL and DTS packages. I am trying to import data from an excel spreadsheet to an SQL server table via DTS package. It seems that the excel task looks at the first few records in a column to determine the datatype for that column. If the first few records are text, the entire column is imported as text. If numeric, the entire column is imported as numeric.
There are about 25,000 records. In one field, the most important one, about half of the records begin with letters and the rest are all numbers. It is the subscriber ID field, and some subscriber IDs are all numbers, some are letters and numbers. The entire column should be imported as text. However, when I run the transform data task from the excel connection, none of the records that are all numbers are imported. I end up correctly importing only 13,000 of the 25,000 records. The rest are imported with the subscriberID field as <NULL>.
I tried using the CAST or CONVERT function in the SQL query, but get the error message "Undefined Function."

Can anyone give me some help? Thanks,
Jim

View 4 Replies View Related

How Can I Specify The Data Type When Importing Txt File Data Via DTS?

Jun 27, 2006

hello,
I create a txt file with a bash script, and i need to use it in a DTS package. But, i don't know how i can specify the type of my column. So in the transformations task, i have an error due to an incompatible type. what can i do to fix this error ?
thanks,

View 8 Replies View Related

Importing Unique Data && MAX Data To Table Using DTS

Nov 28, 2005

I am creating a DTS package that is combining several tables, converting one column of data to a new column removing all special characters, then exporting the unique data based on this column and another column, and the max of other duplicates to a new table.

Now that I have the data in this table, I want to import any data that is not in my main table.

This "CLEANED" table does not have a designated "key" column, but the table I want to import the unique items does have an ID column that is also a primary key column.

DTS seems to want me to have a Key column to reference when importing from the CLEANED table to the MAIN table.

How would I go about checking the MAIN table against the CLEANED table, having DTS import only the unique items from the CLEANED table that are not present in the MAIN table based on three columns? The rest of the columns I want to just extract the MAX data from the duplicates.

Now here is the query I use to extract the unique values from the "CLEANING" table to get the data to the "CLEANED" table, but do not know how to use this to import into the MAIN table using something similar.


Code:



select
partno2,
MAX (partno) as partno,
alt,
MAX (C_alt) as C_alt,
Max (cmpycd) as cmpycd,
MAX (type) as type,
compFN,
MAX (pndesc) as pndesc,
MAX (equipment) as equipment

into tbl_CLEANED
from tbl_CLEANING
group by partno2, alt, compFN
ORDER BY partno, compFN



The three main columns I need to check against are:
partno2
alt
compFN
I have named the columns the same in both tables.

partno2 is the column that has been copied from partno with all special characters & spaces removed. This is the main column I am using as a reference for unique values, then if no match, I have it check against the alt column, then the comFN column. If there are no matches in any of these columns, then I want to extract the data to the MAIN table.

How can I compare these tables and import only unique info to the MAIN table?

In addition, how can I also check items that are the same in both tables and update the MAX info for the other columns (not the three I use for reference - these I need to leave alone) and update those if there is more data in the CLEANED table then in the MAIN table?

View 3 Replies View Related

Trying To Run A Schedule Dts Package Every 20 Minutes To Recreate A Table With Updated Data Into A Excell File

Jun 7, 2007

The only way the job success is if I select the option add rows wich is copying all rows even the ones that are already there, I tried to drop amd recreate the table option but i just dosn't run that way, help please..

View 1 Replies View Related

Importing Data

Feb 12, 2005

I have created a DTS package for SQL Server, saved this as a VB file and upgraded this to the .NET framework.

It is now saved in a file called Shortages.vb.

What I want to do now is add this to an existing ASP.NET project and be able to call this DTS function by the click of a button.

Is this possible and how can I go about it? Can I just add it to the Click event of a button?

View 3 Replies View Related

Importing Data

Mar 17, 2002

Hallo every one,

I', working for a school project.
The setuation is an old database that has to be repaird.

All the data from each table of the old data base are stored in different tabdelimited text files.

I designed a new database, but the tables are a little different.
When I try to import the data I get some errors, because of the changed tables.

An example of my new setuation is a new table thas as colums from 2 different tabdelimited text files.

My question is, how do I solve thos problem!
Do I have to use DTS?
What do I have to do.

tnx
icheron.

View 1 Replies View Related

Importing Data

Jan 12, 2001

How can you can import results into table from system procedures.

For Ex. if you exec sp_spacesued, how do I import the results into the table.

Thank You

View 1 Replies View Related

Importing Data???

Sep 29, 2004

I have a process that calls a proc that BCP's a delimited file into a table. Well the SOX police say a header and footer must be added to the file. Needless to say this screws my BCP process. Does anyone know how to strip a header and footer record from a text file using transact sql or have any other suggestions to strip the records?

View 6 Replies View Related

Importing Data

Dec 4, 2006

I am trying to import data from excel into my server, but get this error message:

Error during transformation 'directcopyXform' for row number 1. Errors encountered so far in this task 1. TransformCopy 'DirectCopyXform' conversion error: Conversion invalid datatypes on coulumn pair 19 (source column '*9' (DBTYPE_WSTR( destination column 'F19' (DBTYPE_R8)

Anyone get me in the right direction?

Scott

View 6 Replies View Related

Importing Data

Apr 24, 2007

we are trying to import data from a flat file using an uptick (`) as a column separator and {CR/LF} as a record terminator. There is a variable number of columns for each record. The initial record in the flat file has 3 columns. Upon processing this record, the import sets all records to 3 columns and does not read the column separators past the second column (even though there may be up to 7 columns in the record).

This method worked ok in DTS2000 and it works with Excel. Any suggestions?



Thank You

View 3 Replies View Related

Importing Data Into SSE

Mar 21, 2007

There's alot of discussion in this Forum concerning the importation of data into SSE.

I recently discovered that you can quite easily export tables directly from MSAccess to SSE.  Simply 1) select the desired MSAccess table, 2) choose 'Export' from the file menu, 3)  in the 'Save as Type' drop down, select ODBC databases(), 4) at this point, an 'Export' dialog with the name of the selected table appears, 5) click 'OK',  6)  the Data Source Manager appears - go to 'Machine Data Source' and select a DSN that connects to SSE (you need to have set this DSN up before), 7)  click 'OK' and  and your table will be exported from MSAccess to SSE.

For whatever reason, the table you exported will always end up in SSEs 'Master' database tables.  There is probably some work-around to correct this, but I havnt spent much time fooling with it.

Now, MSAccess has a very good parser for importing several types of external data, including Excel worksheets, CSV files, and text data.  However, to get a satisfactory into MSAccess, you may need to edit your data files.  Then, to import the data, from MSAccess select 'Get External Data' -> 'Import' from the File menu and just follow the instructions.

BTW, there is alot of confusing discussion concerning the ability of SSE to use DTS.  Some have said that it is not supported, only in the full-blown version.  Others have suggested you can download the DTS application from Microsoft and use it with SSE although it does not show up in the SSEMS directory tree.

I downloaded the file 'SQLServer2005_DTS.msi' and tried to install it.  It appeared to install OK but I cannot find it anywhere on my machine. Weird, eh.

While most of this thread is in the way of a comment, I have a couple of questions:

1 - Is there some way to connect a machine DSN to a database in SSE that is other than 'Master' ?

2 - Are there better ways to import data into SSE ?

3 - Is there some way to move or copy a table from one SSE database to another ?

BTW, please do not respond to this thread by posting some link.  Many of the links I have attempted to follow from this Forum are either irrelevant to the posted problem or are no longer available.   If I wanted that kind of help, I could use Google or go to the library.  Try to answer the question(s) directly.  If you don't know the answer, say so, or at least, don't answer at all.

 

 

View 13 Replies View Related

Importing Data Into SQL

Aug 30, 2007

With SQL 2000 there was an Import/Export facility for importing data into a Sql database. Could somebody tell me how to import an old database which could be in Csv text or Paradox into my new SQL 2005 tables. Thanks

View 4 Replies View Related

Importing Old Data Into Asp_tables...?

Nov 9, 2007

I have created an application that uses the login, create, etc login components in .net. How hard is it to convert all my old users, passwords, usertypes into the new tables. It almost looks like I have to do them by hand and created a new guid(userid), along with the same guid in the aspnet_usersinroles and aspnet_Membership.  Is there a script to do this programatically?

View 1 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved