I was wondering if there is a way to schedule a tast that will dump a
fixed width text file of all the new entries in a table. So if I had
a table with like
username - varchar(20)
created - smalldatetime
I could get a weekly feed each week of all the new users in a text
file. I know I could write a script that would go through and do this
by looking at the time stamp, and the last time that the file
previously ran and get the new dates but I was hoping there was a
built way to do this. Or perhaps a more elegant solution.
My site is hosted with 1and1. They tell me I need to dump my SQL database before uploading it to their site. I have never done this before. I have looked in the online help and have asked them for help, but nothing seems to work. I've tried queries like the following:DUMP DATABASE dbname TO DISK = 'C:folderfilename'It tells me it doesn't recognize "dbname", which is the name of the MDF file (without the ".mdf" extension). So I'm not sure what I should be using for database name.I'm using Visual Studio 2005 Professional. Can someone help me with dumping my database? Thanks!
In an effort to automate a process, I am trying to populate a csv textfile with data from a SQL Server 2000 database that will be imported by a proprietary database; however, not all of the data required to go into the textfile is available in the source db. Fortunately, the data I'm needing has constant values for the fields that I want to populate, i.e Lab Name. Whereas, the Destination database will receive data from other labs but not via this source. Is it possible to use a constant rather than a db field within the SQL query to populate one of the textfile fields. (I placed "LABNAME" in where I would like it to go) A portion of my present SQL statement is: Select LEFT([SAMPNAME], 4) AS IUNUM, RIGHT(LEFT([SAMPNAME], 8), 3) AS SITENUM, convert(varchar,[SAMPDATE], 112) as SAMPDATE, [BDL] AS "SAMPNUM", [ANALYTE], (CASE [STARTDATE]-[SAMPDATE] WHEN 0 THEN '2' ELSE '1' END) AS METHOD, convert(varchar,[STARTDATE], 112) as STARTDATE, [FINAL], [BDL], "LABNAME", [NOTES1], [SAMPLER], [ORDNO], [UNITS] From [CUSTOMER] As you can see, I have already done a lot of formatting within the statement but would appreciate someone's SQL expertise to tell me if using a constant is possible or not. Thanks, Al
Hi, is there a software that can dump the SQL result into a file or to another database? I am using the MSSQL Server and I am accessing it remotely. Is there a program where I can issue an SQL statement and then I can dump the result into a file (in SQL inserts or just plain CSV).
Hey Guys, I do alot of development on my personal machine and have quite few databases in SQL Server. My PC is very prone to crashing :-p so Im writing a custom batch script to backup up all my data. How can I backup my SQL Server databses into a text file. I dont want to use the BAK file. In MYSQL this is a relatively simple process where i just have to use the mysqldump utility which scripts 'all' my databases (structure and data). Can I do the same thing in SQL Server? Please tell me a way. Its going to save me a lot of trouble. Thanks in advance guys.
Hi - is it possible to dump the query text of every query runagainst a particular database, or all databases, in an SQL Server2000 installation? We've an issue whereby a 3rd. party applicationrunning queries over ODBC sometimes claims that gobbledegook isreturned, and it would be very useful to identify the query that'sbeing claimed is the problem. I suspect that it's an applicationissue rather than a SQL Server one.__________________________________________________ ___________Are you Catholic ?http://www.CatholicEmail.com100s of FREE email addresses --->http://www.UltimateEmail.comSend an Online Greeting Card http://www.UltimateEcards.com
how do i do in ssis if i want to read the textfile row by row? i need to do this since every line in my flat file has different conditions to fulfill depending on the content of the row.
Hello, I'm having the following situation. To exchange data between software-programs we have to use textfiles. These files are like: "2100880000095500400600000329000 00000329000 WNOW0121102B 1121". I have to cut the textfiles into lines of 256 characters. Next thing i have to do is cutting that line in chunks (by a defined structure) and use these chunks to insert them into colums in SQL Server. Example, the line above will become something like: "INSERT INTO table (21, 008800000955, '004006', etc, etc) ". The code now is: i've opened a connection, and then I'm doing an INSERT per row, but when i have to insert 800.000 rows like this, it's a real bottleneck. Is there any way to perform the INSERTS quicker ?? By a bulk or something ?? Can anyone help me ??
Using SQL version 6.5, I'm trying to dump our primary database to a backup server with the following command:
DUMP DATABASE TMSEPRD TO Backup_Server WITH STATS = 10, NOINIT, NOSKIP
I created a backup device called Backup_Sever with the location of stlcopadmbkackup. When I try the command I get the error
Can't open dump device 'stlcopadmbkackup', device error or device off line. Please consult the SQL Server error log for more details. (Message 3201)
and in the eventlog it indicates 'access denied' as the cause for the failure.
I've looked in the Books Online and searched the technet database to find an answer. The Books Online does say that in creating the backup device permissions are not checked. I've checked and double checked the permissions on STLCOPADMBK and there shouldn't be any reason why access is being denied.
I've installed the latest service releases for version 6.5 so that shouldn't be an issue.
I am having problems when trying to dump to a network share (NT Network) from SQL Server 6.5 and would appreciate any insight this group can lend.
To setup this dump, I followed the instructions provided in the SQL Server FAQ (www.swynk.com) and have been able to successfully complete this task on a different server (so I know I'm editing the registry correctly etc..etc..etc..)
none the less,
Here is an exerpt from NT's Event Viewer:
SQL Server Scheduled Task: 16, 'Network Dump - EIWDATA' -- Status: Failed -- Task Invoked on: 7/23/99 1:45:18 PM -- Message: Can't open dump device 'crosrvdumpeiwdata.dmp', device error or device off line. Please consult the SQL Server error log for more details. (Message 3201)
Hi:)I need to dump data from database in to *.sql file. I try it using DUMPDATABASE but it not work that I need it.So, how can I dump data from database in to *.sql file and then run itusing: osql -E -i nameOfDumpedDatabase.sqlThanks for help, greetings, gregory :)
Hi, I wish to export a MS SQL table into a text file. I know that I can do this under the DTS wizard but I need the syntax so as to run in on my script. Can anyone help? Thanks!
I ask for your help for problem of SQL Server 7 DTS . I got 6 text data files extracted from a SQL server database Sale1 in Company A. I also got a script produced from Sale1 and run it in my company to build a SQL server database A_Sale which structure is the same as Sale1. Then, I transferred these files into my database A_Sale. I have successfully transferred three files (vlet, vlrl, and vlcv), without any trouble. But I met trouble for transferring other three files (vlcl, vlco, and vlpo). For example, the error messages are as: "Error during Transformation 'DirectCopyXform' for Row number1. …Conversion invalid for datatype on column pair 12 (source column 'Col012'(DBTYPE_STR), destination column 'cl_dob' (DBTYPE_14))." "Error during Transformation 'DirectCopyXform' for Row number 10825. …Conversion invalid for datatype on column pair 17 (source column 'Col017'(DBTYPE_STR), destination column 'c0_stchdt" (DBTYPE_14))." - I don't understand why the datatype is not matched? The datatype is 'int' in both source column and destination column and the data for the source column is like 2533222. And, why the error caused for Row 1 for file vlcl and for Row 10825 for file vlco? How to correct the error? Other errors I met like: "Error at Source for Row number 24391. …Column Delimiter not found." - I don't know how to avoid the error at source? My partner in companyA also cannot control to avoid the error, since the file extracted by SQL server. "Error at Destination for Row number 531377. …Could not allocate space for object '(SYSTEM table id:-269896619)' in database 'TEMPDB' because the 'DEFAULT' filegroup is full." - Why the error caused for row 531377 and how to correct it?
The transaction log does not truncate despite all attempts (no_log,truncate_only, etc...) This database is a subscriber for replication. Tried the reconfigure with override. There are no open transactions (dbcc opentran)
While dumping several tables into the same Backup device, I got this error message in the Log :
dbsvolopen: Backup device 'E:BACKPANBackUp_Of_Focus_Tables.DAT' failed to open, operating system error = 32(The process cannot access the file because it is being used by another process.)
There is no other process using this device running at the same time.
DataBase i am using is Sql Server6.5. In a trigger i had written code to transfer updated records from one table to other table.These updated records needs to be written into a text file. I had used xp_cmdshell but it is taking time.Is there a way to write data to flat file.
when i run this it says "The command(s) completed successfully." but i don't see that file anywhere at all. why is that? is this the right way to do? please help.
i have bulk data that i dont want to have to enter manually. How can i achieve this for sql server. I want to be able to load from a text file (or any text format) with my data separated by delimeters. I know oracle has something that does this called sql loader.
I have some "phantom" tasks running, specifically, DB dumps to a NULL Pipe device and they are not scheduled!!! These do not show up in the SYSTASKS table, where can I look to possibly find these un-scheduled dumps? Thank you....
I have a report that's created each day as a flat textfile.Because I came from the Access world, I created a macro that importsit with a schema that gives meaningful names to the various columns,and then uses a query to massage some of the data for me (deletes thefirst blank row and does a couple of calculations)Then I use DTS to import the Access query as a table.the textfile has a column called "File_num", and among several others,a column called "Serial_num". (the file numbers represent shipments,and sometimes there are more than one serial number in the shipment,etc., so there is a separate line for every serial number)Naturally, I would like to split this info into two tables..one thatdoes not contain the serial numbers and has a primary key on the"File_num" column, and another table that would contain just the"File_num" and "Serial_num" columns. That way I could relate themlater...but most importantly, it will give me a table where I can usethe "File_num" as my primary key.What would be the best way to import these two tables from one sourcetextfile? The other thing that gives me problems is that the textfile has no column names, and the first row is always blank.I'm very new to SQL and DTS and would appreciate any direction.Thanks,Larry- - - - - - - - - - - - - - - - - -"Forget it, Jake. It's Chinatown."
On a 64bit Itanium cluster, I'm having a problem dynamically setting the ConnectionString of a TextFile Source using the "EvaluateAsExpression" = True and setting the ConnectionString using an expression from a variable. I get the error message:
-1073647612,0x,The result of the expression "@[User::vFullFilePath]" on property "ConnectionString" cannot be written to the property. The expression was evaluated, but cannot be set on the property.
This is not a problem at all on a 32bit machine. I can run the same package on 32bit and it runs fine.
The funny thing is, I'm able to dynamically change the ConnectionString of a Logfile that we're writing all of the errors to, which I believe is using a TextFile Destination connection under the covers somehow (I'm using the SSIS Logging, to a textfile).
Can somebody please help me with the following problem.
I want to import data from a textfile called "Links.txt" into a SQL-server database called "LinkData". The data in this textfile is separated by pipelines. And this import should be done every 2 hours automatically. How can I make a proces or something in SQL server 7 that will fill the database with the data out of this textfile every 2 hours. Please help me. I really don't know.
I want to import data from a textfile called "Links.txt" into a SQL-server database called "LinkData". The data in this textfile is separated by pipelines. And this import should be done every 2 hours automatically. How can I make a proces or something in SQL server 7 that will fill the database with the data out of this textfile every 2 hours. Please help me. I really don't know.
I was planning on using DTS for this Only I really don't know how. I tried it with "Data driven query task" in combination with "Text file (source)" and "Microsoft OLE DB Provider for SQL Server". But when I run it, it gives an error. What am i doing wrong??? Help me please
I have an existing table I need to add data to. The data is in a text file, and the existing table already has data in it (I don't want to delete this I want to add to it).
I used Microsoft's import utility but this created a seperate table with generic fieldnames (column01, column02, ect). Is there a step in this wizard I missed?
While working with a vast variety of support projects, i find a sql design where all the fields in a single form (say about 100 fields which are dump data as they are not related to any reports and searching criteria) are dumped in a sql database column in a XML format. See below an example
Hello there guyz!ahm, i just want to ask, how can i upload a MS word text file to a SQL Dbase? I know i must use the Input File, the one in the Toolbox menu for HTML controls. but i need to know 1. How can i link the SQL dbase to the Input file Button? 2. How to put the MS word file onto the SQL dbase? 3. what should i use to declare the MS Word file on the SQL Dbase? is it variable? char? etc... THank you so much! Please i Really need this one to put up my thesis system. and if it isn't much of a task, please can anyone send me the SOURCE CODES for my questions?
I would like to use logging to be able to view information about the package execution afterwards, especially to be able to find out which task that failed and why it failed.
Something similar to this:
----------------------------------------------------------------------------- The execution of the following DTS Package succeeded:
Step 'DTSStep_DTSActiveScriptTask_2' was not executed
Step 'DTSStep_DTSExecuteSQLTask_33' failed
Step Error Source: Microsoft OLE DB Provider for SQL Server Step Error Description:The statement has been terminated. (Microsoft OLE DB Provider for SQL Server (80040e2f): Cannot insert the value NULL into column 'ID_adress', table 'VPKBA.dbo.aktAdress'; column does not allow nulls. INSERT fails.) Step Error code: 80040E2F Step Error Help File: Step Error Help Context ID:0
I am trying to get a result like below without using the cursor.
col1 col2 col3 col4
1111 date uniquenumber 6 1111 date uniquenumber 5 1111 date uniquenumber 4 1111 date uniquenumber 3 1111 date uniquenumber 2 1111 date uniquenumber 1 2222 date uniquenumber 4 2222 date uniquenumber 3 2222 date uniquenumber 2 2222 date uniquenumber 1 3333 date uniquenumber 2 3333 date uniquenumber 1
the column that say unique number is unique and is not duplicated and date column might have duplicates
Please advise whether it is possible to write a query without cursor.
I am looking to create a incremental value based on the resulting insert that I am using. There already is another field being used as an identity field. I have a beginning value that I just want to add the row number to for the insert.
insert into lineitem select substring(group_id,4,len(ltrim(rtrim(group_id)))-3) as co_code, 0,0,(case when enddate < cast(month(getdate()) as varchar(10))+cast(day(getdate()) as varchar(10)) then 'Prior' else 'Current' end ), left(acct_type,2) as bene_type, convert(smalldatetime,left(ltrim(rtrim(eff_date)), 8)), 0,trans_amt,0,0,convert(smalldatetime,left(ltrim(r trim(sett_date)),8)), ltrim(rtrim(b.fname)) + ' ' + ltrim(rtrim(b.lname)) as payee,0,0,a.ssn,'Y',999+count(*), (case when isnull(b.location,'') = '' then '' else b.location end) as location from mbi_tran_temp a left join enrollees b on a.ssn = b.ssn and ltrim(rtrim(a.group_id)) = ltrim(rtrim(b.mbicode))
The '999+count(*)' is where I would like to have the incremental value.