I want to compare the schemas of 2 different databases, I heard that there was a way to dump the contents of a database (tables, stored procedures, objects, etc) to a text file in SQL to use for an easy compare. Anyone know the process in which to do this? Any help is greatly appreciated.
I got SQL server 2000 running on Server 2003 Enterprise edition. And a dump file, which I have no idea how it has been created. All I know is that the data is some relational database.
I need to import this dump file into the SQL server database.
Hi, I just started using SQL Server 2005 and I'm trying to find out how to do a sql dump on a table, but this is proving more challenging then it should be.
I usually use mysql with a program called navicat, and all you do is right click on the table and select dump... Inserting the data back in is just as simple. I have also used sql server 2000 awhile back and I know there was a dumping utility for it.
Can someone point me in the right direction on how to dump data to a .sql file and reinsert that data? Thanks!
I need to dump a table with more than 200 columns to a flat file. I tired in SSIS with the flat file destination, but it only allows 84 columns. What should I do if I want to dump all the columns to a text or cvs file. Thanks in advance
I have several SSIS packages which are run in sequence. But some strange reason sometimes a SQL Dump is created. If I restore database and restart the same ETL is works, or it creates another SQL Dump in a different place. Any idea on how to debug this problem?
SQL Server 2005 Developer Edition SP1 ( should I upgrade to SP2? ) Windows 2003/r2 64b ( latest patches )
Microsoft Visual Studio 2005 Version 8.0.50727.762 (SP.050727-7600) Microsoft .NET Framework Version 2.0.50727
Installed Edition: IDE Standard Microsoft Visual Studio 2005 Premier Partner Edition - ENU Service Pack 1 (KB926601) This service pack is for Microsoft Visual Studio 2005 Premier Partner Edition - ENU. If you later install a more recent service pack, this service pack will be uninstalled automatically. For more information, visit http://support.microsoft.com/kb/926601
SQL Server Analysis Services Microsoft SQL Server Analysis Services Designer Version 9.00.2047.00 SQL Server Integration Services Microsoft SQL Server Integration Services Designer Version 9.00.2047.00 SQL Server Reporting Services Microsoft SQL Server Reporting Services Designers Version 9.00.2047.00
I'm using sqlserver 2000 enterprise edition. I am an oracle dba and we have some tables in sqlserver 2000 that we need to write out to the flat file. I have a procedure in oracle to do this for oracle tables. But, how would I do this in sqlserver 2000. I have 10 columns on this table and I only want 3 columns data to be dumped on the flat file. We are on NT sever 4.0.
Hi, I would preciated if some could help me with this problem. My have used Mysql database and now i have to change my database to sql server. I have made a dump file from mysql db and i would like to insert that dumb file to sql server. Have any idea how it is possible?
I am trying to export a table with ~ 10 Million rows to a flat file and it is taking for ever with SQL2005 export functionality. I have tried creating an SSIS package with a flat-file destination and the results are the same. In each case it does the operation in chunks of about 9900+ rows, and each chunk takes ~1-2 minutes which sounds unreasonable.
I tried bcp, and it fails after a few thousand rows. I tried moving the data to SQL2000 first then to flat file from SQL2K, but the move from SQL2005->SQL2000 was going at the same rate as above.
So, the bottleneck seems to be data going out of SQL2005 no matter what the destination is. I'm wondering if there is some setting that Iam missing that would make this run in a reasonable amount of time?
In my script task I have the following code. The task I'm trying to accomplish is: If the filename on FTP can be found in the local archive folder of e: drive then show message "FileAlreadyThere" (I will ultimatley change it to do nothing); if the filename on FTP cannot be found in the local archive folder of e: drive then transfer the file to the local package folder on d: drive.
While the script task is executing I was watching it closely, but the problem i saw is that: If some files on FTP are already in local archive folder and some are not, then it the files which are already in the archive folder are dumped to the package folder; then after that the files which are not in the archive folder are then dumped to the package folder. But I only want the new files on FTP to be transferred to the package folder for further processing.
Then after this is finished, I saw all the files in the package folder are refreshed one after another, after the first round of refresh the second round starts, after the second round finishes it then stopped. I saw it refreshes itself because the 'Date Modified' of the file changes. And I saw the script task turned green.
I don't see how the code below produced this result. Something is wrong in the logic of the loop? Anyone has any idea why it's behaving the way it is now? And how to change the code to accomplish what I want? Thanks a lot!!
-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Imports System Imports System.IO Imports System.Data Imports System.Math Imports Microsoft.SqlServer.Dts.Runtime Public Class ScriptMain Public Sub Main() Dim cm As ConnectionManager = Dts.Connections.Add("FTP") cm.Properties("ServerName").SetValue(cm, "ftp2.name.com") cm.Properties("ServerUserName").SetValue(cm, "username") cm.Properties("ServerPassword").SetValue(cm, "password") cm.Properties("ServerPort").SetValue(cm, "21") cm.Properties("Timeout").SetValue(cm, "0") cm.Properties("ChunkSize").SetValue(cm, "1000") '1000 kb cm.Properties("Retries").SetValue(cm, "1") Dim ftp As FtpClientConnection = New FtpClientConnection(cm.AcquireConnection(Nothing)) ftp.Connect() ftp.SetWorkingDirectory("/directory") Dim fileNames() As String Dim folderNames() As String ftp.GetListing(folderNames, fileNames) If fileNames Is Nothing Then MsgBox("NoFileOnFTP") Else Dim fileName As String For Each fileName In fileNames If File.Exists("c: emp" + fileName) Then MsgBox("FileAlreadyThere") Else ftp.ReceiveFiles(fileNames, "c: emp", True, True) End If Next End If
I receive the input file with some 100 columns and some 20k+ rows and I want to check the incoming input row is existed in the database or not based on 2 key columns. If the row is existed then I need to check all the columns (nearly 100 columns) values in input and the database are equal or not. If both are equal I need to treat them seperately if not there is a seperate logic. How Can I do that check for each row and for each column?
Basically the algorithm is like this, if the input file row is not existed in the database then treat that as new row else if the input row is existed in the database then check all the columns are equal or not. If all the columns are equal then treat that as existing row and do nothing else if some columns are not equal then treat this row seperately.
I found some thing to achieve the above thing. 1. Take the input row and check in the database. 2. If the row is not found in the database then treat it as new row. 3. If row is found in the database then a) Take the source row and prepare a concatenated string for all the columns b) Take the database row and prepare a concatenated string for all the columns c) Find out the hash code for the 2 strings and then compare hash codes for equal.
The disadvantage of this is running a loop 2*m*n times where m is the number of rows and n is the number of columns. It should be done 2 times for input file row and database row.
Can anybody suggest a good method to do this?
What does the function "GetHashCode" for InputBuffer in method "Public Overrides Sub Input0_ProcessInputRow(ByVal Row As Input0Buffer)" will do? Will it generates hash code based on all the columns values?
when I'm in MediaImportLog , I want use column ImportSource to compare with column ChainCode in table BillerChain ( so I get BillerInfoCode) and then use the BillerInfoCode I got to compare with column BillerCode in Table Bill ( I get CompanyCode) finally I use CompanyCode to compare with column CompanyCode in table DataBackup so I can get the company's keepmonth How can I get the keepmonth? can I use parameters ?
what I want to achieve is to load a text file that has email addreses from disk and using the email addresses in the text file look it up against the email addresses in the database table then once matched delete all the users in the table whose email address were in the text file.
I also want to update some users using a different text file.
Hi, I have added a new SQL Server database to my project and I was wondering how to do generate a SQL dump of the database, tables, and queries? I couldn't find any utilities in Visual Studio 2008 Pro. I also couldn't figure out how to create indexes using VS. Thanks!
1. I need to create a replication enviroment. In that i want to snapshot the database as and when i want from my Java program, so that selective tables are updated with latest data. Is there anyway i can do with out creating native replication methods.
2. Is there any way with out creating a snapshot or replication. I need to copy the data from my master database to child database(my reporting database) with out straining much my master database
I get the following error when doing a dump on a 6.5 database. *** Page '697569' of database 'FOREX_UPLOAD' in the database has the PG_DMPTRLR bit set. Page '698121' of database 'FOREX_UPLOAD' in the database has the PG_DMPTRLR bit set. Page '698233' of database 'FOREX_UPLOAD' in the database has the PG_DMPTRLR bit set. *** I have run SP5 as suggested,and can see that more of the db has been dumped than before but still I get the error I get no errors when doing dbcc's. Any ideas???
When I use "DUMP DATABASE", it will help me to backup both database and transaction log. However, it will not help me to clean up the transaction log. Therefore, I am worry that will the following happen?
1. DUMP DATABASE Start backup database ( there are transactions during the backup process ) Finish backup database Start backup transaction Finish backup transaction
2. DUMP TRANSACTION Start backup transaction and clean up the transaction log
So, will the transactions been backuped twice? If yes, will it cause those transaction do twice when I restore the database then restore the transaction?
I need to dump transaction log every night. Before I do that I'd like to check those open transactions and kill the ones not part of sa or SQL server dba's activity. How can I achieve this?
After Dumping a database device, I still find the xxx.DAT files inthe /Data directory. Can these be 'safely' removed/backed up ? They are taking up WAY too much space.
Question 2: After installation and after creating devices & databases, is there any way to reassign the /Data directory on another drive? Either for future devices/databases or existing?
I understand from the beta sql 7 newsgroups that 7 can't restore a 6.5 dump. This is extremely frustrating that MS wouldn't put this type of backword compatibility into 7.
Has anyone found another way to accomplish this? I can't use transfer manager as the 2 servers are on seperate, non-connected boxes. I ftp'd the dump from 6.5 server to a middle-man machine, then copied it to the 7.0 server.
Am I looking at bcp'ing out all 203 tables in my 1GB db then bcp'ing them back in?
Is there a way to dump sql in Access like in Transact SQL or mysqldump? I've gotta convert an Access Database into a MySQL database, it is a small database, but I'd not want to put it in by hand .
If not, I'll just make a PHP script to rip it from a textfile.
I am trying to import a database from a dump file. The dump file is from a sql server db.
I have created the db on another server and want to import the data from the dump file.
I use the bcp like this :
bcp nnfdb.sa.users in x -Usa -P
nnfdb is the name of the database users is the name of a table x is the name of the dump file.
But I get this error : SQLState = S0002, NativeError = 208 Error = [Microsoft][ODBC SQL Server Driver][SQL Server]Invalid object name 'nnfdb.sa.users'.
also I am sure if I have to run the bcp for each table or if I can import all tables in one bcp run.
****************************************************************************** * * BEGIN STACK DUMP: * 04/25/07 08:32:19 spid 22 * * Exception Address = 00412373 (CTableIsolation::FNeedLockClass + 14) * Exception Code = c0000005 E * Access Violation occurred reading address 00000014 * Input Buffer 132 bytes - * S E L E C T S u p F r o m , S u p T o F R O M S u p p o r t P r * o f i l e s w h e r e S u p P r o f i l e = 3 0 4 5 * * * MODULE BASE END SIZE * sqlservr 00400000 008d2fff 004d3000 * ntdll 77f60000 77fbefff 0005f000 * KERNEL32 77f00000 77f5efff 0005f000 * ADVAPI32 77dc0000 77dfefff 0003f000 * USER32 77e70000 77ec1fff 00052000 * GDI32 78140000 78174fff 00035000 * RPCRT4 77e10000 77e66fff 00057000 * ole32 77b20000 77bd0fff 000b1000 * OLEAUT32 65340000 653dafff 0009b000 * VERSION 77a90000 77a9afff 0000b000 * SHELL32 77c40000 77d7afff 0013b000 * COMCTL32 71710000 71793fff 00084000 * LZ32 779c0000 779c7fff 00008000 * opends60 41060000 41085fff 00026000 * ums 41090000 4109cfff 0000d000 * MSVCRT 78000000 78045fff 00046000 * sqlsort 04000000 0408efff 0008f000 * MSVCIRT 780a0000 780b1fff 00012000 * sqlevn70 410a0000 410a6fff 00007000 * rpcltc1 77bf0000 77bf6fff 00007000 * COMNEVNT 410b0000 410fefff 0004f000 * ODBC32 1f7d0000 1f803fff 00034000 * comdlg32 77d80000 77db1fff 00032000 * SQLWOA 41100000 4110bfff 0000c000 * odbcint 1f8c0000 1f8d5fff 00016000 * NDDEAPI 75a80000 75a86fff 00007000 * WINSPOOL 77c00000 77c17fff 00018000 * SQLTrace 41130000 4117dfff 0004e000 * NETAPI32 4ca00000 4ca40fff 00041000 * NETRAP 77840000 77848fff 00009000 * SAMLIB 777e0000 777ecfff 0000d000 * WSOCK32 776d0000 776d7fff 00008000 * WS2_32 776b0000 776c3fff 00014000 * WS2HELP 776a0000 776a6fff 00007000 * WLDAP32 77950000 77978fff 00029000 * SQLFTQRY 41020000 4103afff 0001b000 * SSNMPN70 41190000 41195fff 00006000 * SSMSSO70 411a0000 411aafff 0000b000 * SSMSRP70 411b0000 411b7fff 00008000 * ENUDTC 74e20000 74e36fff 00017000 * XOLEHLP 74e40000 74e48fff 00009000 * MTXCLU 74e50000 74e5cfff 0000d000 * ADME 74e60000 74e72fff 00013000 * DTCUtil 74e80000 74e89fff 0000a000 * DTCTRACE 74e90000 74e96fff 00007000 * CLUSAPI 7f230000 7f23cfff 0000d000 * RESUTILS 7f250000 7f259fff 0000a000 * MSDTCPRX 74ea0000 74ed0fff 00031000 * DTCCM 74ee0000 74efdfff 0001e000 * msafd 77660000 7766efff 0000f000 * wshtcpip 77690000 77698fff 00009000 * rpclts1 77e00000 77e05fff 00006000 * RpcLtScm 755b0000 755bafff 0000b000 * MSWSOCK 77670000 77684fff 00015000 * rnr20 75600000 7560dfff 0000e000 * RpcLtCcm 756d0000 756defff 0000f000 * security 76e70000 76e81fff 00012000 * msapsspc 756e0000 756f3fff 00014000 * MSVCRT40 779d0000 779e4fff 00015000 * schannel 77400000 7741dfff 0001e000 * MSOSS 75710000 75734fff 00025000 * CRYPT32 75740000 757b4fff 00075000 * MSASN1 757c0000 757cffff 00010000 * msnsspc 757d0000 757eefff 0001f000 * digest 757f0000 757fffff 00010000 * SQLRGSTR 01b60000 01b64fff 00005000 * xpsqlbot 01bf0000 01bf5fff 00006000 * sqlboot 01c00000 01c07fff 00008000 * EntApi 0a000000 0a011fff 00012000 * PSAPI 76bf0000 76bfafff 0000b000 * WININET 01db0000 01e45fff 00096000 * SHLWAPI 01e50000 01eb8fff 00069000 * xpstar 02440000 02470fff 00031000 * SQLWID 02480000 02485fff 00006000 * SQLSVC 02490000 024a8fff 00019000 * odbcbcp 024b0000 024b5fff 00006000 * SQLRESLD 024c0000 024c5fff 00006000 * W95SCM 024d0000 024d7fff 00008000 * SQLSVC 024e0000 024e5fff 00006000 * sqlimage 02590000 025bcfff 0002d000 * * Edi: 00100C70: 000a0100 00030008 00000000 00000000 00100e70 00100e50 * Esi: 00000000: * Eax: 00000000: * Ebx: 00000004: * Ecx: 00000000: * Edx: 00000004: * Eip: 00412373: 007ea6e2 05f61577 01f88318 468b0013 d49f850f 00147e83 * Ebp: 01C8F864: 00000001 00000000 2149a1e4 00000000 00412336 01c8f898 * SegCs: 0000001B: * EFlags: 00010246: 003d0052 00490044 005f004c 00410043 004f004c 005f0053 * Esp: 01C8F854: 00412336 01c8f898 00000000 00000004 2149a1e4 00000000 * SegSs: 00000023: *******************************************************************************
SqlDumpExceptionHandler: Process 22 generated fatal exception c0000005 EXCEPTION_ACCESS_VIOLATION. SQL Server is terminating this process.
Using 'sqlimage.dll' version '4.0.5' Stack Dump being sent to C:MSSQL7logSQL00026.dmp
Sunrise406, creation date and time: 2001/11/16(14:34:00), first LSN: 647599:6555:1, last LSN: 647633:8427:1, striped: 0, number of dump devices: 1, device information: (FILE=1, TYPE=DISK: {'D:MSSQL7BACKUPSunrise406_tlog_200704250000.TRN'}).
Using 'xpstar.dll' version '2000.28.09' to execute extended stored procedure 'xp_regread'.
Database backed up with following information: Database: Sunrise406, creation date and time: 2001/11/16(14:34:00), pages dumped: 47456, first LSN: 647633:8425:1, last LSN: 647633:8427:1, sort order: 52, striped: 0, number of dump devices: 1, device information: (FILE=1, TYPE=DISK: {'D:MSSQL7ackupSunriseDailyBackup9pm'}).
I have a sql 2000 database on my local machine that I need to get to a faraway server so he can make one like it on his server. The admin tells me Ineed to make a dump of it and send it to him. I just don't know how to dothat. Can someone help me out?Thanks,Jerry
Tried googling, but I guess I have the wrong terminology for MSproducts.Using PostgreSQL, or MySQL, if I wanted to tranfer a database from onemachine to another, I'd do a "dump" with a single command, such as:mysqldump [databasename] > [path/filename]Then I'd take the file generated and dump it back onto a new machine.What's the process with MS SQL Server? Can someone point me in theright direction to a webpage dealing with this?--[ Sugapablo ][ http://www.sugapablo.com <--music ][ http://www.sugapablo.net <--personal ][ Join Bytes! <--jabber IM ]
I am having problems and can€™t seem to get a grip as to what is happening. My dump files are filing up fast and I have to go and delete them because they take up too much space.
We started playing with SharePoint 2007 search and think it is the culprit.
Any advise would be great.
Here is a little info form the dump file
BugCheck Dump ===================================================================== This file is generated by Microsoft SQL Server version 9.00.1399.06 upon detection of fatal unexpected error. Please return this file, the query or program that produced the bugcheck, the database and the error log, and any other pertinent information with a Service Request. Computer type is AT/AT COMPATIBLE. Bios Version is DELL - 1 Phoenix ROM BIOS PLUS Version 1.10 A04 Current time is 15:10:02 04/24/07. 4 Intel x86 level 15, 3400 Mhz processor (s). Windows NT 5.2 Build 3790 CSD Service Pack 1. Memory MemoryLoad = 57% Total Physical = 4095 MB Available Physical = 1741 MB Total Page File = 5976 MB Available Page File = 3709 MB Total Virtual = 2047 MB Available Virtual = 290 MB ***Stack Dump being sent to C:Program FilesMicrosoft SQL ServerMSSQL.1MSSQLLOGSQLDump6891.txt SqlDumpExceptionHandler: Process 100 generated fatal exception c0000005 EXCEPTION_ACCESS_VIOLATION. SQL Server is terminating this process.
I need to make a gigantic collection of excel sheets searchable from a web interface.I need entire rows fetched into a webpage depending on the users query.Mind you I'm not searching the names (filenames) of these excel files, butthe contents inside the excel file. E.g. in a file test.xls, if I search for the word"test1" from the web interface, the entire row in the excel file containingthe word test1 should be displayed.One way out is to dump the contents of the excel sheets into a database, andsearch the database via ASP.For this, can anyone tell me how to efficiently dump the contents of an excelsheet into SQL server?Or if anyone can suggest an alternate strategy for searching this mammothcollection of excel files' contents, if would be great.Thanks a lot.