I'm able to pull down all my tables needed, but I need to push a table from my handheld database that was not pulled. Is this possible?
The table is used to save data on some input screen on my handheld application and I need to push it to a 'staging' table on my main sql server before it is inserted into the correct tables.
I would like to create a database for keeping track of payroll data for employees where the supervisors (job coaches) on our workshop floor can use a Pocket PC device to record the hourly employee data on the fly. Then at the end of the day, the supervisor can place the device in a cradle of some sort and synch the newly entered data into the main database.
I'm guessing that SQL Server Compact edition would be perfect for this type of task? Is that correct? Can someone give me recommendations on how to go about setting this up? What should I use as the main database? SQL Server? Access? Any advice is appreciated!
I need to copy a large amount of data from one table and insert it into another table.
The design of the destination table is exactly the same as the source table except for the fact that it has one extra field. Can I copy; in a single SQL statement; all rows in one table (that match given criteria) into another table allowing for the extra field?
I have a production database that I would like to have copied over to a backup database on a separate server every evening. I don't want to mirror, I just want the databases synced up every evening.
The servers are physically attached through a gigabit switch and the database is relatively small, so I don't think that speed will be an issue.
Could someone point me to an article about the best way to accomplish this?
When I try to do a pull using a view I'm getting the following error message:
€œThe query cannot be tracked. There might not be a primary key, or the query might involve multiple tables. [ Query string = SELECT * FROM DATABASE_NAME.viewname ]€?
but when I do a 'select col1 from viewname' I get my data. The view has some inner joins included in it, could that be the issue?
but after the assignment the LocalConnectionString is different from ConnectionStringWithoutCredentials, there are much more ssce parameters.
What I don't understand is:
why I'm getting this error even after closing the connection ?
Since the error report [32] the ssce conflicting parameters shoud be AUTO_SHRINK_THRESHOLD, FLUSH_INTERVAL and MAX_DATABASE_SIZE
I tried to cut & paste the same values from LocalConnectionString into app.config connection string, at first It seemed to work, but now I got the same error.
I have tried using the SqlClient class to connect to the database, but I haven't had much success. Here is my setup: Desktop/Workstation has Windows XP, ActiveSync, .NET Framework 2.0, and SQL Server 2005.
The production Workstation will have 2000; so I do want to ensure SQL Server 2000 compatibility with my solution.
Handheld device is an Intermec 751g with docking station, hooked to PC host using USB connector cable and ActiveSync, Windows CE 4.2, .NET Compact Framework 2.0 with SP1 and latest patch.
Development tools consist of Visual Studio 2005 and C# .NET language.
Note that when I attempt to ping my desktop from the handheld command prompt I get "transmit failed, error code 11010".
Also, I am able to access the Internet from my handheld device.
My error is encountered when I execute this fill code:
int returnValue = this.Adapter.Fill(dataTable);
Please note that all of my code is generated by the VS 2005 IDE at this juncture, as I am simply trying to get the plumbing to work.
Here is the error that I encounter:
System.Data.SqlClient.SqlException was unhandled Message="SqlException" Class=20 LineNumber=0 Number=6 Procedure="ConnectionOpen (Connect())." Server="09Q-GHAZEL2\SQL2K5" Source=".Net SqlClient Data Provider" State=0 StackTrace: at System.Data.SqlClient.SqlConnection.OnError() at System.Data.SqlClient.SqlInternalConnection.OnError() at System.Data.SqlClient.TdsParser.ThrowExceptionAndWarning() at System.Data.SqlClient.TdsParser.Connect() at System.Data.SqlClient.SqlInternalConnection.OpenAndLogin() at System.Data.SqlClient.SqlInternalConnection..ctor() at System.Data.SqlClient.SqlConnection.Open() at System.Data.Common.DbDataAdapter.QuietOpen() at System.Data.Common.DbDataAdapter.FillInternal() at System.Data.Common.DbDataAdapter.Fill() at System.Data.Common.DbDataAdapter.Fill() at DeviceApplication1.AdventureWorksDataSetTableAdapters.ContactTableAdapter.Fill() at DeviceApplication1.Form1.Form1_Load() at System.Windows.Forms.Form.OnLoad() at System.Windows.Forms.Form._SetVisibleNotify() at System.Windows.Forms.Control.set_Visible() at System.Windows.Forms.Application.Run() at DeviceApplication1.Program.Main()
Thanks again for your insight. Sincerely, -Dotnetfellow dotnetfellow@yahoo.com
We are in need of some help around the use of SQL Server 2005 Mobile Edition.
The problem comes in when we have to import a large list of equipment ID€™s into the database.
The way that things work today is that the database only runs on the handheld computer, so we are having to process a large amount of data on the handheld therefore taking a lot of time.
If there were a way to have a €œseed€? database on the server side (running on Windows Server 2003 or similar) and import all of this data on the server side it would be more efficient and require less time.
Once the data was imported on the server side we can simply transfer this seed database down to the handheld
Could you please throw some light on the above queries.
We are performing merge replication between SQL Server 2005 and SQL Server Mobile with more than 1100 handheld, and recently replications are terminating with eror code 80040E19. We can only avoid the problem by deleting the related user's partition directory and reruning the related user's dynamic snapshot.
Hi, I have application in which i am performing synchronization between SQL Server 2000 and SQL Server 2005 CE. I have one table "ItemMaster" in my database.There is no relationship with this table,it is standalone.I am updating its values from Windows Mobile Device.
I am performing below operations for that. Step : 1 Pull To Mobile
Code BlockmoSqlCeRemoteDataAccess.Pull("ItemMaster", "SELECT * FROM ItemMaster", lsConnectString,RdaTrackOption.TrackingOn);
Step : 2 Using one device form i am updating table "ItemMaster" table's values.
So i am getting an error on 3rd step. While i am trying to push it says, "The Push method returned one or more error rows. See the specified error table. [ Error table name = ]". I have tried it in different ways but still i am getting this error.
Note : Synchronization is working fine.There is not issue with my IIS,SQL CE & SQL Server 2k.
Can any one help me?I am trying for that since last 3 days.
Dear all,, I used a code that import data from an excel file into a dataset, now I want to insert the dataset into a table in my database(SQLserver database) using a VB.NET code Could you help me? Thanks in advance,, Here is my code: Imports System.Data.OleDbPartial Class _DefaultInherits System.Web.UI.Page
Protected Sub Button1_Click(ByVal sender As Object, ByVal e As System.EventArgs) Handles Button1.ClickDim connString As String = ConfigurationManager.ConnectionStrings("xls").ConnectionString ' Create the connection object Dim oledbConn As OleDbConnection = New OleDbConnection(connString) Try ' Open connection oledbConn.Open() ' Create OleDbCommand object and select data from worksheet Sheet1Dim cmd As OleDbCommand = New OleDbCommand("SELECT * FROM [Sheet1$]", oledbConn) ' Create new OleDbDataAdapter Dim oleda As OleDbDataAdapter = New OleDbDataAdapter() oleda.SelectCommand = cmd ' Create a DataSet which will hold the data extracted from the worksheet.Dim ds As Data.DataSet = New Data.DataSet() ' Fill the DataSet from the data extracted from the worksheet.oleda.Fill(ds, "Sheet1") ' Bind the data to the GridView GridView1.DataSource = ds.Tables(0).DefaultView GridView1.DataBind() Catch Finally ' Close connection oledbConn.Close() End TryEnd Sub End Class
I'm trying to replicate two very big databases with about 10 million of 4000 characters each. The publisher is SQL 2000, subscriber is SQL 7.0
The subscriber will also perform full text searches.
I'm trying to decide wheter I should use PULL or PUSH. The publisher is operating on a very low quality/speed internet connection, where the subscriber is enjoying a T1.
While trying to push a tracked table using RDA.push, I get the following error:
Error Code: 80004005
The message cannot be built. The make message failed.
Minor Err: 28581
Source: Microsoft SQL server 2005 Mobile Edition.
All other tables in the database are getting pulled and pushed correctly. This table is different only in the larger number of columns, around 150. It has a primary key, no other constraints.
Any help to find the reason for this error will be greatly appreciated.
I am developing an application in which i have to send data from local Sql Server compact edition database[Which is in a Windows Mobile Device,] to central server[SQL Server 2005]. I am using RDA method for communication
Can i use push method to send data from local DB to Central DB?
Is it must to use PULL method before using PUSH method?
Hello all. Please excuse my ignorance, as this is not my territory. I administer a website which is hosted remotely. This site has SQL7 running the data to dynamically build the site. Every Sunday our hosting service runs a DTS package to push the data they have down to us, so we can run reports and analyze it. We recently upgraded to SQL2000, while our host has stayed with SQL7. Now our DTS is failing. They say it is because 7 cannot push to 2000. But they think that we could pull from them. How do I go about setting that up? Will the DTS wizzard walk me through most of it?
I did not see this one coming, and I am not sure if I did something wrong.
How do you push data from sql05 to sql2k?
I set up a data flow task, with one sql05 connection magager and another sql2k connection manager. Then when I tried to map them, I cann't!
The message on the box said: The connection manager uses an earlier version of sql server provider. Bulk insert operations require a connection that uses a sql server 2005 provider.
I have been trying different source, destination and transformation, but seems like missing something.
I am using the Pull command to pull two fields, on is the primary ID (int) non identity and the other is Description which comes down as an ntext type. This works fine but if I change the description and use the push command I get the following error:-
The Query processor could not produce a query from the optimizer because a query cannot update a text, ntext, or image column and a clustering key at the same time.
I am really stuck with this one so if anyone can shed some light on it I would be much appreciated.
Catch RDAConnectionException As Exception MessageBox.Show("Can not push Header Data: " & RDAConnectionException.ToString, "Loading Key Tracker") Finally
RemoteAccess.Dispose() End Try Cursor.Current = Cursors.Default End Sub
TABLE DEFINITION: USE [MobileKeyDB] GO /****** Object: Table [dbo].[KeyHeader] Script Date: 07/11/2007 09:48:24 ******/ SET ANSI_NULLS ON GO SET QUOTED_IDENTIFIER ON GO CREATE TABLE [dbo].[KeyHeader]( [trans_id] [int] IDENTITY(1,1) NOT NULL, [user_id] [int] NOT NULL, [date_stamp] [datetime] NOT NULL, [signature] [nvarchar](100) COLLATE SQL_Latin1_General_CP1_CI_AS NOT NULL, [status] [int] NOT NULL, [id] [int] NOT NULL, [date_stamp2] [datetime] NOT NULL, CONSTRAINT [PK_KeyHeader] PRIMARY KEY CLUSTERED ( [trans_id] ASC )WITH (IGNORE_DUP_KEY = OFF) ON [PRIMARY] ) ON [PRIMARY]
Does anyone know if you can add columns to a local pulled table and if so can you use a select command to push the table back and exclude the added columns.
Basically I need to know if a record has been added on the PDA so I can get an ID from the server before pushing it back. I cannot alter the sql server database because it is part another application.
As I understand replication in Sql2K the only difference in push or pull subscriptions is where the agent runs. If I wanted changes in the publisher to be sent to the subscribers immediately after the change then I thought push would be better. But, if I am equally interested in changes made at the subscriber then what should I use? Or does the agent monitor both the publisher and subscriber at the same time?
I'm able to pull the metatdata down from my sql server to my handheld for table, I can then add data to that table on the handheld, when I try to push it back to Sql Server table I pulled from, I'm getting the following error message:
The Push method returned one or more error rows. See the specified error table. [ Error table name = ErrorTable ]
what does this mean, and how can I push my table back to the table I pulled from? I have trackingOn set on the pull process.
After doing some research it seems like you can only push the same table once using rda.push -- is this correct? If yes, are there any other alternatives for saving changes to the table back to SQL Server aside from merge replication?
One idea I am toying with is to pull the tracked table with 0 records, save changes to the tracked table, push, drop table and pull, repeating this process everytime I push the data. Wondering is somebody has any advice?
I am using merge replication with a push subscription type. I am wondering if the updates of the tables on the subscriber side are push to the publisher from the subscriber or pulled from the subscriber by the publisher when the syncronisation takes place. this makes a big diferrence for me and i can't find the answer to this question anywhere...
if anyone could answer it would be really appreciated
Looking for a faster method of moving data from SQL to Oracle.
I'm attempting to push a sql table into an oracle table (sql server 2000 oracle 7, 8, and 9). I have no problem doing this with either 'Oracle Provider for OLE DB' or the 'Microsoft OLE DB Provider for Oracle'. None of my data is being transformed so its a straight import. With the hardware I'm using it takes nearly 3 seconds to import 1000 rows. While this isn't too bad, I need to import upwards of 4 million rows and this results in unacceptable time results.
I do have an oracle script that imports the csv files of the tables, but I'm looking for an all inclusive sql solution.
Does anyone know of another method in SQL that I can use to push the data faster?