Check Point Manually
May 8, 2007Dear Experts,
please guide me, how can i run checkpoint manually?
and how can i make it done automatically? what is sql server default?
is it a problem (making the checkpoint to run more times)?
Vinod
Dear Experts,
please guide me, how can i run checkpoint manually?
and how can i make it done automatically? what is sql server default?
is it a problem (making the checkpoint to run more times)?
Vinod
Hello folks,
The other day I was checking the dboption and I couldnt see the "Truncate log on Check Point" option in the dboption window. Wondering if i'm looking at the right place.
Appreciated, If any one point me where can I look for this.
Thanks,
Sindi
I have Simple package with DFT with Source and Destination ,check point is enabled to the package
I Have 100000 rows from source , while 5000 rows transferring and committed in destination i lost the connection and leading to failure of the package
what will happen after second time running the package .
I have created a package that use SSIS check pointing for failure-retry mechanism. I knew that when this package fails, on restarting the package it starts from the task where it got failed.
Is it possible for me to override this on a custom condition and start the package at an earlier task where it executed successfully?
Ex:
Check pointing is enabled.
FTP task è Write to Staging Table è Write to Target table.
Assuming I am downloading XML file through FTP and writing to a table.
FTP Download is successful.
Read from XML file and write to Staging table task failed because of downloaded file is not a well formed XML.
Here, FTP task is completed successfully; it failed only in the second task. When I re-run the package; it starts from the second [Read from XML file and write to staging table] task because of check pointing.
Is it possible for me to restart from the FTP task on a custom condition where it executed successfully?
I have huge data and i am loading data from EXCEL to database table, after loading 80 percent data i am getting some error. My package got failed and it has lots of transformation and took around 6 hours to process completely because of that i don't want it to reload from start. if i run it again it should start from next record from where i got the error.
View 3 Replies View RelatedHello all,
First off, I appreciate the time that those of you reading and responding to this request are offering. My quesiton is a theoretical and hopefully simple one, and yet I have been unable to find an answer to it on other searches or sources.
Here's the situation. I am working with SQL Server 2005 on a Windows Server 2003 machine. I have a series of databases, all of which are in Full recovery mode, using a backup device for the full database backups and a separate device for the log backups. The full backups are run every four days during non-business hours. The log backups are run every half hour.
Last week, one of my coworkers found that some rarely-used data was unavailable, and wanted to restore a database to a point in time where the data was available. He told me that point in time was some time back in November.
To accomplish this, I restored the database (in a separate database, as to not overwrite my production database) using the Point in Time Recovery option. I selected November from the "To a point in time" window (I should note that this window is always grey, never white like most active windows, it seems), and the full database backup and the subsequent logs all became available in the "Select the backup sets to restore" window.
I then tried a bevy of different options from the "Options" screen. However, every restore succeeds (ie: it doesn't error out), but seems to be bringing the database back to a current point in time. It's never actually going back to the point in time I specify.
My questions are as follows:
a) Is it possible to do a point in time recovery to a point in time BEFORE the last full database backup?
b) If so, what options would you recommend I use? (ie: "Overwrite the existing database", restore with recovery, etc etc).
I again appreciate any and all advice I receive, and I look forward to hearing from anyone and everyone on this topic. Thank you.
Ryan
I'm running SQL Server 2005 on a Server 2003 machine serving both our home network as well as a remote site through a point-to-point T1. While file transfer speeds are up to par, the remote site's interaction with SQL Server (Point of sale system) is very slow. After testing I am certain that it has nothing to do with the actual physical machine in place neither is it an issue with the program itself since speeds are as they are supposed to be over the home network lan. It seems that there might be a packet size issue or something of the sort. Has anyone dealt with this before or have any thoughts?
Thanks,
Peter
I have multiple ODBC connection and how to check all connection automatically during routine check by using batch file.
View 5 Replies View RelatedHi everyone
In my project, sometimes, I need to run log shipping manually rather that it runs automally when the schedule occured.
manually run log shipping means call the log shipping in code.
Thanks.
Hi All,
This may sound quite easy for you, however I am a newbie in SQL.
So I have an identity field, which I wish to increment automatically every time I do an insert. However the starting integer has to be the MAX value from another table.
So for example I am doing an insert in a #temp table
INSERT INTO #temp(name, surname)
SELECT name, surname from table1
Now the personId of the #temp table has to start from the MAX of table2
Ie SELECT MAX(personId) from table2
The SELECT MAX(personId) from table2 can also be NULL, ie the first time I am inserting, so I also have to cater for this scenario.
can anyone help?
Thanks
Johann
I am new to the ASP.NET and SQL scene.
How do you run a sql script manually?
I have MSDE SP3 installed and running. I am having a problem with the localhost setup for some of the kits and need to run the scripts manually.
Any help would be great, including helpfull learning resources.
Hi:
I have a UDF that splits a string based on a delimiter.The UDF returns a table of records:1- ID2- ValueThe string I am sending to split is:'2-4-07:00 AM-08:30 AM'I am splitting on '-'What I get back is a table of 4 rows.I want to get each value, convert it to the right data type, then assign it to its own variable.How can I get each row alone manually? Can I? Without using Cursors?if I do a "select * from Split(@myData,'-')", then I cannot retrieve each field alone,Thank you.
Hardware:
IBM Netfinity 8500
2 processors Xeon 700
1,5 Gb memory
Windows 2000 Server SP2 Build 2195
SQL Server 2000 Standard Edition 8.00.534 SP2
There is only on Database (DB) of 16 Gb in drive G.
Drive G has 32 Gb space free.
Yesterday we appended tables to the database and in SQL logs appears the next error:
2003-01-23 12:26:42.57 spid101 fcb::ZeroFile(): GetOverLappedResult() failed with error 2.
2003-01-23 12:26:42.61 spid101 Error: 1105, Severity: 17, State: 2
2003-01-23 12:26:42.61 spid101 Could not allocate space for object 'ttdssc030104' in database 'MYDATABASE' because the 'PRIMARY' filegroup is full..
2003-01-23 12:26:48.03 spid101 fcb::ZeroFile(): GetOverLappedResult() failed with error 2.
DB configured to grow automatically by 100 Mb and transaction log Automatically grow in 10% .
Unrestricted file grow selected on both.
I try to expand the DB manually by Enterprise Manager to 20 Gb but not work and in SQL log appears the error
"2003-01-23 12:26:48.03 spid101 fcb::ZeroFile(): GetOverLappedResult() failed with error 2."
In Enterprise Manager-Databases-Properties-General-Size of DB maintain 16Gb.
Windows explorer say MYDATABASE.MDB is 20480 MB.
I delete the tables inserted and the problem persist.
Thanks in advance,
I am getting 228 rows once I manually inserted the value for the ct.code in test region. I would like to insert into other regions using the SQL, how can I do that ?
select distinct
ocmi.name,
ct.code,
ctv.value
from
cv3ordercatalogmasteritem ocmi
join CV3CatalogClassTypeValue ctv on ctv.CatalogMasterGUID = ocmi.guid
join CV3ClassType ct on ct.guid = ctv.ClassTypeGUID
[code]....
I've been studying/experimenting with the replication features and though the Updating Susbcribers Replication Model was the answer to my data environment.
What we want: We have a live server (Publisher) that'll serve content to our users, but we want an "offline/data entry" server (Subscriber) that'll pretty much have the exact same data as the live server for data entry and validation purposes. Changes and inserts would be done on the Subscriber locally, vetted through, and once done, it'd be updated to the Publisher.
I've managed to get some semblance of the above working, but the updates are nearly instantaneous. Which means that changes/inserts that may be erroneous on the subscriber machine would almost always be seen on the publisher machine before our operations can do anything about it.
Is there a way to configure the replication process to hold off until the admin triggers it? Or am I using the wrong feature?
Please help me to clean a Windows 2000 Pro instance which is "dirty"with the MSDE 7.0 and/or/both MSDE 8.0 previously incorrectinstalled/removed.I have a software product installation containing MSDE 7.0. When I triedto install MSDE 7.0 I got an error "You have installed MSDE".*** Sent via Developersdex http://www.developersdex.com ***
View 5 Replies View RelatedHi all
I'm trying to use the Slowly Changing Dimension transformation and it seems that in order to get the transformation to work as expected the custom property "sql command" has to be manually changed so that a "RTrim" function is applied to all string columns referenced. As an example consider this:
SELECT Initials, Department_Name from dimEmployee
has to be changed to
Select RTrim(Initials) as Initials, RTrim(Department_Name) as Department_Name from dimEmployee
where Initials is the "bussines key" and "department_name" a historic attribute.
Is this the normal behaviour ??
Regards, Steen
Hi,
I have a table
Code Snippet
CustomerDetail
customerid
customername
status
app_no
deskno
Now everytime a new customer is created, i allocate the deskno. The logic I use is
Code Snippet
create proc spoc_incrementdesk
as
DECLARE @deskno int
select @deskno = (SELECT max(deskno) from customerdetail)
Insert into customers(1,'rick',1,23232,@deskno)
If there are already records present in the customer table, it works well. However if the very first record is getting inserted it inserts null in the deskno because when no records are present (max(deskno) = null).
At time time of insertion, how do i check if @deskno = null and allot it 1.
Vidkshi
I have a report scheduled for email delivery at a certain time of the day. Is there a way I can execute the subscription ON DEMAND so that when ever I execute it, the report is processed and emailed? Is there a way to run the subscription from a batch file/ssis/tsql etc?
any help appreciated.
thanks
Jacob
When I force Mirroring to fail over while watching the Mirroring Monitor I notice something interesting. Before I failover, the servers are synchronized and there is zero unsent log and zero unrestored log. When I failover, the mirror changes to "restoring", stays there for about four minutes, switches to "Principal, restoring", and then the unsent log and unrestored log numbers jump to values in the tens of megabytes, and we spend time restoring.
We wait about five minutes between requesting the failover and having the servers back up again, answering requests.
Why does failover take so long? We've been thinking that failover shouludn't take so long because the servers think they're synched. What causes the log spooling spike? Is the state refelcted by the mirroring monitor inaccurate?
Is there a way to do this? It would be useful to test whether a subsciption works. At the moment I am setting the scheduled time a few minutes ahead and waiting for the report to run.
Subsciptions are also a convenient way to store different parameter settings for a report.
My preference would be to to do this through the Report Server UI (presumably from the subscriptions tab).
Alernatively, being able to call a report (with paramters) from a stored procedure woukd be useful.
Thanks,
Dick Campbell
Hello,
I use Reporting Services in my solution. The data i have as source is very uncertain. Sometimes i manually have to delete a record from the report (not the database itself). I woluld like to have a checkbox or something simular to delete from report and totals, grafs would be updated. At the same time I would like the report to make a comment that a record was taken away, alternative let the user make a comment. Is this possible? DO i have to make a own application for that?
Thank you for your help!
Best regards,
Luskan
hi i have an sqldatasource which has an insert command - a stored procedue is used.I have a text box with a button next to it . it is not in a datagrid.on the onclick event I would like to pass the value of the text box to the sqldatasource insert parameter ( it only expects this one parameter , and use the sqldatasource to do the insert basically doing a manual insert using the sqldatasource.does anyone know if this is possible thanks
View 2 Replies View RelatedHi, I have limited control on the server I use. I can not create data bases but can add tables and edit the data base that has been created for me. So what I would like to do is be able to use a tool such as aspnet_regsql through SQL Server Express to add the tables that Membership would automatically create. Any help in pointing me in the right direction would be great
Thanks
howlinhuskie
Recently i got one assignment from my senior, I have to merge multiple sql server databases,
Where to start?
How to start?
What prerequisites should consider?
What factor will cause error, how to tackle that error.
declare @kk int
set @kk=0
insert into tblSSAppsOrgEntityToEmployerMapDiffer
(Id,
OrgEntityCode,
EmployerId,
[Default],
[Code] ...
In above example Id is PK for Differ tbl and Temp tbl not having field related to this. thats why i have to take and increment that Id value manually.... but like above way i m getting error ..........
My client has a number of jobs that are run overnight. We've set themup to email me when they're completed. Every morning I get in to abunch of emails like this:<quote>JOB RUN:'Tech Pubs Email Notification' was run on 18/03/2006 at00:00:00DURATION:0 hours, 0 minutes, 0 secondsSTATUS: SucceededMESSAGES:The job succeeded. The Job was invoked by Schedule 10 (SendMail). The last step to run was step 1 (Send Mail).</quote>However, the most important job - the database backup - fails everytime.<quote>JOB RUN:'DB Backup Job for DB Maintenance Plan 'DB Maintenance Plan1''was run on 20/03/2006 at 18:00:00DURATION:0 hours, 0 minutes, 2 secondsSTATUS: FailedMESSAGES:The job failed. The Job was invoked by Schedule 7 (Schedule1). The last step to run was step 1 (Step 1).</quote>What's strange is that the job runs successfully if you kick it offmanually (in EM: right-click and "Start Job")!!! Does anyone have anyidea of why that might be? Where to look for diagnostic information?TIAEdward
View 6 Replies View RelatedI've mined countless threads all over the Net where people are having my problem, but none of the solutions I've seen seem to apply.
When I run a DTS package from Enterprise Manager, everything works great. When I schedule the same package as a job, it fails immediately and reports this error:
Executed as user: SERVER-NAMESYSTEM. DTSRun: Loading... Error: -2147221005 (800401F3); Provider Error: 0 (0) Error string: Invalid class string Error source: Microsoft Data Transformation Services (DTS) Package Help file: sqldts80.hlp Help context: 713. Process Exit Code 1. The step failed.
My hunch is the SERVER-NAMESYSTEM.DTSRun account doesn't have sufficient privileges, but I can't find how/where to fix that.
Help, please.
Thanks,
Doug
I know it will have SMK undescrypable in change the service acount of sql server manually From Laurentiu Cristofor's blog . But i don't know how what is "manual service account change", and i only know is that change service account by Configuration Manager.
any suggestions?
thanks
Hi guys,
You know when restoring a database in SQL Server, exclusive access is required. With sp_who I can see the clients. But I don't know how to terminate the client connections easily. Usually I have to stop the server and restart it. I DO think it's too troublesome. Do you have any ideas? Is there any sp or command, or tool that may help? How do you deal with this issue?
Thanks!
Hello,
I've heard rumors of best practices to include manually stopping SQL instances before doing a reboot of the system. Is this true and if so, where is this documented. I've searched all over and come up empty. So I was wondering if this was just some old DBA saying but no longer relevant?
Thanks
Blue
Hi Andrea,
I have made a table which contain data inserted manually and also data that was inserted by using bulk insertion. I have no problems using the table with Grid View. But when I try to use a query like the following:
SELECT *
FROM dbo.last
WHERE VMake = 'Honda'
AND VType = 'sedan'
AND VColor = 'Red';
I would only get the data that were inserted manually.
When I use the same query to filter data from a table that the data was inserted by using bulk insert, I get column names but no data.
Any help please.
Juvan
Dear all,
Why I always fail to manually train a model in Management Studio?
ZhaoHui Tang recommends me to untrain the model first. So I untrained it like this.
DELETE FROM Decision_Tree.CONTENT
Then I train it like this.
INSERT INTO Decision_Tree
(Age, Bike_buyer, Customer_Id, Gender)
OPENQUERY ([AdventureWorksDW], 'SELECT Age, Bike_buyer, Customer_Id, Gender FROM Training_table');
The error message is
Error (Data mining): The mining structure , Decision_Tree is already trained and does not
support incremental updates. Before using the INSERT INTO statement, use DELETE FROM <object>.
Why is that so? I already untrained the model. The model has been made and processed
outside Management Studio, i.e., in the Visual Studio.
Thank you,
Bernaridho