Log Shipping Transaction Log Backup Files Have Wrong Timestamp

Jun 23, 2007

I'm experiencing a weird problem with log shipping in SQL 2005.

I've setup Log Shipping for a production database between two sites. The standby database is being updated correctly and everything seems to be working as expected but for one detail: the name of the transaction log backups are generated with an UTC timestamp instead of my local timezone.

The the data below extracted from the backup history:

2007-06-23 17:30:00.000 D:BackupDatabasesmydbmydb_20070623073000.trn
2007-06-23 17:15:00.000 D:BackupDatabasesmydbmydb_20070623071500.trn
2007-06-23 17:00:00.000 D:BackupDatabasesmydbmydb_20070623070000.trn
2007-06-23 16:45:00.000 D:BackupDatabasesmydbmydb_20070623064500.trn

My timezone here is GMT+10.

Although it's not affecting Log Shipping, it's very confusing as the full backups have a timestamp in the local timezone!

Has anyone seen experienced something similar to this? Please see below my SQL details:

1 ProductName NULL Microsoft SQL Server
2 ProductVersion 589824 9.00.3042.00
3 Language 1033 English (United States)
4 Platform NULL NT AMD64
5 Comments NULL NT AMD64
6 CompanyName NULL Microsoft Corporation
7 FileDescription NULL SQL Server Windows NT - 64 Bit
8 FileVersion NULL 2005.090.3042.00
9 InternalName NULL SQLSERVR
10 LegalCopyright NULL © Microsoft Corp. All rights reserved.
11 LegalTrademarks NULL Microsoft® is a registered trademark of Microsoft Corporation. Windows(TM) is a trademark of Microsoft Corporation
12 OriginalFilename NULL SQLSERVR.EXE
13 PrivateBuild NULL NULL
14 SpecialBuild 199360512 NULL
15 WindowsVersion 248381957 5.2 (3790)
16 ProcessorCount 4 4
17 ProcessorActiveMask 4 f
18 ProcessorType 8664 NULL
19 PhysicalMemory 4095 4095 (4294037504)
20 Product ID NULL NULL

Thanks,
André

View 3 Replies


ADVERTISEMENT

SQL 2012 :: Local Backup Strategy On Primary With Transaction Log Shipping Enabled To Secondary DR Site

Oct 2, 2014

I have a scenario where a customer is going to be using Log Shipping to the DR site; however, we need to maintain the normal backup strategy on the current system. (i.e. Nightly Full, Every 6 Hour Differential and Hourly Transaction Log backup)I know how to setup Transaction Log Shipping and Fail-over to DR and backup but now the local backup strategy is going to be an issue. I use the [URL] .... maintenance solution currently.

Is it even possible to do regular backups locally keeping data integrity for your backup strategy with Transaction Log Shipping enabled?

View 2 Replies View Related

For Loop - Iterate From Older Files To Newer Files Based On File's Timestamp

Mar 13, 2008

In the For Loop, How to Iterate from Older flat files to Newer flat files based on File's Timestamp. If there are some older files in that folder, it should be processed first and then continue with the newer one.

Any Suggestions?

View 3 Replies View Related

SQL 2012 :: Log Shipping - Secondary Database Path Showing Wrong Location?

Mar 17, 2014

I've configured log shipping to use for DR purposes. I'm concerned that the physical location of the secondary is mis-reported by SQL Server Management Studio.

Viewing the secondary location (with Studio DB_name Properties Files) shows the path of the primary DB (I expected it to show the path of the secondary).

This SQL command shows the correct/actual paths of both primary and secondary DB's when run on their host servers.

SELECT name, physical_name AS CurrentLocation, state_desc FROM sys.master_files

Is this just cosmetic?

Here is an Example of how the Studio shows the incorrect path for the secondary.

Example:

(Primary) servername=prodSrv, DBname=aquaDB, Actual_Path=G:aquaDB, SQL-studio-Properties-Path=G:AquaDB, sys.master_files Path=G:AquaDB,

Log shipped to

(Secondary / Read Only) servername=DRSrv, DBname=aquaDB, Actual Path=F:aquaDR, SQL-studio-Properties-Path=G:AquaDB(WRONG), sys.master_files Path=F:aquaDR

View 2 Replies View Related

Processing Files With Timestamp

May 8, 2007

hi guys,

i have a unique problem here.i get four different type of flat files,which i need to pick up and process parallelly,this i am doing using four different foreach containers for the four load process as i can get multiple files ( i had other issues like rollback etc for using forecah containers).

my problem is,my input file names will have format like this
yyyymmdd_salesdataforproduct_yyyymmdd_hhmmss.txt



here the first date is bussiness date and the second one is the sysytem date..if i get multiple file i have to proceess considering the second date.but by default the foreach loop is considering the first date,what can i do to ensure that only second is used to process my files.



thanks in advance

srikanth

View 1 Replies View Related

Timestamp Error While Restoring Transaction Logs

Feb 10, 2000

I have scheduled a database backup at 9 pm and 5 Transaction Log backups. The transaction Log backups happen at 11 am, 1 pm, 3 pm and 5 pm (device does not get initialized) and one at 11 pm where i initialize the backup device. now i have another server with the same database on it. I restore these backups on to this server by simple copying the files and then doing Restore ...From Device...Add file...option. I am able to restore the DB backup but when i try to restore the log backups it get the error msg -

'Specified file 'G:...' is out of sequence. current time stamp is Feb 9 2000 3:54 pm while dump was from Feb 9 2000 3 pm.'

When i am restoring the logs i start from the 1st backup buyt still i get this error. Can someone pls let me know what the problem is and how to resolve it. Thanks.

View 2 Replies View Related

Timestamp Of Backup Is Different Than OS Time?

Jun 5, 2002

SQL7, sp3
I run a backup each night via a job, using a backup set. The question/concern is: the timestamp on the physical backup set file doesn't not change but SQL says in the history of the job that the backup is successful. I checked the system time on the server and it's current, it's todays date, but again last night's backup says May22. Should this be cause for alarm? Bug??

thanks,

View 2 Replies View Related

Log Shipping Transaction Logs.

Feb 22, 2004

Hi,

We currently have a couple a large Databases running on SQL 2000 SP3 Clustered Windows 2000 SP3 environment.

Log Shipping is enabled for both databases shipping to a Standalone SQL 2000 SP3 Windows 2000 SP3 box.

Log Shipping occurs every 15 mins with the Transaction Files on average being no more than 500KB in size. However, every now and then a Transaction Log comes through and it can be as big as 3.52GB.

Not sure why this is happening. Anyone got any ideas?

Regards

Paul Towler

View 3 Replies View Related

Log Shipping Transaction Log Question

Jul 23, 2005

I am going through a security audit on our servers. We use log shippingfor a standby database. One of the questions in the audit has melooking for answers."Are the transaction logs that are being shipped to the standbydatabase encrypted?"I am assuming no. However, I need to know definitively. I have not beenable to find an answer in BOL or in Google. If the logs are notencrypted, is there an option where I could send them encrypted, ifnecessary?Thanks,Jennie

View 2 Replies View Related

Transaction Log Shipping Over WAN With Compression

Jan 18, 2006

Hi guys,

I have a server in a datacenter (SQL 2005 ent) that collects large quantities of data from our visitors. I need to set up a secondary database in our office (different geographic location) that will server 2 purposes, 1, a backup of the database and 2, allow us to perform complex queries on the data.

There is no updating of the data on the secondary server so no changes need to go back to the primary server. A database in standby mode is fine and users on the secondary server can be disconnected when it's being updated.

I have transaction log shipping working well in a staging environment (LAN). My first question is is there any reason why transaction log shipping would not work over a WAN with a VPN connection?

And my second question is can I compress the trn files for transport over the WAN. If I manually compress the files with winzip they compress by 98%. That translates into a huge saving when I am leasing a line to transport these files.

Thanks in advance

Stephen

View 4 Replies View Related

Problem Returning A Timestamp Column Inside An TSQL Transaction

Jan 15, 2007

I cannot manage to fetch the new timestamp value inside a TSQL Transaction.  I have tried to Select "@LastChanged" before committing the transaction and after committing the transaction. A TimestampCheck variable is used to get the timestamp value of the Custom Business Object. It is checked against the row updating to see if they match.  If they do, the Update begins as a Transaction.  I send @LastChanged (timestamp) and an InputOutput param, But I also have the same problem sending in a dedicated timestamp param ("@NewLastChanged"):  1 select @TimestampCheck = LastChanged from ADD_Address where AddressId=@AddressId
2
3 if @TimestampCheck is null
4 begin
5 RAISERROR ('AddressId does not exist in ADD_Address: E002', 16, 1) -- AddressId does not exist.
6 return -1
7 end
8 else if @TimestampCheck <> @LastChanged
9 begin
10 RAISERROR ('Timestamps do not match up, the record has been changed: E003', 16, 1)
11 return -1
12 end
13
14
15 Begin Tran Address
16
17 Update ADD_Address
18 set StreetNumber= @StreetNumber, AddressLine1=@AddressLine1, StreetTypeId=@StreetTypeId, AddressLine2=@AddressLine2, AddressLine3=@AddressLine3, CityId=@CityId, StateProvidenceId=@StateProvidenceId, ZipCode=@ZipCode, CreateId=@CreateId, CreateDate=@CreateDate
19 where AddressId= @AddressId
20
21 select @error_code = @@ERROR, @AddressId= scope_identity()
22
23 if @error_code = 0
24 begin
25 commit tran Address
26
27 select @LastChanged = LastChanged
28 from ADD_Address
29 where AddressId = @AddressId
30
31 if @LastChanged is null
32 begin
33 RAISERROR ('LastChanged has returned null in ADD_Address: E004', 16, 1)
34 return -1
35 end
36 if @LastChanged = @TimestampCheck
37 begin
38 RAISERROR ('LastChanged original value has not changed in ADD_Address: E005', 16, 1)
39 return -1
40 end
41 return 0I do not have this problem if I do not use a TSQL Transaction. Is there a way to capture the new timestamp inside a Transaction, or have I missed something?Thank you,jspurlin  

View 1 Replies View Related

Log Shipping Dies Files Set To A

Aug 21, 2006

I have log shipping setup at a client site (it is their network). Log shipping runs fine until midnight, and then stops processing files. What I notice on the primary server (where the files are backed up to), all the files after midnight are set to A attribute, none of the files before midnight have this. The short of it it does not process these files and log shipping dies and has to be rebuilt. Any network admins have any ideas (I did have them stop backing up these files with their backup software) still dies

View 3 Replies View Related

Log Shipping: Clean Up Log Files

Mar 3, 2004

Hi,
I need some advise in the log shipping. The log files in the primary server get cleaned up according to what I have specified in the maintenance plan. But the log files that got shipped to the secondary server stay there for ever wasting my hard disc. Will it make any problem if I remove them or can I set it up to remove all files earlier than past 2 hrs? Please advise.
Thanks

View 6 Replies View Related

Transaction Log Shipping Script Errors

Dec 10, 2007



Greetings:
When I script out my log shipping configuration from the GUI and subsequently drop the log shipping and try to recreate it with the created script, the backup and restore functions do not seem to be working; please see script below. Is there an additional step (or steps) that the SSMS GUI does not output when it creates the script for log shipping? I noticed in the GUI after I run the script that the destination folder for copied files is blank as well.

Example error from backup/restore job - Error: The path is not of a legal form.(mscorlib)


-- Execute the following statements at the Primary to configure Log Shipping

-- for the database [rdevsql2].[SymbolLookUp],

-- The script needs to be run at the Primary in the context of the [msdb] database.

-------------------------------------------------------------------------------------

-- Adding the Log Shipping configuration

-- ****** Begin: Script to be run at Primary: [rdevsql2] ******



DECLARE @LS_BackupJobId AS uniqueidentifier

DECLARE @LS_PrimaryId AS uniqueidentifier

DECLARE @SP_Add_RetCode As int



EXEC @SP_Add_RetCode = master.dbo.sp_add_log_shipping_primary_database

@database = N'SymbolLookUp'

,@backup_directory = N'm:ackups'

,@backup_share = N'\rdevsql2m$ackups'

,@backup_job_name = N'LSBackup_SymbolLookUp'

,@backup_retention_period = 60

,@monitor_server = N'RDEVSQL1'

,@monitor_server_security_mode = 1

,@backup_threshold = 60

,@threshold_alert_enabled = 1

,@history_retention_period = 60

,@backup_job_id = @LS_BackupJobId OUTPUT

,@primary_id = @LS_PrimaryId OUTPUT

,@overwrite = 1

,@ignoreremotemonitor = 1



IF (@@ERROR = 0 AND @SP_Add_RetCode = 0)

BEGIN

DECLARE @LS_BackUpScheduleUID As uniqueidentifier

DECLARE @LS_BackUpScheduleID AS int



EXEC msdb.dbo.sp_add_schedule

@schedule_name =N'LSBackupSchedule_rdevsql21'

,@enabled = 1

,@freq_type = 4

,@freq_interval = 1

,@freq_subday_type = 4

,@freq_subday_interval = 1

,@freq_recurrence_factor = 0

,@active_start_date = 20071207

,@active_end_date = 99991231

,@active_start_time = 0

,@active_end_time = 235900

,@schedule_uid = @LS_BackUpScheduleUID OUTPUT

,@schedule_id = @LS_BackUpScheduleID OUTPUT

EXEC msdb.dbo.sp_attach_schedule

@job_id = @LS_BackupJobId

,@schedule_id = @LS_BackUpScheduleID

EXEC msdb.dbo.sp_update_job

@job_id = @LS_BackupJobId

,@enabled = 1



END



EXEC master.dbo.sp_add_log_shipping_primary_secondary

@primary_database = N'SymbolLookUp'

,@secondary_server = N'RDEVSQL1'

,@secondary_database = N'SymbolLookUp'

,@overwrite = 1

-- ****** End: Script to be run at Primary: [rdevsql2] ******



-- ****** Begin: Script to be run at Monitor: [RDEVSQL1] ******



EXEC rdevsql1.msdb.dbo.sp_processlogshippingmonitorprimary

@mode = 1

,@primary_id = N'4d80db8c-e090-4dc0-8af6-d5f5802c4207'

,@primary_server = N'rdevsql2'

,@monitor_server = N'RDEVSQL1'

,@monitor_server_security_mode = 1

,@primary_database = N'SymbolLookUp'

,@backup_threshold = 60

,@threshold_alert = 14420

,@threshold_alert_enabled = 1

,@history_retention_period = 60

-- ****** End: Script to be run at Monitor: [RDEVSQL1] ******



-- Execute the following statements at the Secondary to configure Log Shipping

-- for the database [RDEVSQL1].[SymbolLookUp],

-- the script needs to be run at the Secondary in the context of the [msdb] database.

-------------------------------------------------------------------------------------

-- Adding the Log Shipping configuration

-- ****** Begin: Script to be run at Secondary: [RDEVSQL1] ******



DECLARE @LS_Secondary__CopyJobId AS uniqueidentifier

DECLARE @LS_Secondary__RestoreJobId AS uniqueidentifier

DECLARE @LS_Secondary__SecondaryId AS uniqueidentifier

DECLARE @LS_Add_RetCode As int



EXEC @LS_Add_RetCode = rdevsql1.master.dbo.sp_add_log_shipping_secondary_primary

@primary_server = N'rdevsql2'

,@primary_database = N'SymbolLookUp'

,@backup_source_directory = N'\rdevsql2m$ackups'

,@backup_destination_directory = N''

,@copy_job_name = N''

,@restore_job_name = N''

,@file_retention_period = 4320

,@monitor_server = N'RDEVSQL1'

,@monitor_server_security_mode = 1

,@overwrite = 1

,@copy_job_id = @LS_Secondary__CopyJobId OUTPUT

,@restore_job_id = @LS_Secondary__RestoreJobId OUTPUT

,@secondary_id = @LS_Secondary__SecondaryId OUTPUT

IF (@@ERROR = 0 AND @LS_Add_RetCode = 0)

BEGIN

DECLARE @LS_SecondaryCopyJobScheduleUID As uniqueidentifier

DECLARE @LS_SecondaryCopyJobScheduleID AS int



EXEC rdevsql1.msdb.dbo.sp_add_schedule

@schedule_name =N'DefaultCopyJobSchedule'

,@enabled = 1

,@freq_type = 4

,@freq_interval = 1

,@freq_subday_type = 4

,@freq_subday_interval = 15

,@freq_recurrence_factor = 0

,@active_start_date = 20071207

,@active_end_date = 99991231

,@active_start_time = 0

,@active_end_time = 235900

,@schedule_uid = @LS_SecondaryCopyJobScheduleUID OUTPUT

,@schedule_id = @LS_SecondaryCopyJobScheduleID OUTPUT

EXEC rdevsql1.msdb.dbo.sp_attach_schedule

@job_id = @LS_Secondary__CopyJobId

,@schedule_id = @LS_SecondaryCopyJobScheduleID

DECLARE @LS_SecondaryRestoreJobScheduleUID As uniqueidentifier

DECLARE @LS_SecondaryRestoreJobScheduleID AS int



EXEC rdevsql1.msdb.dbo.sp_add_schedule

@schedule_name =N'DefaultRestoreJobSchedule'

,@enabled = 1

,@freq_type = 4

,@freq_interval = 1

,@freq_subday_type = 4

,@freq_subday_interval = 15

,@freq_recurrence_factor = 0

,@active_start_date = 20071207

,@active_end_date = 99991231

,@active_start_time = 0

,@active_end_time = 235900

,@schedule_uid = @LS_SecondaryRestoreJobScheduleUID OUTPUT

,@schedule_id = @LS_SecondaryRestoreJobScheduleID OUTPUT

EXEC rdevsql1.msdb.dbo.sp_attach_schedule

@job_id = @LS_Secondary__RestoreJobId

,@schedule_id = @LS_SecondaryRestoreJobScheduleID



END



DECLARE @LS_Add_RetCode2 As int



IF (@@ERROR = 0 AND @LS_Add_RetCode = 0)

BEGIN

EXEC @LS_Add_RetCode2 = rdevsql1.master.dbo.sp_add_log_shipping_secondary_database

@secondary_database = N'SymbolLookUp'

,@primary_server = N'rdevsql2'

,@primary_database = N'SymbolLookUp'

,@restore_delay = 0

,@restore_mode = 0

,@disconnect_users = 0

,@restore_threshold = 45

,@threshold_alert_enabled = 1

,@history_retention_period = 60

,@overwrite = 1

END



IF (@@error = 0 AND @LS_Add_RetCode = 0)

BEGIN

EXEC rdevsql1.msdb.dbo.sp_update_job

@job_id = @LS_Secondary__CopyJobId

,@enabled = 1

EXEC rdevsql1.msdb.dbo.sp_update_job

@job_id = @LS_Secondary__RestoreJobId

,@enabled = 1

END



-- ****** End: Script to be run at Secondary: [RDEVSQL1] ******


help is much appreciated,
Derek


View 1 Replies View Related

DB Re-index Explodes Log Shipping Files

Mar 14, 2002

Hi,

Is it a recognised limitation of log shipping over a WAN, that if you re-index the primary db then the result is gigabytes of trn log to send over the WAN, effectively breaking the log shipping sync?

Are there any known work-arounds?

Would anyone recommend using db replication instead to maintain a DR sql server?

Any comments greatly appreciated.
Steve

View 1 Replies View Related

Recovery :: Log Shipping Transaction Log File Corrupted

Aug 7, 2015

Log shipping was configured 6 months back. A Transaction log file got corrupted today. How to resolve this?

View 20 Replies View Related

Slow Transaction Log-Shipping Restores (SS 2005)

Sep 21, 2007

Hello,

We have log-shipping set up between a source and 3 destination SS 2000 databases. Two of the destination servers actually perform their log restores across the network from the other secondary server. This allows us to only copy the files once from a remote location. All three servers stay caught up within 15 minutes of each other.

Recently, I added a fourth server to this that has SS 2005 SP2 (X64). I wrote a stored procedure that restores log backups from the same single location as the maintenance plan jobs. The problem that I'm experiencing is that this fourth server is not keeping up with the other three. It seems to take longer to restore the same log backups. The destination servers are all on the same domain. This fourth server was previously part of the same maintenance plan configuration as the others prior to rebuilding it for SS 2005 SP2 (X64). During that time, it stayed caught up with the other servers. There is another database on the new server that I am log-shipping to in the same manner and it stays caught up, though, for the most part, the log backups are smaller. There is a file on the fourth server with a ckp extension for the database in question that doesn't seem to exist for the other databases on this server and the other servers.

Any information on this behavior would be appreciated.

View 1 Replies View Related

Transaction Log Shipping Secondary DB In Restoring... State

Apr 23, 2008

I am new to this environment and was asked to ensure that the transaction log shipping for SQL 2005 on W2K3 boxes is working properly. I noticed the db's on the secondary server are show "Restoring..." I am not sure if these were set up in No Recover Mode or Standby Mode. I have no access to the secondary db's. I get an error message when trying to access them (error 927). Monitoring was not set up initially and as you may or may not know can't be turned on after the fact...unless you delete the job and start over.

My question is is "Restoring..." normal and what does it indicate?

View 3 Replies View Related

Setting Up Transaction Log Shipping In Prod Environment

Aug 21, 2007

Hi,

I currently have a 2000 Ent. production server and a stand by server ready for transaction log shipping.

Is it possible to setup transaction log shipping on a live environment without any interruptions?

I'm currently backing up the log every 1 hour, I'd like to increase to 15 minutes.

Any help would greatly be appreciated.

Thanks,
- Gary

View 4 Replies View Related

Transaction Log Size Increases Enormously(log Shipping)

Feb 4, 2008

Hi,

Im in the process of setting up logshipping on sqlserver 2005 enterprise edition.
My scenario is like this:
My Avg size of my tlog is 500MB and im planning to set the log shipping at 30mins interval(ie backup job schedule,Copy,restore job schedule).But at some part of the day the Tlog suddenly increases up to 1.5GB - 2 GB .So i wanted to know, wht if that 1.5GB-2GB tlog file is unable to get backed up,copy and restore at 30mins interval?.How to deal with this kind of issues where the size of tlogs are increased suddenly.I cannot do it at15mins interval due to some network restrictions at my office.

i have One more doubt about the setting on Logshipping screen:
Now let us suppose my settting on log shipping screen 'Alert if no restore occurs within' is set to '180mins', then does this setting mean that the restore job will keep on looking for the copied file in the folder on secondary for next 90mins and if its not able to find any, it will generate an alert after 90mins ??? or it will generate an error if its nt able to find any copied file after the first restore job execution.???
in the same way,

Thnx in advance for any help.

Regards
Arvind L

View 1 Replies View Related

Recovery :: Point New Log Shipping Files To Different Drive

Jun 9, 2015

I have two databases on a Production Server that I want to Log Ship to a Test Server.  According to the sys.master files the physical File Location is on an E drive. Early attempts at Log Shipping these two files error'd out due to space issues on the E drive (one Log Shipped and then one err'd out).  I was subsequently informed from the server group that they would prefer that I Log Ship these two database files over to the M Drive where more space is available.  In fact, they modified the Server Properties / Database Settings / Database Default Locations (for Data and Log) to the larger M drive (I'm not really sure why they just don't increase the E drive space but there is proabably a good reason).

Okay, so now my problems have been solved.  Easy enough.  Now I deleted the successful Log Shipped database and started from scratch.  However, as before, one db restored and one failed (due to space issues).  Apparently, both db are pointing towards the E drive.  How is that possible? 

So here I am with one successful database and the normal sys databases pointing to the E drive.  What is the best way of approaching this  move to the larger and preferred M drive?

View 6 Replies View Related

SQL 2012 :: FOR FILES Command To Delete Old Backup Files On Remote Server?

Feb 24, 2015

I have the need to delete old backup files via TSQL job. Found this solution online:

PushD "
emoteservershareDIFF" &&(
forfiles -m *DIFF*.sqb -d -1 -c "cmd /c del /q @path"
) & PopD

It works remotely if I run it via command prompt. But when I add this to a TSQL job on my remote SQL instance, it runs without deleting anything. What I'm missing?

View 6 Replies View Related

Recovery :: Log Shipping Transaction Log Backups Not Being Deleted On Primary

Oct 12, 2015

I've got log shipping set up, and everything seems to be working fine, but the log files are not being deleted from the primary server despite configuring log shipping to retain them for 3 days.  I see no errors concerning the log shipping, but did not configure a monitor. What process is responsible for deleting the older log backups, and how can I look for errors.  I could simply set up a jog to delete the older files, but that will only mask the issue.

View 3 Replies View Related

Recovery :: Validate Data In Transaction Logs Shipping

Jul 16, 2015

Out of using stored procedure, reports and all this staff, I want to know the possible way to make sure that the data inside my Secondary Server Read only database are same as data in my primary server database.

So what is the simple way to do this check?

View 4 Replies View Related

Full Backup In Log Shipping

Dec 5, 2007

Hi all,


We have log shipping set up for some our databases.


So, my question is, Can i take a FULL BACKUP of databases which are involved in log shiping at primary.


--> Please give your comments on this.


Thanks.

View 4 Replies View Related

SQL Security :: Full Backup Needed After Restoration Of Database Before Transaction Log Backup

Jul 15, 2015

We take a full backup in the early morning and hourly transaction log back during the working hours for one database in the production server. The application team made certain changes to the design of the said database in their development server. The backup from the development server was restored to the production server during working hours. After the restoration should we take a full backup before next transactional logbackup? Would the transactional log backup with out a full backup after the restoration of a database be valid?

View 5 Replies View Related

TLog Backup Then Start Log Shipping

Apr 9, 2008

Hello,

I have an SQL 2005 maintenance plan that backups up the t-logs of my production db every 15 minutes. I want to begin log shipping over to a warm-standby secondary server. The network share, log shipping folder is not the same folder location as my maintenance plan folder. My question is, do I need to disable the maintenance plan for tlog backups in order for log shipping to be sucessful? Or will the dual backups to seperate backup folder locations cause a failure in the secondary server restore process?

I suspect the latter but wanted a second opinion.

Thanks.

View 4 Replies View Related

Delete Log Shipping Initial Backup?

May 29, 2008

Does anyone know if you can delete the initial backup file created (xxx_init.bak) on the secondary server after setting up Log Shipping?

View 1 Replies View Related

Log Shipping - Backup Job Missing On Primary

Jan 17, 2008



Hi everyone,

I'm trying to configure log shipping on a 2005 sql server. I follow the wizard's instructions (see http://technet.microsoft.com/en-us/library/ms190640.aspx) and everything looks right except for the backup job that somehow is not being created on the primary server. Secondary server contains copy, restore and alert jobs.

No errors reported on job monitor.

Any ideas?

Regards,

View 9 Replies View Related

Tail Log Backup + Log Shipping Fail-over

Jul 14, 2007

Hi!

i have a question regarding tail log backup.
I have set up logshipping between primary and secondary servers. Now when i want to fail-over to secondary, I am planning to take the tail log backup on primary

backup log primarydb
to disk=N'..',
with norecovery


and then I will restore this tail log backup to secondary

restore log secondarydb
from disk=N'',
with recovery

Will there be any issue with this approach?

Will I be able to recovery the primarydb if i want to, in future? Taking tail log backup doesnt cause any issues? Or do i need to take the full backup of primary before I take tail log backup..in case if I can not recover the primary db after taking the log backup?

How big the tail log backup can be? Will it be of similar size of other normal transactional log backups.?

Any help would be appreciated.

Thanks

View 3 Replies View Related

SQL Server 2008 :: Count Number Of Backup Files Deleted From A Default Backup Location

Mar 13, 2015

I've written a custom script to delete backup files from location. But unable to modify now to count the number of files are deleted. How to modify the script...

/* Script to delete older than N days backup from a specific directory */

USE [db_admin]
GO
IF OBJECT_ID('usp_DeleteBackup', 'P') IS NOT NULL
DROP PROC usp_DeleteBackup
GO

[Code] .....

View 2 Replies View Related

SQL 2012 :: Can See Last Copied Files In Log Shipping Monitor Reports As Well In Proc

Jan 30, 2015

At one of our client side a wired log shipping issues has come up.while monitoring those two server i noticed that although log-shipping report says both server are in sync, report displays information related to both backup and restore , it doesn't shows information related to copy that is when was last file copied and last file copy column is showed blank. Same is when i execute proc "sp_help_log_shipping_monitor" . I get same result ...

When i expand copy job history to analyse its what i found is although job has executed successfully , but in depth reading each steps says that no .trn file was copied .

My copy directory is at secondary server itself where .trn files are placed.And from this location itself files are begin restored.

SQL server and agent on both servers are run by same domain account ....

View 5 Replies View Related

SQL Server 2000 : Log Shipping, Backup Plan

Dec 8, 2006

Hey all-

I am working to establish a backup plan for SQL Server 2000. We are currently doing log shipping between servers.

A full backup once a week
transaction log backup every hour

We run a bulk deletion (delete * from table where date < ninetydays) which deletes 22,000 rows approximately. We are running in Full Recovery mode.

I wonder if there was anyway to improve my backup plan, preferably minimizing logging when the deletion takes place but still backup the data with log shipping.

Let me know if any of you have ideas. Thanks

Nate

View 1 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved