FillFactor Considerations

Apr 19, 2007

I have a table with about 200 million rows of data. I add a couple million rows of data each week to the table in a single load process. The table is used for reporting purposes only and there are never (not intentionally at least) any updates or deletes to the table. The data is always being added to the "end" of the table with the new AsOfDate being the main factor in the clustered index.



My question is this: Since I'm not "inserting" rows that would split pages, should I have my FILLFACTOR for the table set to 100, or am I missing something? I obviously want to save physical hard drive space, but I also don't want to slow down the import process.



BTW, I'm using SQL2000

View 1 Replies


ADVERTISEMENT

Fillfactor

Oct 26, 2000

I have 100.000 rows with 120 bytes each.
I want load 50.000 rows more.
How calculate the best fillfactor ?
Thanks you everybody.
Marc

View 1 Replies View Related

Question About Fillfactor

Aug 19, 2005

Hello,

from what I gathered, if writes on table are more often than reads, then you should use a lower fillfactor - to reduce page splits (talking about clustered index here). But what if primary key = clustered index is incremental integer, and there NEVER is any insertion into middle/rename of rows, only appends, can I use 100% fill factor then?

thanks

View 3 Replies View Related

Alter Index FillFactor

Dec 5, 2007

If you perform an Alter Index All without the FillFactor (example below) and the previous fill factor for the table and it's index were (80%), will the fillfactor be set to the previous value (80) or the sql server default index fill factor value (0 or 100%)?



ALTER INDEX ALL ON [Sales Header] REBUILD
WITH (SORT_IN_TEMPDB=ON, STATISTICS_NORECOMPUTE=OFF)

Thanks,

Ian

View 3 Replies View Related

How To Remove WITH FILLFACTOR = 100 When Generate Sql Script ?

Jul 23, 2005

Hello,I'm using Entreprise Manager (for Sql Server 2000) to generate mydatabase's script. By mistake, i've changer FillFactor one time. And,now I can't remove this data from generated sql script. How to removethat ?Thank's a lot.

View 6 Replies View Related

Network Considerations For Mirroring

Apr 19, 2007

I would like to know people's thoughts on any special network considerations to take for mirroring and the logic behind them. Is it best to segregate mirroring traffic from other network traffic? Use a VLAN? Dedicate one NIC for mirroring and the other for general network traffic or just aggregate the two and let both types of traffic share the bandwidth?



I haven't seen much in this area from Microsoft's best practices and wanted to know what those who have implemented it have done and why. There are pros and cons for each method: Letting everything share one massive pipe with load balancing vs. trying to segregate traffic in some way so that general network connections etc. do not impact the mirroring capability.



I look forward to hearing from you.



-M-

View 4 Replies View Related

Integration Services Considerations

Sep 14, 2007



We have about 150 SQL servers and basically we're considering the pros and cons of installing SSIS on a central SSIS server - that is responsible for all DTS jobs - as opposed to installing SSIS on the local SQL instance.


On the plus side so far:


1./ Central administration, alerting, change management etc

2./ Possible performance gain on the local instance not having SSIS installed?



On the negative side:


1./ Central point of failure

2./ Possibility that it would need to be a clustered...

3./ Compatibility issues may mean having to make the central SSIS server 32-bit?

4./ Possible performance cost of remote SSIS?

5./ With multiple DTS packages running at different times, when would we take the server down for maintenace...?

Would appreciate your thoughts.

View 1 Replies View Related

Database Mirroring && Replication Considerations

Jun 21, 2007

Hi all,



Can a publisher be mirrored? What are the implications, issues, gotchas? Transactional, Merge or Transactional w/ Updating Subscribers is what I'm considering.



Bottom line is I would like to use mirroring, but only one mirror will not suffice.



Thank you in advance.



Ray Nichols

View 1 Replies View Related

Considerations When Delete Records From Table.

Nov 15, 2006

Hello, I'm developing application which monitors network packets. The monitoring data are saved into table. Monitoring table maintains the data for fixed quantum time,for example during one 1 hour. So, every minute before or after insert new data, I delete the time-expired data. I doubt that the endless delete operation would results in some problems(increasing index,etc..).

Is this mechanism safe to the dbms?

Aren't there round-robin(?) style table?

View 1 Replies View Related

Instead Of Trigger Considerations For Multi-Row Insert

Apr 12, 2006

I would like to know how to, if at all possible, to reconstruct the following trigger as to be able to handle multiple row insert when a single insert command is used - because the trigger will only be called once...I'm not familiar and don't know anything about cursors and i've read that its not the best way to go.

TRIGGER ON childtable INSTEAD OF INSERT
AS
BEGIN
DECLARE @customkey char(16);
DECLARE @nextchild int;
DECLARE @parent int;
DECLARE @date datetime;

SET @date = getdate();

SELECT @parent = parenttable FROM inserted;

SELECT @nextchild=count(*)+1 FROM childtable WHERE parenttable = @parent;

IF (@nextchild >= 9998) return;

SET @customkey = €˜type€™+ convert(char(4),year(@date)) + convert(char(2),month(@date)) + convert(char(2),day(@date))+convert(char(4),@nextchild + 1);

INSERT INTO childtable (customkey,parent) VALUES (@customkey,@parent);
END

View 7 Replies View Related

Table Design Considerations For Merge Replication?

Oct 31, 2006

Can someone point me to a resource for Table Design Considerations for Merge Replication? I have an ASP.Net/SQL2K5 app that I need to run on disconnected machines, then allow dfor data sync through merge replication. I assume that the first step is getting my tables indexed in a replication friendlt manner?

Many Thanks to anyone who can point me in the right direction!

View 3 Replies View Related

License Considerations For SQL Data Source Of Reports

Dec 10, 2007

I may be overthinking this, but I want to make sure this is right. If you have a processor license of SQL Server Standard running both Reporting Service databases and the IIS interface, isn't it true that the underlying licenses of other servers containing your data are irrelivent in the context of serving the reports over the web? Example. Server 1 has SSRS as described above, processor license of Standard. Server 2 has user license of SQL Enterprise and serves data to a couple of reports on Server 1. This does not violate a license, correct? Doesn't Server 1 just take one of the CALs from Server 2?

View 1 Replies View Related

What Are The High Level Considerations For 24x7 Installations?

Feb 13, 2008

in a prior "legacy" life we couldn't imagine 24x7 implementations because it was important to 1) reorganize databases periodically to remove fragmentation that adversely affected performance and 2) back up databases just in case.

In a 24x7 SQL Server 2005 implementation, high level only, how are these and other maintenance related things accomplished with confidence?

I dont think SQL cleanses its own page splits unsolicited. Are DBAs totally reliant on logs in full recovery installations where db must be up 24x7? What if the devices those logs sit on fail? What if the logs become too large? Is it likely that if you want 24x7 you're looking at Enterrise Edition only?

I'm totally aware of and confident in the sliding window partitioning thing but it seems to me there must be more out there in terms of periodic, more frequent maintenance activity.

View 3 Replies View Related

Variable Memory Utilization And Performance Considerations

Apr 26, 2006

Can someone point me to some good articles or perhaps directly supply some words of wisdom with regard to wise utilization of variables within a T-SQL script from and standpoint of conserving memory usage and improved execution cost?

For example:

(1) Is it better to use varchars, nvarchars, etc. defined with minimal lengths to support the needs of the script or is it just as efficient to declare all with a length of say 4,000?

(2) I've seen behavior that leads me to believe that when passing a variable as a parameter in a nested procedure call, if the declared types of the parameter and the variable being passed in don't match (i.e. one is numeric(38,10) and the other is int), then implicit type conversions hurt performance. Is this true and how broadly does it apply?

(3) Does the number of variables declared in a script materially impact the performance and / or resource utlization?

(4) Is it more efficient to have a series of variable value assignments in a single SELECT statement versus a series of SET statements? Should I always perfer one to the other? Only within a looping construct?

Thanks,

Shadowraven

View 1 Replies View Related

Considerations... Backing Up IIS 6 Web Files And SQL Server 2005 Express Database

Mar 1, 2006

Anyone know of a good "free" way to back up web files and SQL Server 2005 Express Database?
I was able to use Windows Server 2003 Backup utility to back up the folder where the Databases were stored, as well as the web files, with no errors.
But I have heard a lot of discussion that you can't just simply backup SQL Server data files?
I'm wondering how sound the backup I've created is...
Any suggestions?

View 1 Replies View Related

SSIS Task Or Design Considerations For Copying/updating/replicating Tables From One Server To Another

Dec 13, 2007




I need to copy all the data from all the tables in a database to a copy of this database on another server.
What feature of SSIS should I take advantage of to accomplish this?

We have an SLA for 8am, most times the data warehousing jobs complete at 8:05am. Adding an additional process/set of tasks to this package would obviously make matters so I'm trying to update/copy/replicate the data in the fastest manner. Typically we're talking 2 marts (10-20GB) with 2 large tables (5-10 mill records) and 20 marts (0.5 - 5 GB) with many more smaller tables (~40 tables with record count ranging from 1 to a million)

Additionally please indicate if the design/feature you suggest can handle (pushing schema changes and additions to the target server) schema changes or new tablesviews added to the source database.

My only idea so far...is using the import wizard (in Management Studio) to create an SSIS package (top copy all the tables from one server to another) and saving it to the server, Then executing this package after the job is complete. However this would not work if the schema of a table changed, or if a a table is added. Moreover I don't think I can edit this package in visual studio.

View 3 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved