Storage Snapshot
Jul 20, 2005hello group,
is it possible to do a storage snapshot of a running ms-sql database
without losing transactions?
What tasks must be done before such a snapshot.
thanks in advance,
Bernhard
hello group,
is it possible to do a storage snapshot of a running ms-sql database
without losing transactions?
What tasks must be done before such a snapshot.
thanks in advance,
Bernhard
How to implement distinct storage tiers on SQL Remote BLOB Storage (RBS)?
I want to use this SQL Feature to move files(images, videos, pdf files) from a database to a distinct database dedicated to RBS. Then I want to have several storage tiers, where objects will be saved and moved according access frequency. Old data will be arquived in cheap storage, but it must be always accessible if needed.
Description:
- 1st and main tier: new and frequently accessed objects stored in high performance storage;
- 2nd tier: automatically move older or less accessed objects to an inexpensive and different storage tier;
- in all cases, all objects must be accessible to all users, but accessing to archived objects(2nd tier) will be much slower;
I am a Windows developer for the IBM Tivoli Storage Manager Server (TSMS) product.
Our product installation is built with InstallShield and uses the Windows Installer.
On a new installation of Windows 2003 x64 Storage Server R2, at a customer's site, the TSMS product fails to install.
The install of the OS has version 3.01.400.3959 of the Windows Installer and I see no newer version that installs.
Part of our product is 32 bit (console) and another part is x64 (server).
When installing I can see that the install's default is being redirected/reset to C:Program Files (x86)TivoliTSM after it is explicitly set by a custom action to ..Program Files.. . I further observe that our custom actions to write 64 bit registry entries are being refused.
REGSAM samMask = KEY_ALL_ACCESS;
if ( regIsWow64Process () ) samMask = samMask | KEY_WOW64_64KEY;
lStatus = RegCreateKeyEx( hLocalConnectKeyRoot,
szSubkey,
0L,
NULL,
REG_OPTION_NON_VOLATILE,
samMask,
NULL,
hKey,
&dw ) ;
The above fails to create the key.
We have tried four versions of our TSMS spanning many changes but the install acts the same.
This does not happen on any other Windows OS we test on but we do not test on Windows 2003 Storage Server R2 being that it is an OEM product. We did test on Windows server 2003 R2 x64 and do not see this problem.
Do you have any suggestions on how to tackle this problem?
I have full installation traces but can only see that the registry work is being refused. I can't see why.
Hi
I seem to have a strange problem when applying a snapshot when the tables in the publication have been updated while the snapshot was being generated.
Say for example there is a table called RMAReplacedItem in the publication. When the snapshot starts being applied to the subscriber, a stored procedure called sp_MSins_RMAReplacedItem_msrepl_css gets created that handles an insert if the row already exists (ie it updates the row rather than inserting it). However, after all the data has been loaded into the tables, instead of calling this procedure, it tries to call one called sp_MSins_RMAReplacedIte_msrepl_cssm - it takes the last letter of the table name and adds it to the end of the procedure name.
The worst part is that this causes the application of the snapshot to fail, but it doesnt report what the error is, and instead it just tries applying the snapshot again. The only way i have managed to find which call is failing is to run profiler against the subscriber while the snapshot is being applied and see what errors.
I have run sp_broswereplcmds and the data in there is what is applied to the subscriber - ie the wrong procedure name.
All the servers involved are running sql 2005 service pack 2. The publisher and subscriber were both upgraded from sql 2000, but the distribution server is a fresh install of sql 2005.
Hello Everyone and thanks for your help in advance. I am developing a document storage application for an intranet that will store various Word, Excel, and PDF documents. Most of the examples I see utilize SQL Server and an image field rather than the FileSystem Object to store documents. My concern with this method is that some of the documents may be several hundred pages (not exactly sure of the actual file size yet, but they must be fairly large). My question is, where does the use of SQL Server become impractical for this type of application? Any insight would be greatly appreciated. Thanks.
View 1 Replies View RelatedHello,
Does anyone know the upper limit of data size that one SQL 2K table can hold. I've seen 50GB tables in some warehousing servers, but what is the true limit. Soes the limit vary with the SQL2k version?
Thanks!
I have an MSDE installation on Windows server 2003. It looks like the C: drive is taking the brunt of the data when I load up the database. I would like to specify a different drive for data...Is there a way to do this?
View 1 Replies View RelatedHow should i know size of the table in the DB. suppose my DB has 5 tables and the size of the DB is 500 MB. How can I know size of the indivdual table.
Thanks.
greetings!!!
Help me please!
I'm migrating a images DB of a system
I know the structure of the data tables and all type of data in it
How can I learn about the STORAGE of IMAGES? In sql Server
Where can I found information about that?
I need to know something about that topic
usually, whats the way for image’s storage ?
Hi guys.
I am currently developing a system thats stores exchange stats in a db. Since our customers are companies with 20 employees up to 5 000 there a a big difference in the volume of data needed to be stored.
We currently thinking of supplying a SQL Server Express DB to the small customers and suggest a SQL Server to the bigger.
But since I would like to use the same structure for both types of customers I wonder how should i design the storeage.
Since the could be from 500 records a day up to 20 000. There are quite simple recordes with only simple datatypes. about 15 fields with no more than 10 chars each, mostly 2.
Should i separate the data in diffrent tables for a week or a day etc.
Since I am only going to filter data on 1 or 2 fields the data will be easly indexed.
The reports generated will almost always only use 1-3 months of data, but historical reports have to be possible.
My question are ofcourse:
Whats the best solution for me?
Thanks in advance:)
/Johan Wendelstam
Sweden
In MSQL Server 2000 how can I expand or use multiple transaction logs because the hard disk i am using windows dont have more than 4 GB free and the query i want to run overcomes this space.
I have another one HDD with 20-30 GB free space and i want to use this disk so to use a second transaction log or move this log to this disk.
Can this be happen and how ????
Thank you in advance
additional to data, what other type of information can be store in sql databases, i need to store pictures and mp3's that can be done, if not do you know what storage can be used for this purpose?
View 2 Replies View RelatedI have recently designed and built my first database using SQL server 2005 express. I have included an image (BLOB) column in one of the database tables. This is a bad idea according to some experts, and some say it is OK!
I am currently carrying out a trial with just 3 pictures via Visual Basic 2005 express forms, and there is no problem so far as the images are displayed for each record. But I anticipate between 300 - 1000 images for the table, and this could pose real problems for SQL server 2005 express and Visual Basic 2005 express, I guess.
I have just been reading that the cost of storing large images in the database is too high! I have also read it's better to store images (BLOB) into the file system because it is cheaper to store them no matter how many there are.
But the question is how I can reference an image in this path: C:PictureProductGrocery 0052745.jpg in the database table, so that when I select a record Visual Basic 2005 forms the image is displayed accordingly, similar as when stored directly in the database table? Your help very much appreciated.
From what I've read, if a row contains more than 8060 bytes and has varchar(MAX) columns in it, the data in those varchar(MAX) columns will be stored off-row. But what happens if you have two varchar(8000) columns instead and both contain more than 4030 bytes, is the data for both stored off-row? If so, just for that row, or for all rows in that table? And is there ever a good reason to have two varchar(8000) columns in a SQL Server 2005 table, instead of using varchar(MAX)?
View 6 Replies View RelatedHey,
i don't see in my sql server 2005 the 'storage' under
Databases/MyDatabase/STORAGE for the full text catalogs.
How can I activate it? On an other computer I'm running the same version of my sql server 2005 i can see the storage label to click on...
I've searched so long without any mathces. Hope you can help me!!!
Thanks,
Steven
Hi All
from following link
http://msdn2.microsoft.com/en-us/library/ms143432.aspx
i got this information
Bytes per row
8,060
Bytes per varchar(max), varbinary(max), xml, text, or image column
2^31-1
Characters per ntext or nvarchar(max) column
2^30-1
so is it means we can store in nvarchar(max) or varchar(max) upto 2^31-1 characters/bytes. but on the other hand we can store only 8060 bytes in a row. so the concept of Row-Overflow Data Exceeding 8 KB comes into picture.
my questions is suppose i declare a variable of type varchar(max) then i can store upto 2^30-1 bytes in it and SQL Server on its own determine how and where to stores it.
Hi,
i want to know how many data can storage into sql server compact edition. I've got a db into a pocket pc that has a table with about 2000 records inside; are they too records?
it sounds like a column can be added to each row in a table that is the checksum or binary_checksum of an expression. How many bytes do each of these occupy? Does the answer depend on the number and/or length of items in the expression?
View 4 Replies View Relatedi am using compact framework do create applications now i am facing a problem....
i am getting a "Not enough storage is available to complete this operation" exception
i am testing my application on O2 pocket pc. running Sql mobile edition.......... and this error is causing me trouble this is espically happening when i am opening sql server connection or calling complex joined queries..... or so... any ideas???
I am new to windows mobile development, and I am looking to create an application (C#) which uses the SQL server CE. I am curious what to do about data loss. I have a Windows Mobile 2003 PDA, which means half of my storage is main memory (volatile) and the other is built in storage (no volatile).
I expect I would want to install my application to the main memory, along with the database. I Planned to give the user the abililty to select a subset of properties for running the application and storing them in the DB for running. But if the user's battery runs out I wouldn't want them to loose all of the information in the database (ie configurable property selections chosen by the user).
If the user looses the application, I plan to have them reinstall it (which would recreate the database). But I would like them to be able to load a 'configuration' file from their Built in storage area when the app is reinstalled, so that they don't loose all of their settings.
Expererience with this? or reccomendations on how to go about it?
hi all,
I have a field which name is Information
and it type is Varchar (8000),but some time data access than 8000 character, my client told me,make this field to store Unlimited data.
So how can i achive this task, i m using VS 2003 (ASP.NET with VB.NET) with SQL 2000.
Thanks
Shally
Does anyone know if SQL7 will work with Storage Area Networks(SAN's)? I've read that SQL2000 implements something called a Virtual Interface System Area Network (VI SAN) that allows communication with devices connected via a SAN.
My site is installing a SAN and I need to know if SQL7 can utilize those resources (Storage,etc) and how reliable if so.
Randy
Hi,
I am having trouble trying to import a big file (aprox 250Mb is size) into an SQL Server database and I keep getting the message:
"Not enough storage is available to complete this operation".
The application tries to import the file by executing a stored procedure:
CREATE PROCEDURE sp_updateMaterialBlob
@MaterialId Int,
@BLOB image
AS
BEGIN
Update Material SET blob = @blob where id = @materialId
END
The application uses an ADO connection. I've tried increasing the memory of the client machine but that didn't work. Whenever I do run the import, nearly all the memory on the machine is used up but every time after several hours I get the same error message. What is the cause of the problem and how do I resolve it? Ideally I want to use my application to do the import rather than anything bespoke.
Can anyone explain to me how a column defined with a "bit null" datatype is physically stored in MSSQL? Is it stored like a "tinyint null" physically? In other words, how many bytes on the row on the page does a "bit null" datatype consume (assuming a non-null value 0, or 1 is the current value).
Is there any good documentation about the physical storage layout for a data page?
Thanks -
I am using Delphi 7 and need to save mulitple documents that range from 2K KB to 200K KB. These need to be stored in the database for inclusion in the database backup.
When i am working on the 200K KB file i continusely run out of memory, how can i resolve this.
I'm in the situation where we are suffering of poor performance on our SAN storage (VPlex) but it is mainly due to the quantity of data of different types which are on it (other applications, other I/O profile, bad storage usage...). As we plan to dedicate an ESX for SQL Server, we decide to have a new storage type. So we will go with NETAPP Clustered Data ONTAP on NAS technology.
Storage team want to enable only NFS protocol, so I'm wondering how SQL Server will handle that ? I read that NFS wasn't optimize for SQL Server and that block level (SAN) should be preferred.
Just curious to know what you think.
View 7 Replies View RelatedI need to bulk insert very large amount of data into several MSSQLtables.The first Data model definition used identities to mantain relationshipbetween those tables but we found that natural keys (compound) arebetter forbulk insert (there is no need to obtain the identity first)My question is, changing the identities to natural keys (in some tablesinorder of 4, 5 attributes) will enlarge my database storage?I think MSSQL implements relationships with pointers (or hashcodes), sothestorage size will be similar, right?Regards,
View 3 Replies View RelatedHi All,As per BOL, XML data type can store up 2 GB of data.My question is when a row is inserted in a table, for its xml column,2GB of space will be resered.In other words, how xml is internally stored. Is storage allocation issimilar to varchar(max) data type?Thanks in advance for everything.
View 1 Replies View RelatedGreetings.I need to convert some columns of type numeric(12, 0) to hold floating pointinformation scale and precision I can't determine in advance (customer datacan vary wildly) so I wanted to use the datatype that offers the maximum scaleand precision..I'm targeting sql server 2005 systems (not 2000).It seems my choices are real and float, and the docs seem to indicate thatfloat offers with widest ranges.I'm trying out using the 'float' for the new data type as the defaultprecision is said to be 53.. Does this mean the total number of digits is upto 53?I don't know if there is anything else I need to take into account since thesetwo columns are part of a primary key, and I supposed, therefore, are indexed.thanksJeff Kish
View 6 Replies View RelatedHi ,This is related to datawarehouse , data mining . We are told that datamining tools such as spss or sas , need a large table which has lotsof columns inside . Based on our project , out final table which willbe used in data mining , has lots of computed values . The finalnumber of columns of the table is 7800!(lots of calculated values)First of all , i dont see the reason of this requirement of the datamining tools. I will be appreciated if someone can clarify this: Whydoes a data mining tool need such a large table?The main issue is if i create table with 7800 columns inside ( in factoracle only allows to create table with 1000 columns) , i believe itcant be queried.My basic calculation shows that the average row size of this tablewill be 160kb. , considering my db_block_size of 16kb. this means 10blocks for 1 row. (The table will have more than 10.000.000 rows) Nomatter how fast my disk subsystem is , i think the queries againstthis table will fail. So what can i do? May be I need a differenttype of storage technique for instance column based storage ( i heardthat sysbase has this feature , dont know the details / purpose..)How can i solve such a problem? The database server really does notmatter , it can be oracle,sqlserver , sybase,informix , etc... I willbe appreciated if someone can help me abut the issue.Kind Regards,hope
View 1 Replies View RelatedHi!
I'm investigating if it is an good idea to use some form of "bitmap index",
e.g. I have an normal table with 100-200 columns, but sometimes I just want to know:
"Is there an value in the cell" (...where col20 is not null)
So I will measure if I get some benifit, so If you tried this, let me hear, but otherwise see only the rest:
If I remember correctly SQL Server will group bit-columns in some way minimizing storage needs,
but my questions: how much storage is needed for an bit column:
1 or 2 bit (the second bit, because we can have three states: 0,1 and null)
Will 8 bit-columns need 1 byte or 2 byte?
I tried some tests with a number table (1mio rows), but It didn't give me the answer.
Best Regards
Bjorn
create table tmpbit (a bit)
create table tmpbit8 (a1 bit, a2 bit, a3 bit , a4 bit, a5 bit , a6 bit, a7 bit, a8 bit)
insert into tmpbit
select cast(1 as bit) from numbers
insert into tmpbit8
select cast(1 as bit) , cast(1 as bit) , cast(1 as bit) , cast(1 as bit) ,
cast(1 as bit) , cast(1 as bit) , cast(1 as bit) , cast(1 as bit)
from numbers
select object_id('tmpbit8') -- for input to dbcc
DBCC SHOWCONTIG scanning 'tmpbit' table...
Table: 'tmpbit' (1269891891); index ID: 0, database ID: 6
TABLE level scan performed.
- Pages Scanned................................: 1359
- Extents Scanned..............................: 171
- Extent Switches..............................: 170
- Avg. Pages per Extent........................: 7.9
- Scan Density [Best Count:Actual Count].......: 99.42% [170:171]
- Extent Scan Fragmentation ...................: 1.75%
- Avg. Bytes Free per Page.....................: 1.8
- Avg. Page Density (full).....................: 99.98%
DBCC SHOWCONTIG scanning 'tmpbit8' table...
Table: 'tmpbit8' (1301892005); index ID: 0, database ID: 6
TABLE level scan performed.
- Pages Scanned................................: 1359
- Extents Scanned..............................: 171
- Extent Switches..............................: 170
- Avg. Pages per Extent........................: 7.9
- Scan Density [Best Count:Actual Count].......: 99.42% [170:171]
- Extent Scan Fragmentation ...................: 0.58%
- Avg. Bytes Free per Page.....................: 1.8
- Avg. Page Density (full).....................: 99.98%
both tmp-tables uses 10,617 MB
hello forum friends,
i need to know where the database table stored in,
where to find the data table without distrubing the
sql server editor.whtr it is possible to get it from
"C:Program FilesMicrosoft SQL ServerMSSQLData"
or some where else.
Regards
prathap