SQL 2012 :: Model Database Files Missing - Inaccessible Or Corrupt
Aug 12, 2015
Been practicing DR strategies with a test SQL instance by following the scenarios listed here: [URL] ....
> Took a backup of the Model database
> Stopped SQL Server
> Deleted model database data & log file
> Started SQL Server and it obviously wouldn't start because TempDB needs a model database present.
> Started SQL instance with trace flags 3608 & 3609
> Connected to SQL instance using command prompt.
> Issued restore command but was met with this error:
Shared Memory Provider: The pipe has been ended.
Communication link failure
And found this in the SQL log..
2015-08-12 16:21:32.83 spid51 Starting up database 'tempdb'.
2015-08-12 16:21:36.88 spid51 Error: 3456, Severity: 21, State: 1.
2015-08-12 16:21:36.88 spid51 Could not redo log record (59:136:21), for transaction ID (0:0), on page (1:20), allocation unit 458752, database 'tempdb' (database ID 2). Page: LSN = (30:165:3), allocation unit = 458752, type = 1.
Msg 945, Level 14, State 2, Line 1 Database 'Northwind' cannot be opened due to inaccessible files or insufficient memory or disk space. See the SQL Server errorlog for details.
I keep getting that error for just trying to USE Northwind. I don't understand why this problem is occurring. It was working fine for the past few weeks. Yesterday something happened to the DB and it said something about recovery to which I accidentally clicked cancel. From then on I can't access some of the DB that I have on my computer.
I'm fairly new to the SQL Server Environment so I would appreciate any and all help that can be provided. I've google searched the details but I can't seem to get any of those to work. Thanks in advance for your reply.
Does anyone knows what is the reason behind following error ? And resolution for that ?
Server: Msg 945, Level 14, State 2, Line 1Database 'Blaster' cannot be opened due to inaccessible files or insufficient memory or disk space. See the SQL Server errorlog for details.
I have an existing project that I have added a simple text file. I am using the Project Deployment Model for this project. I save the project, close it and open the project and the file is there under the Miscellaneous folder. I successfully deployed the project to the server. When I retrieve the project using the Integration Services Import Project Wizard, all of my package modifications are there and the packages up to date but the txt file I added to the Miscellaneous folder is not there.
Last week I backed up my SQL Server by using BE 2012. I named the file "SQL Server BAK" which contained copies of my SQL Server databases. A few days ago I lost some part of my data due to accidental deletion. I backed it up, so I tried to restore the database from the .bkf file. The problem comes here, when I try to to restore my .bkf file, it becomes inaccessible.what causes this?
I have a 3-nodes AlwaysOn cluster (Windows Server 2008 R2 SP1 + SQL Server 2012 RTM), Node Majority quorum, the quorum vote for each node is 1.
Today the AlwaysOn AG was suddenly down due to the cluster service on node 1 stopped and can't be started.
The error in eventlog is -
The cluster database could not be loaded. The file may be missing or corrupt. Automatic repair might be attempted. The Cluster Service service terminated unexpectedly. It has done this 2 time(s). The following corrective action will be taken in 120000 milliseconds: Restart the service.
The failover cluster database could not be unloaded. If restarting the cluster service does not fix the problem, please restart the machine.
The Cluster Service service terminated with service-specific error The system cannot find the file specified..
The error log in cluster log is -
0000156c.000008f8::2012/09/05-08:09:36.057 INFO [DM] Key RegistryMachineCluster.restored does not appear to be loaded (status STATUS_OBJECT_NAME_NOT_FOUND(c0000034)) 0000156c.000008f8::2012/09/05-08:09:36.057 WARN [DM] Node 1: Failed to unload restored hive from the registry with error STATUS_INVALID_PARAMETER(c000000d) 0000156c.000008f8::2012/09/05-08:09:36.057 INFO [DM] Node 1: loading local hive 0000156c.000008f8::2012/09/05-08:09:36.057 ERR [DM] Node 1: failed to unload cluster hive, error 2.
Now the cluster service can't be started on node 1, error code 2. Looks like the clusdb in C:windowscluster is missing or corrupted. How to restore the clusdb file? And how to prevent this happen again?
All nodes were well patched, AlwaysOn and cluster related hotfixes were all installed. [URL] .... doesn't wok.
I am using sql2012 se and we want to use couchbase to store some data as documents. So the database will be Hybrid(partly SQLServer and Partly couchbase). However the database is still in the design phase. What are the things that we should keep in mind when designing this database from a design perspective? Our database which was completely SQLServer based(RDBMS) was using GUIDs everywhere based off NEWID() and prime goal is to get rid of GUIDs for the most part.
I'm intrested in all possible ways to corrupt data or log files in SQL server 2000 except of manualy editing them or cutting the comp's power abruptly.
I scheduled automatic backup process but its only showing backup of the only one .sql file in the backup folder. Other created .sql files are not backed up. Why is it so?
I have a problem with a database on our SQL 7 system . I had set away a restore from tape, but cancelled it after users complained that a different database was performing slowly . However, this has resulted in the database being marked as inaccessible when trying to access it through SQL Enterprise Manager and trying to access it through Query Analyser produces message 927 informing me that the database is in the middle of a restore . How do I remove this error and make the database accessible again??
Many thanks in advance
Peter Burton (IT Support) Durham Aged Mineworkers Homes Association
missing witness server information and the fail-over is broken suddenly? 4:00am no maintenance job. I have one sql job on 10pm for backup on database transaction log only.
I can see the primary have problem then perform fail-over to mirror database, the auto fail-over was broken.
I re-build the sql mirror is OK , but i want to find the root cause.
Windows application event was full when there have many failed event, i have increase log size for application event.
I want to create a new report model project, however in VS Business intelligence Development Studio, I could find the report model template in the New project dialog box. Anyone coulf tell me how can I add the templeate? Thanks
student - sno (pk), sname subjects - subno(pk), subname marks - sno(fk), subno(fk), marks
While creating the report model:
1.Created the data source - no problem 2. created the data source view - no problem added all 3 tables here. 3. created the report model - only subject and student show up here. Where is the marks table?
When I add a certain Table into my report model, one of the fields is not automatically converted into an attribute, but I'm not sure what the exact pattern is.
This table has 3 fields as its key, two of them get included and one does not. The one that does not, is also added as a Role as it is used in a relationship within the DSV (Data Source View).
Does anyone know what rules BIS (Business Intelligence Studio) uses in deciding which fields to automatically convert using the wizard and which to skip?
Perhaps I'm doing something wrong, or there is a workaround?
If anyone can shed any light in the issue, I'd greatly appreciate their comment.
CREATE SESSION MINING MODEL test ( HCVS_MemberId Text KEY, HCVS_MeasureDate_Float LONG KEY TIME, HCVS_MeasureDate Date Discrete, SysPressure LONG CONTINUOUS PREDICT, DiaPressure LONG CONTINUOUS PREDICT, Pluse LONG CONTINUOUS PREDICT ) USING Microsoft_Time_Series ;
How do I specify that I would like to Mean to fill the gap of missing values?
I've stepped into a new environment and have never dealt with multiple data files on user databases only with Temp db.What would be the best way to get all my data files in sync. I have done this on databases that aren't that big in size or off in size by a lot. Here is what I have
I can detach/attach SSAS database.But I have a software that protects(backs up) the files of the SSAS Database.
What the customer needs is to be able to take these backed up files and port it to a different server and attach it there.But the new server complains that these files have no corresponding detach-log files.
The customer doesn't want to backup and restore the SSAS databases.
I have a small project to be done in which I need to fetch the pdf file from a my system and save it in database and also fetch the name of it and save it in the database.
When opening .sql files, I get a connect to database engine prompt every single time. how to stop this from prompting vs. just using my current active connection?
I am testing out a blank database created over two physical files on two separate disks with one table called data which has one column called values nvarchar(max).
I filled the table up with a whole load of data and ran a select * against it. If I run Permon at the same time I can see that the read load has been spread over multiple disks as each of these disks is getting read from in parallel. If I create the same database on a single file and run the same select * again it takes much longer, proving that the read load has been distributed across multiple disks.
Now moving onto writes, this is where the confusion lies. I understand that SQL server fills files evenly until they need growing, after which it will then fill files individually until they are full in a round robin fashion unless you have trace 1117 turned on. What I don't understand is why the writes aren't distributed out whilst it is filling these file groups.
I ran an continual insert into my table with go 1000000 to monitor how the files are being filled up. I monitored where SQL server was physically placing the files as they were being inserted by running the following query:
;WITH CTE AS (SELECT sys.fn_PhysLocFormatter (%%physloc%%) col1, RIGHT(LEFT(sys.fn_PhysLocFormatter (%%physloc%%),2),1) AS [Physical RID], DATAID
[Code] ....
I could see that it would a thousand or so records into file 1, then a thousand or so into file 2, then a thousand or so into file 1 etc etc. In another words it would hit one disk, then another disk, then back to disk one to fill the file evenly. Is there any way to make SQL Server distribute the writes out in parallel so that both disks are writing in tandem?
By the looks of it, multiple disks only scale reads, as with writes only one disk is ever written to at once which is annoying. Any way to harness the write power of multiple disks?
I was in the process of migrating a server from one physical box to another. They are identical drive setups, same OS (2003), same SQL install (2005). Our server team did a 'PlateSpin', which copies the drives from one server to another, as long as the files are not in use. I did not reinstall SQL on the new box, i let the 'PlateSpin' tool copy everything over for me. I then stopped the SQL services on the old server and new server and copied over all of the system database (.mdf & .ldf) files. As soon as i started up the services on the new server, it looked great with one exception. The TempDB was only showing one datafile. When i queried sys.master_files, it was showing me 8 TempDB files. I tried restarting the services, but i still saw the same, only 1 file. I then tried to re-add TempDB files with the same name, but it would error saying they already existed. In turn, i could add new files with different names and they showed up fine. However, on a restart, they would not show up in the properties of the TempDB.
When i queried, sys.master_files again, i now had 16 Temp db files listed in the results. I deleted all but the original single file that was recognized out of the sys.master_files table and re-added the additional 7 files with he original names, restarted the service and then they all appeared.
It works remotely if I run it via command prompt. But when I add this to a TSQL job on my remote SQL instance, it runs without deleting anything. What I'm missing?
I have seen this before. A 2000 restore fails, leaving the database thinking it is being restored but the restore job failed and errors when it is restarted. EM is clueless. I believe there is a proc to reset some flag. Can you share it with me???
My Snapshot Agent for SQL Server 2005 sometimes does not fully complete. I am expecting around 2073 files in the snapshot folder but only get 492 (this number does change). I have immediate_sync set to true, so I should get a full snapshot every time. It seems to quit after all data has been BCP'ed out and before the schema generation starts. I do get the "The replication agent has not logged a progress message in 10 minutes. This might indicate an unresponsive agent or high system activity. Verify that records are being replicated to the destination and that connections to the Subscriber, Publisher, and Distributor are still active. " message in the MSSnapshot_history table.
I am guessing that there is some sort of blocking that is preventing the snapshot agent from continuing and then it times out and dies. But that is a total guess as I have not been able to observe the behavior, only the outcome.
Any suggestions to what the cause is or how I can fix it?
I am working in a test environment where I need to corrupt a database for test purpose/R&D. After corrupting the database, I ran the DBCC CHECKDB (pubs, repair_fast) and then DBCC CHECKDB (pubs, repair_rebuild) commands. But DBCC is not fixing the corrupt database.
My question is why DBCC is not fixing the corrupted Database? Is that a limitation of DBCC or is there any other way/command or tool to fix the corrupted database other then restoring from the previously backed up file.
The following are the steps that I followed to corrupt the database: I used the following script (available in the script section of www.swynk.com) to corrupt the pubs database.
sp_configure allow, 1 go reconfigure with override go update sysindexes set FirstIAM = 1234 where id = OBJECT_ID('roysched') go sp_configure allow, 0 go reconfigure with override go
Then I did a select * from roysched. I got the error: Server: Msg 5180, Level 22, State 1, Line 1 Could not open FCB for invalid file ID 53764 in database 'pubs'. Table or database may be corrupted.
The pubs database got corrupted. Now, I want to fix this error. So I set the pubs database to single user mode and ran the DBCC command : DBCC CHECKDB (pubs, repair_fast) that doesn't help so I ran the DBCC CHECKDB (pubs, repair_rebuild) command.
I got big DBCC output. I doesn't seems to fix the pubs database. When I did a select * from roysched again it gave the same corruption error.
Very strange. I suspect McAfee Anti-Virus was run against SQL 6.5 files, but don't know for sure. Now server won't start. Rebuild master, it starts. Restore Master, it hangs. Rebuild master, Disk ReInit & ReFit devices, user database and server hangs while recovering db log on restart. A backup was made after the corruption got into db, so restoring db has same result, hang. Cannot open db to begin anything. Help?
To make a very long story short, we had a SQL server that experienced problems as a result of bad sectors within one of the drives within a RAID5. Although the drive was replaced the server was never the same and we were forced to perform a migration.
After the migration we began noticing that the server had corrupt data. This was noticed after our applications began erroring out. We determined the corrupt table using DBCC CHECKTABLE and moved the data from those tables into a new table.
Our applications are now working without error, but now it seems as though our indexes are broken which would make sense becuase the tables have essentially been renamed. During an attempted DTS, I selected the option to NOT copy over indexes. I am getting the same error within a DTS as I am within the servers event logs.
Page (3:68379), slot 23 for text, ntext, or image node does not exist. or [SQL-DMO]Code execution exception: EXCEPTION_ACCESS_VIOLATION
This is one of the errors that helped determine that the database was corrupt in the beginning. To make it worse, our backups are useless becuase they will not restore. In order for me to use a backup, I have to copy the MDF and LDF from the production server and copy to the test environment server and then attach it. From the test server is where I was able to succesfully backup and restore.
I would have contacted MSFT a long time ago when this all got started. Unfortunately, I work for a company that will not spend the money on asking MSFT for help. I fear that the long this goes on, the worse it gets.
Please keep in mind that their is a lot of detail that is too must to detail here.
My question is, what would be the best way to essentially start over without losing the data? I assume that since a DTS is failing I am basically "SOL".
We have a table (> 4,000,000 rows) that appears to be missing data, but if we copy the table using "insert into new_table from bad_table" some of the missing data mysteriously reappears. I am assuming there is something wrong with the database, but I don't know where to begin. Can anyone give me some idea of what might be wrong? Thanks in advance!