I've installed SP2 on my server. If I run the package on the server, the package worked there before the installation of SP2. Now with SP2 it doesn't work anymore. In VS2005 on my computer the package works before and after the installation of SP2.
He gives an error with the execution of a sql-task on an oracle server:
Error :Executing the query "insert into cube_content values (trim(?), trim(?), trim(?), trim(?), trim(?), sysdate) failed with the following error: "ORA-01401: inserted value too large for column". Possible failure reasons: Problems with the query, "ResultSet property not set correctly, parameters not set correctly, or connection not established correctly.
I've a problem which I think a lot of SQL Server 2005 users had already ...however I didn't found any answers so far. Maybe anyone of you might know.
I have installed SP1 for SQL Server 2005 on 2 systems, one is an Enterprise Edition (english) on a Win. 2003 SP1 system and one is a Developer Edition (also english) on a Win. XP Prof. system.
On both systems the installation process of the SP1 worked fine except that it gave me the message "Reboot required" on the step installation of SQL Server Native Client.
After my reboot I reexecuted the installation of SP1 with the result of no error nor reboot message. Everything seemed fine....
However when I checked the status of the services I recognized that the SSIS has switched from LOG ON AS ...my specific login.... to NT AUTHORITYNetworkService and the automatic startup failed.
I tried to change the login to my previous login which also failed. It tried to change it to any other possible login but the startup with all logins failed with the error message:
The service did not respond to the start or control request in a timely fashion.
The event log of the system does tell me the same error message. The summary log of the SP1 installation doesn't show any error or warning message except "Failed to read registry key: Debug". Enclosed is the log file:
05/02/2006 22:11:09.609 ================================================================================ 05/02/2006 22:11:09.625 Hotfix package launched 05/02/2006 22:17:22.296 Attempting to install instance: SQL Server Native Client 05/02/2006 22:17:22.312 Attempting to install target: MHPA0F7C 05/02/2006 22:17:22.312 Attempting to install file: sqlncli.msi 05/02/2006 22:17:22.328 Attempting to install file: e:de43f41a74855a4cfc41014b760758HotFixSqlncliFilessqlncli.msi 05/02/2006 22:17:22.328 Creating MSI install log file at: D:WINNTHotfixRedist9LogsRedist9_Hotfix_KB913090_sqlncli.msi.log 05/02/2006 22:17:22.343 Successfully opened registry key: SoftwarePoliciesMicrosoftWindowsInstaller 05/02/2006 22:17:22.343 Failed to read registry key: Debug 05/02/2006 22:17:38.437 MSP returned 3010: A restart is required to complete the install. This message is indicative of a success. This does not include installs where the ForceReboot action is run. This error code is not available on Windows Installer version 1.0. 05/02/2006 22:17:38.515 Successfully opened registry key: SoftwarePoliciesMicrosoftWindowsInstaller 05/02/2006 22:17:38.515 Failed to read registry key: Debug 05/02/2006 22:17:38.515 Pending reboot, successfully installed file: e:de43f41a74855a4cfc41014b760758HotFixSqlncliFilessqlncli.msi 05/02/2006 22:17:38.515 Successfully installed target: MHPA0F7C 05/02/2006 22:17:38.515 Successfully installed instance: SQL Server Native Client 05/02/2006 22:17:38.515 05/02/2006 22:17:38.515 Product Status Summary: 05/02/2006 22:17:38.515 Product: SQL Server Native Client 05/02/2006 22:17:38.515 SQL Server Native Client (RTM ) - Reboot Required 05/02/2006 22:17:38.515 05/02/2006 22:17:38.515 Product: Setup Support Files 05/02/2006 22:17:38.515 Setup Support Files (RTM ) - Not Applied 05/02/2006 22:17:38.531 05/02/2006 22:17:38.531 Product: Database Services 05/02/2006 22:17:38.531 Database Services (RTM 1399 ENU) - Not Applied 05/02/2006 22:17:38.531 Analysis Services (RTM 1399 ENU) - Not Applied 05/02/2006 22:17:38.531 Reporting Services (RTM 1399 ENU) - Not Applied 05/02/2006 22:17:38.531 05/02/2006 22:17:38.531 Product: Notification Services 05/02/2006 22:17:38.531 Notification Services (RTM 1399 ENU) - Not Applied 05/02/2006 22:17:38.531 05/02/2006 22:17:38.531 Product: Integration Services 05/02/2006 22:17:38.531 Integration Services (RTM 1399 ENU) - Not Applied 05/02/2006 22:17:38.531 05/02/2006 22:17:38.531 Product: Client Components 05/02/2006 22:17:38.531 Client Components (RTM 1399 ENU) - Not Applied 05/02/2006 22:17:38.531 05/02/2006 22:17:38.531 Product: MSXML 6.0 Parser 05/02/2006 22:17:38.531 MSXML 6.0 Parser (RTM ) - Not Applied 05/02/2006 22:17:38.531 05/02/2006 22:17:38.562 Product: SQLXML4 05/02/2006 22:17:38.593 SQLXML4 (RTM ) - Not Applied 05/02/2006 22:17:38.687 05/02/2006 22:17:38.687 Product: Backward Compatibility 05/02/2006 22:17:38.687 Backward Compatibility (RTM ) - Not Applied 05/02/2006 22:17:38.687 05/02/2006 22:17:38.687 Product: Microsoft SQL Server VSS Writer 05/02/2006 22:17:38.687 Microsoft SQL Server VSS Writer (RTM ) - Not Applied 05/02/2006 22:17:38.687
I know there are a lot more LOG files in the directory ..HotFix.. which might tell more about the problem. However I don't know in which file I have to look.
Made the jump from SQL2000 to 2005 - i'm not a DBA but like to play around with such nice tools- and tried to convert my DTS package into a 2005 one.
I've got a csv file containing some log in information.
As users log in everyday, this keeps getting added to the file.
I'm only interested in the latest appearance of a certain machine on the network so i filter by creating an unique Index
the CSV file get's imported into a temporary table first and then i executed the following on SQL2000
CREATE TABLE [patchings].[dbo].[allfilter] (
[username] varchar (20) NULL,
[machine] varchar (20) NULL,
[os] varchar (20) NULL,
[sp] varchar (20) NULL,
[date] smalldatetime NULL,
[time] varchar(10) NULL,
[logonserver] varchar (20) NULL,
[country] varchar (20) NULL,
[domain] varchar (20) NULL,
varchar (20) NULL,
[brand] varchar (20) NULL,
[model] varchar (20) NULL,
[speed] int NULL
)
CREATE unique INDEX [my filter] ON [dbo].[allfilter]([machine], [os]) with ignore_dup_key ON [PRIMARY]
go
insert into "Allfilter" select * from temptable order by date desc
This worked perfectly in SQL2000.
when i try to do the same in SQL2005, it doens't start inserting the most recent machines (filtered on the date) but machines at random
selecting the select * from temptable order by date desc shows me the correct order for the machines by date so i'ld presume that it adds the most recent combination of machine and operating system from "today" and ignores the older appearances.
if the temptable contained for instance..
MY-MACHINE MY_USER Windows XP Service Pack 2 2006/06/13 10:02:34 Mydomaincontroller BE MY_DOMAIN my_ip IBM type 2593 MY-MACHINE MY_USER Windows XP Service Pack 2 2006/06/15 9:59:59 Mydomaincontroller BE MY_DOMAIN my_ip IBM type 2593 MY-MACHINE MY_USER Windows XP Service Pack 2 2006/06/16 12:33:09 Mydomaincontroller BE MY_DOMAIN my_ip IBM type 2593 MY-MACHINE MY_USER Windows XP Service Pack 2 2006/06/19 10:48:56 Mydomaincontroller BE MY_DOMAIN my_ip IBM type 2593 MY-MACHINE MY_USER Windows XP Service Pack 2 2006/06/20 11:03:49 Mydomaincontroller BE MY_DOMAIN my_ip IBM type 2593
under SQL2000 i'ld see one record in the final "allfilter" table
MY-MACHINE MY_USER Windows XP Service Pack 2 2006/06/20 11:03:49 Mydomaincontroller BE MY_DOMAIN my_ip IBM type 2593
under SQL2005 i don't see the last one though coding is the same
MY-MACHINE MY_USER Windows XP Service Pack 2 2006/06/16 12:33:09 Mydomaincontroller BE MY_DOMAIN my_ip IBM type 2593
I'm pulling my hair out. After several attempts I got the sp_OAMethodto execute without error. Unfortunately the DTS package isn'texecuting. It also isn't returning any error. What could I be doingwrong? Any help would be appreciated.This is theEXEC @hr=sp_OACreate 'DTS.Package', @oPKG OUTPUTIF @hr<>0BEGINEXEC sp_OAGetErrorInfo @oPKG,@src OUT, @desc OUTSELECT hr=convert(varbinary(4),@hr), Source=@src, Description=@descRAISERROR (@desc,16,1)RETURNENDEXEC @hr=sp_OAMethod@oPKG,'LoadFromSQLServer',NULL,@ServerName='CAMDEV 0',@PackageName='TestPkg',@Flags=256IF @hr<>0BEGINEXEC sp_OAGetErrorInfo @oPKG,@src OUT, @desc OUTSELECT hr=convert(varbinary(4),@hr), Source=@src, Description=@descRAISERROR (@desc,16,1)RETURNEND--Execute the pkgEXEC @hr=sp_OAMethod @oPKG,'Execute'IF @hr<>0BEGINEXEC sp_OAGetErrorInfo @oPKG,@src OUT, @desc OUTSELECT hr=convert(varbinary(4),@hr), Source=@src, Description=@descprint @descRAISERROR (@desc,16,1)RETURNEND
I have created a package that do a file search on an AS400 box using activex scriptand UNC path to do the file search. When I run it locally, it's fine. When I run it on the server, it fails. The login setup for the sql server agent service and the job is the same and they both have admin rights.
In addition, I also have another package pointing to the same path, but the job is to create a text file to the UNC path. It works even when I schedule it.
May someone please help me to solve this problem ?
I have been working with this for about a month now, and no similar problems to date. Today I am trying to introduce 4 configuration flags that control whether optional ETL stage feeds are executed. I did this by adding a do-nothing script component. The precedent and constraint is used, and it checks the boolean variable flag. The first package executes fine. But it never returns from there. This precedent has nothing fancy on it either. It simply does not run any more of the package, make any more conditional checks, nor the common completion tasks. It just seems to think it is done.
The optionals all fire execute package tasks. One thing that might be tripping it up is that I attempt to run one package twice, each time with varying parent package variable set to control it to use a different destination database for each run. Should this not be OK to do?
Pump file data into sql server Move file to "archive" directory(file system task) Delete File (file system task) End Loop
Unmap Drive (batch file)
The Map/unmap code is in a batch file c:windowssystem32et use \10.10.10.10ShareName MyPassword /USER:MyUserName /YES
Unmap: c:windowssystem32et use \10.10.10.10ShareName /DELETE /YES
Here are the results when running this package: 1. Running in BIDS on separate workstation. Everything OK. 2. Running on Server by right clicking on package in Integration Services (SSMS) and choosing "run". Everything OK 3. Running as job with SQL Agent: Package succeeds but no action was taken on the files, files in "ShareName" still there, so therefore no data pumped into SQL Server.
Now, the difference is the SQL Agent jobs are running using a domain account proxy. I'm not sure how that would affect things though--I have the tasks in the package set to fail the package if they fail, so they are not failing, the drives are being mapped o.k.
The computer with the share is non-domain, but that shouldn't matter--I am specifiying the local username and password in the batch file as you can see, and as you can see it works from the workstation in BIDS on a separate machine, and works on the server too as long as I don't run it as a job. The batch file sits on both the server and the local workstation with the same local path.
Any idea why the files aren't actioned when run as a job?
I have several DTS packages saved 'locally' to the SQL server. I want to duplicate a package, so that I can make some changes then replace the original. I certainly don't want to rebuild the entire package from scratch. So, I open up the original package, go to the 'Package' menu and choose 'Save As', then give it a new name and press OK. No errors, all appears well, the title bar even shows the new name of the package. But, when I close the package and go the the 'local' package list, it (the new package name) doesn't appear in the list. Refresh, exit SEM, reboot - doesn't show up. I even looked in the MSDB table where packages are supposed to be stored (at least the name / package id / etc), and it doesn't show there as well. Tried from several client machines.
OS: Windows 2000 Server (advanced) SP2 SQL: SQL 2000 Server (no SP's)
I have many jobs on sql 05 and all work but one. This one writes to an Access DB on the same server as SQL. The package works fine. But when executed in the context of the SQL Agent job, it fails.
Jobs that write to a text file work fine. The Access DB has no password required. By the way, that job in sql 2000 worked fine.
I have this SSIS package which just doesn't seem to run when executing as a sql job and I keep getting this error:
"The command line parameters are invalid. The step failed."
I read some of the comments in forums and they were suggesting to verify the command line for the sql job since there is known bug in the command line for sql job.
But that didn't seem to resolve it and the reason could be one of the variable values that I am trying to set.
In this package one of the variables that I am trying to set is the connection string and my command line looks like this
When I try to run this from the command line I get the error as:
Argument " package.variables[User::MetaDataConnectionString].Value;Data Source=[SVRNAME];Initial Catalog=[DBNAME];Integrated Security=True;"" for option "set" is not valid.
I think the issue is in the set parameter where it seems to be intepreting the ;'s in the connection string as part of its command (which I seem to be escaping by putting them in quotes but it seems to be stripping them off)
Has anyone else encountered this issue? Is there any other escape character that I should be using?
I'm stuck with this and I haven't idea how to solve it. I'm trying to migrate a dts 2000 package from BIDS and I obtain this message:
This wizard will close because it encountered the following error:
Index was out of range. Must be non-negative and less than the size of the collection. Parameter name: index (mscorlib)
I'm going to Migrate DTS 2000 Package select my current sql2k production server (it has almost 600 dts although I think that is not any problem at all)
Wizard recognize without problems my server and then I put a folder for save them but on the next step appears the aforementioned message.
I have an SSIS package is made up of SQL tasks and dataflows. The dataflows connect to an Oracle database using Native OLE DBOracle Provider for OLE DB (10g). This is the first package dealing with oracle that runs on the server. I can execute the package manually by right clicking and going to 'Run Package' while logged in remotely from the server, but it gets hung up and does nothing if I run it as a job. I always have to quit the job. I can disable everything but the dataflows in the package and the job completes and runs fine.
I have a simple parent that uses an Execute Package Task to call a simple child package. The child package has a Data Flow that I disabled. When running the child package by itself, the data flow task is bypassed. When running the package via the parent the data flow task executes.
Second issue is when you disable the package configurations in the child, the parent doesn't recognize that it's diables and tries to load the configurations.
It's almost like there are some settings in a child that get ignored by the parent??
I am trying to execute an SSIS package from an MS Access 2003 database that imports a table from the Access database into a target table in SQL 2005. I saved the package in SQL 2005 and tested it out. If I run it from the Management Studio Console with Run->Execute ... everything works just fine. However, if I try to run it using the following line command "Exec master.dbo.xp_cmdshell 'DTExec /SER DATAFORCE /DTS SQL2005TestPackage /CHECKPOINTING OFF /REPORTING V'" the execution will always fail when the Access database is open (shared mode). The connection manager looks like this: "Data Source=E:Test.mdb;Provider=Microsoft.Jet.OLEDB.4.0;Persist Security Info=False;Jet OLEDB:Global Bulk Transactions=1". The error is listed below:
Code: 0xC0202009 Source: NewPackage Connection manager "SourceConnectionOLEDB" Description: An OLE DB error has occurred. Error code: 0x80004005. An OLE DB record is available. Source: "Microsoft JET Database Engine" Hresult: 0x80004005 Description: "Could not use ''; file already in use.".
Has someone managed to pass successfully a variable from a parent package to a child package? Ive tried a zillion permutations and I cant get it to work. The strange thing was that I was able to successfully do this with pre-RTM builds. Basically, what I am trying to do is:
The parent package has a variable, e.g. ExecutionID which I set using a script to System::ExecutionInstanceGUID. I verified that the variable is set correctly by dumping it to a SQL Server table. I created a child package variable with the same name. In the child package, Ive created a parent package configuration that points to the ExecutionID variable. I am trying to read the variable in a Derived Column Task in which I have a column linked to @ExecutionID. This doesnt work. Step-by-step instructions from someone who managed to concur this will be greatly appreciated.
Oh, I didnt have any luck hitting a breakpoint in a script task inside a child package with both in and out of process execution also.
I have a package with a custom log provider, which runs in BIDS. However when I deploy the package onto SQL Server and run it on the deployed machine, if fails:
"failed to decrypt protected XML node DTS:Password...key not valid for use in specified state..."
Now this is definately to do with the custom log, as if I take it out & redeploy, I can run it on the deplyed server + also run it within a job. I have entered the custom log provider library (+ other required DLLs) in the GAC on the deployed machine, but I'm clearly missing something.
This problem is a bit weird but I'm just wondering if anybody else experienced this.
I have a package that has file system tasks (copying dtsx files actually). Basically the package copies other packages to a pre-defined destination. Thing is, it only works if one of the packages it is configured to copy has some sort of sensitive data (e.g., a connectionstring with a password), otherwise it reports a success message on execution but doesn't actually do anything. I've checked the forcedexecutionresult and it is set to None for that matter.
Just wondering if anybody else experienced this problem and of course if there's a way to solve it.
I've got SQL 7.0 installed on a desktop to develop DTS packages. I've been using it fine for a month or so. Now, all of the sudden, I cannot connect to the server using Enterprise Manager, SQL Analyzer or anything. I get an error stating 'Connection could not be established: ConnectionTransact(GetOverLappedResult())'. The error log states that the max number of connections has been reached. Also, when restarting the server, there is an inticonfig statement that's says it's only allowing one connection. Why? Nobody/nothing else is connected to this server. How do I change it so it'll allow more than one connection? I can't connect to change this setting.
Hi, I'm using SQL Server 7 and have it installed on one PC, Not networked. Without any explanation Enterprise manager won't let me look at my databases. It lets me look at a server on the web but not my own Server on this PC.
It has worked without any problems until now. I receive this message when I try and look at the servers databases.
**A connection could not be established to H3G5Z8- Cannot open user default database <ID> using master database instead.
Please verify SQL Server is running and check your Server Registration properties.
********************* Total mystery right out of the blue.
This same code used to work but doesn't anymore. The connection times out and the error points to the dbCMD.Execute command. Here same things I have found: 1. If I point the database calls to the development server it works. If I point the database calls to the production server it fails. I am to get to production through other calls just fine.
2. If I comment out the dbCmd.Execute or dbCmd.CommandText = "knet_user_setLastLogin" seperately the code works.
What is the dbCmd.Execute doing when it is called? Could it be a problem with the store procedure? Could it be a software upgrade on the Database server that is causing the problem?
I rebuilt my machine a few weeks back - re-installed the developer version...The app in question can see all it's tables in the SQL back end via ODBC andeverything looks normal via Enterprise Manager, but I can no longer createan ADO Connection.Source code is unchanged, what's changed is the PC rebuild and consequentre-installatin of SQL Server developer version.I suspect it's something to do with the provider (at least that's the only thingI can see in the .Connect string that's not verifiably correct....)The error looks like this after I trap and format it:-----------------------------------------------------------------12/30/03 21:34:30v0.71 Userid: UPQC on SAGProc: basADO: ADO_ConnectionCreate-2147467259: [DBNETLIB][ConnectionOpen (Connect()).]SQL Server doesnot exist or access denied.Errors encountered when trying to connect:''''''''''[DBNETLIB][ConnectionOpen (Connect()).]SQL Server does not exist oraccess denied.----------Connect String:Provider=SQLOLEDB.1;SERVER=SAG;DATABASE=TRETS;UID= Trets;PWD=trets-----------------------------------------------------------------Can anybody see anything obvious to look for?I'm thinking I need to know how to verify/validate the presence of provider"SQLOLEDB.1"... But how?--PeteCresswell
I have an ASP.NET/ADO.NET page i wrote. It was based off of the development database during most development, and it switched over seamlessly to the production database.
Except for yesterday, when i went back to edit the page. It simply stopped returning results on my queries. To see if it was caused by bad values being passed to the sqldatasource, i tried it in the "Configure Data Source..." window, and it returns zero rows. When passing all of the exact same parameters to the same procedure in the same DB engine, even using the same login information, it works.
Just not in my ADO.NET app anymore...
Can anyone tell me what the hell could be going on here, please? I already rebuilt the page from an older version this morning, being very minimal on changes, and the last thing i changed before it stopped working was the connection string... *BUT* I have triple-checked that against the canon string we have always used in every working copy.
I'm losing a lot of hair over this . Any help would be greatly appreciated.
I'm unsure what happened, but I'm using SQL Server 05 Developer Edition. I've been working with a local database where I tried to add a clustered index on a huge table (about 50M records). Well, I had to stop the query since it was maxing out my work computer. Since then, I've been unable to open the database or run queries against it.
When I open SQL Server Management Studio the database says In Recovery. Once it's done, when I try to expand tables, I eventually get this error:
TITLE: Microsoft SQL Server Management Studio ------------------------------ Failed to retrieve data for this request. (Microsoft.SqlServer.SmoEnum) For help, click: http://go.microsoft.com/fwlink?ProdName=Microsoft+SQL+Server&LinkId=20476 ------------------------------ ADDITIONAL INFORMATION: Lock request time out period exceeded. (Microsoft SQL Server, Error: 1222) For help, click: http://go.microsoft.com/fwlink?ProdName=Microsoft+SQL+Server&ProdVer=09.00.1399&EvtSrc=MSSQLServer&EvtID=1222&LinkId=20476 ------------------------------ BUTTONS: OK ------------------------------
If I run a Select * from INFORMATION_SCHEMA.TABLES query, it will go on for a while
If I try to access the properties of the Database (right clicking), I get this error:
TITLE: Microsoft SQL Server Management Studio ------------------------------ Cannot show requested dialog. ------------------------------ ADDITIONAL INFORMATION: Cannot show requested dialog. (SqlMgmt) ------------------------------ An exception occurred while executing a Transact-SQL statement or batch. (Microsoft.SqlServer.ConnectionInfo) ------------------------------ Lock request time out period exceeded. (Microsoft SQL Server, Error: 1222) For help, click: http://go.microsoft.com/fwlink?ProdName=Microsoft+SQL+Server&ProdVer=09.00.1399&EvtSrc=MSSQLServer&EvtID=1222&LinkId=20476 ------------------------------ BUTTONS: OK ------------------------------ I'm unsure where to go from this point forward.
Yesterday I made the HelloWorld_CLR sample work and figured today I would create a simple windows service to immulate its functionality. When this was not working I tried the sample again and it stopped working. I have tried several things and nothing has worked. Has anyone seen this before or have any ideas on how I can get it working again? I even tried removing SQL Server and its samples then reinstalling to give me a clean slate but that did not work. I did not make any changes to the software nor did I make any changes to the database however the sample did so I need to make sure I can fix this incase it happens in development not to mention I need to continue development.
The Client Queue is normally empty but sometimes has a request to two in it. The Server Queue has alot in it that are both Request and an unspecified type or has a URL for a type. I would paste it in but it takes too much space. Please let me know if you have any ideas because I am not sure what else to try.
Bellow is a sample of the output from running this sample.
Connecting to SQL Server instance Connected to SQL Server instance
Transaction 1 begun Dialog begun from service (HelloWorldClient) to service (HelloWorldService) Message sent of type 'Request' Transaction 1 committed
Transaction 2 begun Waiting for Response.... No message received - Ending dialog with Error Transaction 2 committed
I installed the dts designer component for 2005 so I could migrate over some 2000 packages and was able to successfully open and run dts packages under 2005.
Then all of a sudden I get this error msg and can't open them, migrate them, export them, etc:
"here were errors loading the package "MyPackageName": [DNETLIB][ConnectionOpen (Invalid Instance()).] Invalid connection. (Microsoft.SqlServer.DtsObjectExporerUI)"
That is not a typo, it actually says "here were errors..." instead of "There were errors..."
I've tried the obvious, repair the dts designer component and rebooting, no luck.
When I try to run my report manager which is in my local machine, I get the following error message.
The report server cannot open a connection to the report server database. A connection to the database is required for all requests and processing. (rsReportServerDatabaseUnavailable) Get Online Help
An error has occurred while establishing a connection to the server. When connecting to SQL Server 2005, this failure may be caused by the fact that under the default settings SQL Server does not allow remote connections. (provider: Named Pipes Provider, error: 40 - Could not open a connection to SQL Server)
I can not open Report server data base and reportserver temperary data base, even in Management Studio. I can not see them there anymore.
Also When I try to open data base engine in my local machine through Management Studio, I can not open it. I get the following error massege.
An error has occurred while establishing a connection to the server. When connecting to SQL Server 2005, this failure may be caused by the fact that under the default settings SQL Server does not allow remote connections. (provider: Named Pipes Provider, error: 40 - Could not open a connection to SQL Server) (.Net SqlClient Data Provider)
Recently, I changed data base connection properties of all my reports. coz earlier I used a data base which was in the local machine. But then I connected all my reports to the same data base which is on another server. Then I changed all connection properties of each report. After that I deleted the data base in the local machine.
Would that be the reason for this problem?
Also SLQ server(MSSQLSERVER) and SQL Server Agent has been stopped running. I can not start it from SQL Server Configuration Manager.
Please can anyone tell me how to fix this problem and get my work done as usual?
I've a problem with my database. Till yesterday the option for Auto Grow of Database (10 %) was working very fine, but now it seems to be some problems with it. Finally I had to specify a restricted size for the database and then it again startd to give me some space in the database to write in. Ideally it should have worked automatically, isnt it ???
There is no problem with the space on the drive, I still have some 76 gb of free space there ...
"SSIS 2012 Catalog doesn't have option to give read access to SSIS Catalog to view package run reports" ... Any luck allowing power developers / operators access to READ the SQL 2012 SSIS Execution Reports without granting them SSIS_Admin or Sysadmin?
According to this link posted back in 2011 (w/ Microsoft's feedback in Nov 2011: "We’re closing this issue as “Won’t Fix.” At this point the bug does not meet our bar for resolving prior to SQL Server 2012 RTM. As we approach the SQL Server 2012 release the bar for making code changes gets progressively higher." URL....Regarding Permissions to SSIS Catalog, here are the findings. We can give access in three ways:
1. READ Access – We can provide a user db_datareader access. With this the user can see the objects within the SSIS catalog database, but cannot see the reports.
2. SSIS_ADMIN – Add the user to this database role in SSISDB. With this the user can view the reports. But it also provides them privileges to modify catalog information which is not expected. We can add it using below script EXEC sp_addrolemember 'ssis_admin' , 'REDMONDPAIntelAnalyst'
3. SYSADMIN - Add the user to this server role. This will make the user an admin on the SQL server. This is not intended. Is there any method available which will have provision to give read only access to see SSIS Catalog package execution reports and not having modify Catalog access.
Hi,SQL Server 2000 SP3Windos 2000 Server SP4I have a DTS package that imports data from a dBase IV databse withfiles located in two folders (dBF1 and dBF2). I use a transform datatask to transform the data.They were running properly, but last week we installed W2K SP4, andnow the transform task for files from dBF2 are not working properly.I have two tranform tasks to extract data from files in dBF2 folder.If I double click to open the transform data task of either of them,Enterprise Manager crashes with the errrormmc.exe applicatio ErrorThe instruction "xxxx" referenced memory at "xxx". The memory couldnot be read.Althoug the transform task for one of the files will run, the otherwill not run giving the messageError Source: DTS packageCatastrophic failureAlso, I have an Access database that has links to the same dBasefiles. For files from dBF2 folder, I'm able to see the data from oneof the files, but if I double click to see the data from the other,access crashes with no specific error message. Nothing has changed ondBase related files (permission wise).The transform tasks to extract data from the other dBase folder (dBF1)files are working fine, and data is accesible from Access.Any advice how to tackle this one?
I use ASP.NET 2.0 and SQL 2005, the SQLCacheDependency didn't work stable for me, it usually works and after a while, it stops working. Recently, It doesn't work. Today I reset service broker by
Alter DB set new_broker
It seems all work, but too early to be happy, it doesn't work now anymore. I don't see any record in
sys.transmission_queue
And I don't see new entries in
sys.dm_qn_subscriptions neither when I modify a record which bind to SQLCacheDependy.
I see some errors in log like
Message The query notification dialog on conversation handle '{8F8CC642-6340-DB11-8F09-0014227B7B80}.' closed due to the following error: '<?xml version="1.0"?><Error xmlns="http://schemas.microsoft.com/SQL/ServiceBroker/Error"><Code>-8490</Code><Description>Cannot find the remote service 'SqlQueryNotificationService-f79776f7-9ca5-4c5f-8a66-0d81f7673683' because it does not exist.</Description></Error>'.
This happens only on multiprocessor machines. SQL Service pack 3. The only way to fix it is to restart the SQL service. If you have a performance tool based on that value, like Idera's SQLdm- sorry, it will show zero.