I am working on on SQL Server Integration Services and facing few problems in it.
Actually am supposed to create a package that would automatically pick excel spreadsheets with a specific format and import it onto the SQL server.
(Lets say , there is a company named AB and they have got various products named CD,EF and GH and each product has its own spreadsheet in which its monthly sales(JAN, FEB,...NOV, DEC) is given. So i have to build a generic package for each product (lets say CD) so that i don't have to import every spreadsheet seperately for each month.
To summarize i just have to build a package where i can deploy the various spreadshhets again and agian instead of building a package for each and every month spreadsheet seperately.
I have tried and used lots of combinations like Loop
conatiners etc. but still am not able to find a solution to it.
I'm working on a web application that uses SQL Server 2000. Once I'm ready to upload my website to a production server, I need to know how to upload my SQL Server database files to the server. Are there actual files that can be uploaded? I've only deployed a website that works with MySQL to a production server, and there were no files to upload... the database had to be created on the server, and then the data itself had to be exported to text files (on my computer), then imported/uploaded to the production server using SQL statements...
Hi..I would not like my customers to have access to any SQL code that isbeing packaged with our product and hence would like to encrypt it insome fashion. What's your recommendation on packaging/deploying .sqlfiles?Would appreciate any feedback.Thanks,Sandeep Madduri.
I am deploying SSIS packages using the deployment utility. We are starting to use configuration files for storing connection information. I have the configuration files in a directory on the server I am deploying the packages to, I want all new packages deployed to use these files. When I deploy a new package it wants to overwrite the config files on the server with the ones in the package. I have tried to remove the .dtsconfig file from the package deployment folder but then the package will not use the config files on the server. How do you deploy a package and get it to use the .dtsconfig files that are already on the server?
I generated and transfered reports using RSscripter. I created those reports in my local machine. Then I transfered to prodcution machine. But the problem is there is no Visual Studio in prodcution machine. So I'm wondering how to deploy those reports now. Can any one tell me how to do this please?
I am hoping someone with more experiencing deploying packages from dev to other environments can help me out here. When developing in BIDS everything works fine. When I deploy I want to use a configuration file. This does seem to work fine by specifying the configuration file on the DTEXEC command. However, if it is specified incorrectly the package STILL executes using the information it was created with. I am attempting to control things like connection strings and passwords via configuration files so that I can take the same package and deploy to different locations and have it run against different sources and destinations. What I am afraid of is that things could easily get screwed up if the configuration file gets moved or renamed, or if the dba that is scheduling the job (outside of sql agent using a different job scheduler on a machine that only has the SSIS service) messes up typing the config file name. Shouldn't it fail if you specify an invalid parameter. Isn't there a way to make it do that? Any help or insight is appreciated. BTW I have used environment variables as well and the issue is the same. If it cannot locate it, it still will execute.
I've written a ClickOnce app that deploys a Sql Server Express data file (.mdf) and log file (.ldf) along with the other application files. SQL Express SP2 is a prerequisite for this app. After the applicaiton installs and starts up it copies these files to a specific folder and attaches them by calling the AttachDatabase() method of a Microsoft.SqlServer.Management.Smo.Server object.
This method works fine using Windows XP, however when I run it on Windows Vista I get the following exception:
Microsoft.SqlServer.Management.Smo.FailedOperationException: Attach database failed for Server 'localhostSQLEXPRESS'. ---> Microsoft.SqlServer.Management.Common.ExecutionFailureException: An exception occurred while executing a Transact-SQL statement or batch. ---> System.Data.SqlClient.SqlException: Unable to open the physical file "C:WashMasterDatabasewasshmaster.mdf". Operating system error 5: "5(Access is denied.)".
This exception occurs when I try to programmatically attach the database.
I've searched through the MSDN forums for a solution and I've only found two suggestions:
1.) Create all database connections using a User Instance. 2.) Create the database (using SQL) in the target folder, rather than copying it over from the deployment folder.
I don't like the idea of using User Instances because my entire data access layer is currently written using Ole and I would rather not rewrite it if I don't have to. I cannot find any way to use User Instances with Ole.
I also don't like the idea of creating the database using SQL because it is fairly large.
Here are my questions:
1.) Can User Instances be used with Ole?
2.) Can I somehow programmatically set the permissions of the copied data and log files so that SQL Express can access them without error?
3.) Is there a better method for doing what I'm trying to do or am I destined to have to either rewrite in data access layer so I can use User Instances or create my large database rather than copy it to the target directory?
Can any one pls tell me or send me any links of how to deployee SSIS Package in SQL server 2005. I had used fallowing steps but i am unable to find my package under MSDB.
1) Created Package Configuration of XML type. 2) Under Project properties i changed CreateDeploymentUtility to True. and build my application. 3) Under binDeployment of application folder i executed my Menifest. to SQL
This maybe simple question but I can not seem to figure it out. If you use the deployment utility to deploy your package how do you specify a different SSIS directory to install it to. For instance when I'm in SSIS under stored packages > MSDB > I created a folder called package 1. I want to install to that folder. The only way I can do it is to manually import the package from SSIS.
I have created a package which reads Informix data & puts it into Sql Server
I get these warnings when I do the build:
Warning 2 Validation warning. load tm ty depts: OLE DB Source [1]: Cannot retrieve the column code page info from the OLE DB provider. If the component supports the "DefaultCodePage" property, the code page from that property will be used. Change the value of the property if the current string code page values are incorrect. If the component does not support the property, the code page from the component's locale ID will be used. load_daily_sls.dtsx 0 0
but when I run it in development, all of the tasks complete
I am trying to deploy the package
I set CreateDeploymentUtility to true. When I do the build, I get no errors but I also get nothing in bindeployment or anyplace that I can see
It is not doing anything because of the warnings or something else? Exactly where is bin?
I used to write DTS Scripts in SQL Server 2000 and schedule them as jobs with out problem.
This was normally done within SQL Server its self.
Now that I've moved to using SQL Server 2005 I've been learning how to use SSIS.
I've successfully developed a package and managed to create a .dtsx file. Now I have 2 large books on the subject of SSIS but none seem to go into any detail on what to do next.
So here€™s my newbie question (I apologise if I sound dumb!):
I don't want to run my package manually as the books keep telling me how to do.
I need to have my package added into SQL Server 2005 somehow and then schedule it as a reoccurring job.
I hope someone can help becuase this problem is issue us several headaches.
We are currently trying to deploy an SSIS package to a production server. The deployment goes fine, the package runs ok when executed manually. The issues start when we try and execute it under the SQL agent.
Having gone back to the drawing board and spent much of the day reading various articles and applying the various options (especially those within the MS KB article 918760), we are still no closer to a resolution.
The SSIS package was created under an Administrator, and the SQL agent runs under a different Domain Admin account.
When we set up the Schedule to read from SQL Server or the SSIS Store the standard "Executed as user: DOMAINUSERNAME. The package execution failed. The step failed" in the history.
We tried to create the package as a file access and now get "Package could not be found" even though you can browse to i in the schedule list. The Domain account as full access to the folder where the package resides.
Has anyone else come across this issue, or have a workable solution?
Hi, I am trying to deploy SSIS package through manifest file on a server. It gives the following message.
'Could not save the package "C:Package Pathpackagename.dtsx" to SQL Server "(local)"'. Additional Information: The SaveToSQLServer method has encountered OLE DB error code 0x8004E14 (invalid Target Directory. You must choose a project folder to store your package. If this is a permanent package, contact your DBA to create a project folder.
I have tried to create the package folder on the same path it is trying to save the package. But still i get the same error. Is this a permission issue? I have all rights on sql server. Thanks
I modified an existing package created by someone (who left the organization). Its a ftp bulk insert from flat files to the tables. The need was to updates 2 tables out of 9 with additional columns. I made the changes to the package and it runs successfully on my laptop.
After that I build the package and tested it on the development server, although the package runs successfully but the data is written to my DB tables in my laptop instead of development server DB tables!.I see "." in the place of the connection manager server name. The DB name is the same in all the three environment (local, development & production). There is a configuration table used in the package.
Brief overview...Running SQL Server 2003 Server Enterprise 64 bit - All Service Packs and patches current SQL Server 2005 Enterprise Edition 64 bit Build Microsoft SQL Server 2005 - 9.00.3054.00 (X64) Mar 23 2007 18:41:50 Copyright (c) 1988-2005 Microsoft Corporation Enterprise Edition (64-bit) on Windows NT 5.2 (Build 3790: Service Pack 2)
I cannot import any SSIS packages nor crete any new folders under stored packages. I hve googled the news groups and looked at BOL to no avail. HELP!!!!
I have a requirement where in i have around 15 different flat files , filenames are fixed but folder path can be changed(i think i should use a variable for folder path). These 15 files data should go to their respective tables in the database.
Whether I need to create separate data flow task for each file or separate package? In addition to these, example : while importing product data into product table, if product ID already exists, we need to ignore it and upload only the new records.
I have a couple of hundred flat files to import into database tables using SSIS.
The files can be divided into groups by the format they use. I understand that I could import each group of files that have a common format at the same time using a Foreach Loop Container.
However, the example for the Foreach Loop Container has multiple files all being imported into the same database table. In my case, each file needs to be imported into a different database table.
Is it possible to import each set of files with the same format into different tables in a simple loop? I can't see a way to make a Data Flow Destination item accept its table name dynamically, which seems to prevent me doing this.
I suppose I could make a different Data Flow Destination item for each file, in the Data Flow. Would that be a reasonable solution, or is there a simpler solution, or should I just resign myself to making a separate Data Flow for every single file?
I am trying to restore multiple .bak backup SQL database files onto a new server. However, I have found that it will not allow me to restore multiple databases at once. Is there a way to do this so that I do not have to manually upload one at a time? I tried adding all the .bak files at once to the backup device window but it only did the first one listed. It would be so much easier to restore them all at once so that I do not have to continue this manual process. I am restoring them via device.
I need to be able to bulk insert a bunch of tables from their corresponding flat file. I have created an XML file (see below) which has the file name/table name pair at each node. I then created a ForEachLoop task and used the Node enumeration type and the following OuterXpathString: ReferenceFiles/File. At this point I get lost. How do I pass the 2 inside node values (file name and table name) to variables which I can then use as expressions for the bulk insert task inside the Foreach?
I used the data export wizard to export a single table to a single flat file (multiple wasn't allowed). I saved the package as a *.dtsx file which I'm attempting to edit to add the additional tables.
Creating additional sources is fairly easy copy of the first source and change to the table name.
I've tried copying the destination connection and changing to a new text file, but can't get past having to add each column manually to the new destination.
How can I duplicate the mapping that must be taking place in the wizard in the *.dtsx editing environment?
This seems like a simple / common task, but I've been unable to find a solution.
I have searched but not found quite the best way to look at this so far..
I have an application that outputs data to several text files (up to 30). These have commonality by an object name, but then contain completely different column data.
In DTS I had each of the source text file connections going to one OLE DB connection and then individual transform data tasks pointing to the one OLE DB connection.
Looking at SSIS, it would appear that I would need to have one source and one destination for each of these and therefore 30 parallel data flows?
Just wondering if there is a neater way of doing this??
It is a regular data import that happens a few times a day - the text files are named the same as the SQL tables - ie app_userdata.txt goes to app_userdata table.
I have a multiple package solution that I've deployed using the manifest file produced with the development environment. If I need to make a change to a single package, how do I then deploy this package? Is it a case of rebuilding the entire solution and re-running the manifest file, or is there a simpler way?
I am trying to write (my first, unfortunatly) DTS, and am having some problems.
I need to be able to import multiple flatfiles (all in the same format, just with different schema), each one going into a different table. I have written an application to call my DTS, sending it variables for the tablename and the filename. This works fine when I test it on a single flatfile.
My problem is, the Tranformation object does not reset after each DTS call, so I get "Column does not exist" errors after the first successful import. I can go into the DTS Manager and reset the Transformation options, but that would defeat the purpose of automation. Is there anyway to reset, or another technique, the Transformation object so that it will continuosly work on files that use different schema?
I am very new at DTS, so please consider me "ignorant" when replying.
I have multiple .sql files, exe each one manually is a pain, so how do you run multiple .sql files all at once? Beside creating a batch file, are there t-sql commands that could execute .sql files?
I have a rather large sale transaction DB. Basic header, and detail tables. I am providing a third party company with daily sales information, and I need to give them back data from about 8 or 9 months ago. I currently have a DTS package that gets sales for the current day, but since I have to go back, I have to manually edit the query in the DTS package, and change the date range...UNLESS ...
Blah, blah, blah. The problem is that they can only take the data in Daily files. So, there would be ONE file for each day. I really don't need to be manually running these jobs, so I'm wondering if someone could point me to a way of writing a package (maybe ActiveX, not sure) that would run through a loop, basically, of dates, and create a seperate file for each day. Versus having to edit a generic DTS package, and changing the date range 350 times...
Using an expression to set the log filename to include the date and time results in 3 log files being created.
Ummm. Why? I only ran the package once. Is SSIS not sharing my log file connection among the different components?
On first thought my workaround would involve using a script task to "set" the log file name to a variable and use an expression to set the log connection to the variable. But the problem with that is that logging starts BEFORE the script task is run...
I have a job with a single step that executes a stored procedure that performs the following steps:
1. Checks for the existance of a file A in a folder A
2. If it exists,
a. executes the cmdshell to run a DTS package to drop a table, recreate it and load the data in the file A to table X
b. runs other stored procedures that use the data in table X to create other tables Y and Z
c. executes the cndshell to remove and rename the file A from Folder A into Folder B
What I'd like to do is use this same stored procedure if possible, but create a job or another store procedure that would loop thru and process multiple files in Folder A instead of just one.
hi, all; i'd like to resolve this problem: We are Developing packages in Dev.Server; if we are Deploying our packages to Prod.Server; we need to modify the connection string and related properties of the connection manager. 1. How to modify the connection manager while deploying to production server? 2. can we create 2 dtsconfig file for a single package; is there any mechanism for instructing sequel server to use a particular dtsconfig file for our package.
i modified the config file in order to change the connection string. but its affecting only the config file not the package. while running the package its reffering the old connection string only.
Hi, I have about 300-400 XML files I want to load in my SQL database (2005). The following code will load one (1) file. How do i do a mulitple collections? INSERT INTO MEL (DATA) SELECT * FROM OPENROWSET (BULK'C:TempCHAPTER1.xml', SINGLE_BLOB) AS TEMP Thanks,
Usually, our in house ERP software has 1 database and 1 database file. After an upgrade from MS SQL 6.5 to MS SQL 7.0 I have a database who's properties show that it is made up of multiple datafiles. What is the easiest and safest method to return this database to only have 1 datafile?