Although unsupported / undocumented, using FN_DBLOG(..) is inevitable.
While retrieving [RowLog Contents 0]/ [Log Record] transaction log information
I encounter a possible limitation regarding the size of the fetched data.
There is a suspect that data exceeding 8000 bytes is truncated.
I do not manage to get more than 8000 bytes - neither at the Studio Management Query pane nor with a programmatic ODBC piece of code.
There is more data for sure - no doubt about that!!
Can anyone tell me if SQL 2000 has a record limitation on tables? The table that I have is very basic. It contains 6 columns, none of which can contain more than 10 characters, but I'm already sitting at 2,500,000 records and expecting up to 20-30 million records. Can it handle it? Thanks in advance.
I'm got a "folder" structure application which we'll be using as an in-house directory viewer. (In case you're wondering, it doesn't relate to any "real" folders, so using xp_cmdshell is out! )
Each folder and file record can have its own permissions, however these are assumed to inherit from the parent folder if no specific access rules have been set, basically in the same way file systems work. Each file record can only have one parent, and a folder can either have a parent or be at the root level.
Right now I'm having an issue with the inheritance of permissions. Say if I want to grant access to "Folder 1" to "Group A", then "Group B" shouldn't be able to see it. However, if I grant access to "File 1" in "Folder 1" to "Group B", then "Group B" should be able to see "Folder 1", but only see "File 1" and not the rest of the contents.
I thought I could do this with a CTE, but I'm having a bit of difficulty..
Here's the code:
CREATE TABLE #FileSystem ( FSIDINTEGER NOT NULL IDENTITY(1,1) PRIMARY KEY ,ParentFSIDINTEGER NULL ,NameVARCHAR(100) ,RecordTypeVARCHAR(1)-- (F)older, or Fi(L)e
Gurus,How do I choose multi-name 'columns' like "Current LSN" or "TransactionID" from ::fn_dblog?I CAN select single 'column' names like 'Operation' and 'Context'.I do NOT want all 'columns' returned:- I don't want Select (*) from ::fn_dblogDoug
Hi All,I have come up against a wall which i cannot get over.I have an sql db where the date column is set as a varchar (i know, should have used datetime but this was done before my time and i've got to work with what is there). The majority of values are in the format dd/mm/yyyy. However, some values contain the word 'various'.I'm attempting to compare the date chosen on a c# .net page with the values in the db and also return all the 'various' values as well.I have accomplished casting the varchar to a datetime and then comparing to the selected date on the .net page. However, it errors when it comes across the 'various' entrant.Is there anyway to carry out a select statement comparing the start_date values in the db to the selected date on the .net page and also pull out all 'various' entrants at the same time without it erroring? i thought about replacing the 'various' to a date like '01/01/2010' so it doesn't stumble over the none recognised format, but am unsure of how to do it.This is how far i have got: casting the varchar column to datetime and comparing. SELECT * FROM table1 WHERE Cast(SUBSTRING(Start_Date,4,2) + '/' + SUBSTRING(Start_Date,1,2) + '/' +SUBSTRING(Start_Date,7,4) as datetime) '" + date + "'"Many thanks in advance!
I am currently designing a SSIS package to integrate data into a data warehouse fact table. This fact table has about 70 columns among which 17 are foreign keys for dimension tables.
To insert data in that table, I have to make several transformations and lookups. Given the fact that the lookups I have to make are a little complicated, I have about 70 tasks in my Data Flow. I know it's a lot, but I can't find a way to make it simpler. It seems I really need all these tasks.
Now, the problem is that every new action I try to make on the package takes a lot of time. At design time, everything is very slow. My processor is eavily loaded each time I change a single setting in one of the tasks, and executing the package in debug mode takes for ages. If I take a look at the size of my package file on disk, it's more than 3MB.
Hence my question : Are there any limitations in terms of number of columns or number of tasks that can be processed within a Data Flow ?
If not, then do you have any idea why it's so slow ?
I have result set as the attached in screenshot, where the candidateid gets duplicated, but then the column university and careerchoice column has different values for the same candidate. (the CareerChoice and University comes from different tables and i have used union clause)
What want is e.g. CandidateId 186 repeats three times, it should be one row, with three columns all filled with the values which are in different rows.
So that I don't have duplicate candidate ids and i have single row with all column data.
Say a record has four columns in { Emp#, firstName, secondName, thirdName, spouseName }. Now if you configure a sql datasource for a dropdownlist to return all columns, only Emp# will be displayed in the dropdownlist, the other columns wont be listed. Is there a way to list the values of all columns in the dropdownlist using the sql datasource or in ADO.net code. Thanks.
IdnCol ColumnA ColumnB ColumnC ColumnD 1 CC DD EE FF 2 DD GG HH JJ 3 HH JJ KK HH
How I can check for the repeated value in the column. for example in the 3rd record, in columnA and ColumnD the value HH is repeated so is there a way to find if there are values repeated in any of the columnA ColumnB ColumnC ColumnD.
May i know what is the SQL query to create another intermediate column which is used to store the sum of two columns of each and every record. Thank you.
I am looking for the best way to check to see if any columns are still NULL in a record. I have a form that gets filled out by users and the values entered into TableA. There are 6 columns in the table, 5 are responses and column6 indicates if the record is complete. So I want a way to see if all of the first 5 columns are NOT NULL and, if so, mark column6 with a 1.
I am thinking this would be a good thing for a trigger on INSERT or UPDATE to check to see if the first 5 columns are filled in and then mark the record as complete.
I am looking to create a script that will go through a table a pick out the necessary columns to create a unique record. Some of the tables that I am working with have 200 plus columns and I am not sure if I would have to list every column name in the script or if they could be dynamically referenced. I am working with a SQL server that has little next to no documentation and everytime I type to mere some tables, I get too many rows back.
I am trying to figure out how to use the select maximum command in SQL Server 2005. I have already created a database and I have it populate it with multiple fields and multiple records. I Would like to create a new column or field which contains the maximum value from four of the fields. I have already created a column and I am trying to figure out how to use a command or SQL statement which is entered into the computed equation or formula in the properties for this field/column.
Any help you can provide will be greatly appreciated!
I'm new to SQL so don't laugh if this is easy to spot but I'm having trouble with a select statement but I do not think it's the syntax etc, I'm convinced it must be due to some restrictions and limitations within SQL.
Could anyone shed some light? I've been 'googling' for two days now and can't find anything, much appreciated! thanks.
INNER JOIN tblAction ON a.action_id = tblAction.action_id INNER JOIN tblClassification ON a.class_id = tblClassification.class_id INNER JOIN tblType ON a.type_id = tblType.type_id INNER JOIN tblUnit ON a.unit_id = tblUnit.unit_id INNER JOIN tblSystem ON a.system_id = tblSystem.system_id INNER JOIN tblLibrary ON a.lib_id = tblLibrary.lib_id INNER JOIN tblTransport ON a.trans_id = tblTransport.trans_id INNER JOIN tblPOC ON a.poc_id = tblPOC.poc_id
WHERE a.medno_id = 327 ORDER BY a.med_effdate
I've spaced out the query so it's easy to read
I've started the query from scratch building my SQL adding one field at a time and it works perfect until it reaches a certain number of fields, it's as if SQL has a limit to no of fields it can return/be used in the select part of the statement. Are there any limitations I should know about? or am I being an idiot and doing something I shouldn't?
What are the limitations of the SQL 2005 Express Management Studio tools, as opposed to the full blown 2k5 Tools (i'm not referring to the Server, just the tools). I ask because the install for the 2005 Management Studio Tools alone is 878 megs.
I've been using CE 3.1 in a .NET/C# application. My knowledge of database technology is very basic, and I'm wondering how far I can take my application with CE, given its limitations, which I stumbled across when I read the following about VistaDB: http://www.vistadb.net/compare_sql_compact.asp
For right now, my application uses a CE database as a fancy log file. There are only two tables in the database -- one to hold the log information, and one to hold information about different "runs" of my application. The only time a row is ever updated is when a row in the latter table is updated (once) to indicate that that "run" has completed. Other than that, I only add and delete rows.
Right now, two different processes on the same CPU (the same CPU as the database) write rows to the database during a run of my application. They do this by accessing a singleton object, hosted in a Windows service, via .NET remoting. One of these processes also polls the database continually to read it for display purposes.
So. My questions are:
I can live with the single-user limitation, given my application. But what exactly is it about using CE that limits me to a single user? I never actually specify user information, as far as I know. Am I missing it?
Can I assume that the "single CPU supported" limitation simply means that it only runs on a 32-bit Windows machine, and that it has nothing to do with multiple databases running on different CPUs at runtime?
Why is it that the two processes in my application seem to be able to connect to my database concurrently, when this is apparently a limitation? As I said, my knowledge of this technology is quite elemental. Forgive me. But I figure I'm in the right place to change that. :-)
I am receiving an error message while using the System.Transactions.TransactionScope class. The error message that I am receiving is "Communication with the underlying transaction manager has failed". This error seems to only appear when I have my web application one server, Server1, and my database on a second, Server2. When I run the web app on the same server as the database (i.e., web site and database on Server2), I don't receive this error. So, this leads me to believe this has something to do with MS DTC. Is there a limit to how much data MS DTC can manage for a given transaction? If so, is it configurable? When I run my code, the application fails after a certain number of steps (this is repeatable). See sample code below. When I execute the code below, the error occurs on UpdateBody2();. If I comment out UpdateBody2(), the error will now occur on UpdateBody3();, and so on. This leads me to believe that I have hit some upper limit. My code follows a pattern similar to this:using {TransactionScope scope = new TransactionScope()){ UpdateHeader(); UpdateBody1(); UpdateBody2(); UpdateBody3(); UpdateFooter();}Where each of the classes follows a pattern of:UpdateHeader(){ using (SqlConnection conn = new SqlConnection()) { conn.Open(); // Do something conn.Close(); } } Environment:ASP.NET 2.0SQL/2005 StandardWindows Server 2003 Thanks.Steve
I'd like to use MSDE since it's free instead of SQL Server for my database. I will be hosting a portal type site. If all goes well and my site is wildly successful are there any limitations in MSDE that I need to worry about?
For instance, I thought there was a limit to the number of connections. I thought I remember seeing 5 or 50 on the microsoft download site.
If there is a connection limit of say 10, what happens when connection 11 comes in? Does it just wait for a free connection or does it fail?
Is anyone using is as the database on a large portal site? How many users are there in total? How many connected at the same time?
I am under the assumption that it is SQL Server underneath, so I assume the performance and abilities are very good. Is this a fair assumption?
Hi, I was curious if anyone knew if there was a way to get around SQL Express Edition 2005's limitations to support remote connections. If I install SQL Server 2005, will there be a smooth transition between the two on my IIS 5.1 server? My database is extremely simple and only consists of a single table without any complex queries. Idealy, I would like to spend no money since I am a poor collge student doing this as a project. Suggestions?
I updated a varchar field fom 500 length to 800 and now nothing works. Are there limitations to how big a varchar table entry can be and if so what is the alternative?
Hi all there,I'm a newbee to this forum. I've a question, is there any limitationon "IN" clause in select query.for example :"SELECT EMP_ID, EMP_NAME FROM EMPLOYEE WHERE EMP_ID IN('EMP1001','EMP1002','EMP1003', etc, etc, so on)".I've read at some documentation that there is a limitation for Columnsin a table and i.e. we can have only 1024 columns per table, is thistrue?Plz help me !!Thanx in advance.Kind Regards,Harry
I'm thinking about using the FTP task in an integration i am developing, but before i do, i need to get an idea that what i want to do is possible.
With the FTP task could i get a list of directories on the FTP server, and compare the folders with a result set returned from a SQL query. If any of the directories on the FTP site then i want to download each of the relevent directories and there contents to the local machine.
Could someone give me an idea as to if this is possible via the task, or if indeed a script task would be better?
I had a question about the CPU limitations of SQL express. It says limited to 1 CPU but what if I have multiple instaces since the application is a new service and runs in it 'sown memory space how is it limited to 1 CPU? does the service use multiple processors and the queris are binded to one?
With the limitation of 1Gb of ram and 1 CPU for SQL Express, does that mean we cannot load it on a server with multiple CPU's and more ram, or that we can load it and it will only use 1CPU and 1Gb of ram?
As SQL Server XE can run on one PC maximum, does it mean that it can't support multi users??? I mean could it be installed on one server and to be used by multiple users connected to server with their PCs?
I was looking around, and I found this info on SQL Server Express:The 4 GB database size limit applies only to data files and not to log files. However, there are no limits to the number of databases that can be attached to the server.Does this mean I can use several different databases in my website, like one separate db containing forum entries, and one with user details and related stuff?
We are primarily a DB2 and Oracle shop. There are movements towards SQL Server for smaller applications. We would like to know if anyone out there has shop standards in determining which database should go with what type of application. What we are looking for are possibly database sizes, number of concurrent users, hardware and software cost, network infrastructure, security, scalability, and development tools. Any information would be appreciated.