Dimension Loading Best Practice?

Aug 14, 2007



It has been suggested that when loading dimensions, the first step should be to clear the current dimension table and reload in total from the data source. I use integer identity PK fields in all my dimension tables.

When loading my fact table I'm only loading rows that have been added or changed since the last SSIS run.

Wouldn't this introduce the possibility of mis-matched key relationships if, for some reason, the order of the rows in the dimension table changed and a fact row that doesn't get modified in this SSIS run pointed to a surrogate key in a dimension table that was changed?

I suppose in most cases a change in the OLTP that caused a reordering of the dimension table - say a customer was dropped - would show up as a modification in the fact load and cause the row to be updated. However if someone changed the indexing in the OLTP it might cause the customer order to chagne causing them to be loaded in a different sequence with different surrogate keys in the dimension table.

It just concerns me that a possibility for corruption seems to exist in this type of scenerio. It seems like one should load the dimensions and facts in the same manner.

View 9 Replies


ADVERTISEMENT

SSIS Dimension Loading Help.

Dec 5, 2007

Hi,

Can anyone provide me with some sample examples on how to load dimension tables in SSIS. Any help is highly appreciated!

Thanks in advance!

Regards,
Vivek.

View 3 Replies View Related

Dimension Loading Problem

May 27, 2008



I have a dimension table that uses slow chaning type 1 and 2€¦



Source Table

CD Category Object_key

-1 Not Provided NULL

1 CADA 1002

2 DFG 1889

3 GHJ 456

4 CACA 54635



I am using ObjecT_key as a business key

Destination ( Dimenstion)



Key Category business_key

-1 Not Provided NULL

1 CADA 1002

2 DFG 1889

3 GHJ 456

4 CACA 54635

5 Not provided NULL



Problem is When a business_key is NULL , slow changing dimension wizard doesn€™t recorgize whether it€™s a new or existing value, so it keep trying to add a new row..

How can I fix this issue?

View 1 Replies View Related

Problem Loading SSAS Dimension With SSIS

Dec 15, 2007

I am trying to perform an incremental (ProcessAdd) load on a dimension using SSIS and the Dimension Processing data flow component. Whenever I try the load it fails with the following error:

The PrimeOutput method on component "OLE DB Source" (1) returned error code 0xC02020C4

OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:44 PM,12/14/2007 7:44:44 PM,-1055719414,0x,Parser: An error occurred during pipeline processing.
OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:44 PM,12/14/2007 7:44:44 PM,-1055719414,0x,Internal error: The operation terminated unsuccessfully.
OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:44 PM,12/14/2007 7:44:44 PM,-1055719414,0x,Errors in the OLAP storage engine: An error occurred while the 'ORDER LINE STATUS' attribute of the 'DIM ORDER LINE' dimension from the 'ACRDvel' database was being processed.
OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:44 PM,12/14/2007 7:44:44 PM,-1055719414,0x,Errors in the OLAP storage engine: An error occurred while the dimension, with the ID of 'Q DIM ORDER LINE', Name of 'DIM ORDER LINE' was being processed.
OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:44 PM,12/14/2007 7:44:44 PM,-1055719414,0x,Errors in the high-level relational engine. The data source view does not contain a definition for the 'Q_DIM_ORDER_LINE' table or view. The Source property may not have been set.
OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:44 PM,12/14/2007 7:44:44 PM,-1055719414,0x,Internal error: The operation terminated unsuccessfully.
OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:44 PM,12/14/2007 7:44:44 PM,-1073450974,0x,SSIS Error Code DTS_E_PROCESSINPUTFAILED. The ProcessInput method on component "Dimension Processing" (112) failed with error code 0x80004005. The identified component returned an error from the ProcessInput method. The error is specific to the component, but the error is fatal and will cause the Data Flow task to stop running. There may be error messages posted before this with more information about the failure.
OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:44 PM,12/14/2007 7:44:44 PM,-1073450975,0x,SSIS Error Code DTS_E_THREADFAILED. Thread "WorkThread0" has exited with error code 0x80004005. There may be error messages posted before this with more information on why the thread has exited.
OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:46 PM,12/14/2007 7:44:46 PM,-1071636284,0x,The attempt to add a row to the Data Flow task buffer failed with error code 0xC0047020.
OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:46 PM,12/14/2007 7:44:46 PM,-1073450952,0x,SSIS Error Code DTS_E_PRIMEOUTPUTFAILED. The PrimeOutput method on component "OLE DB Source" (1) returned error code 0xC02020C4. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing. There may be error messages posted before this with more information about the failure.
OnError,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:46 PM,12/14/2007 7:44:46 PM,-1073450975,0x,SSIS Error Code DTS_E_THREADFAILED. Thread "SourceThread0" has exited with error code 0xC0047038. There may be error messages posted before this with more information on why the thread has exited.
OnInformation,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:46 PM,12/14/2007 7:44:46 PM,1074016264,0x,Post Execute phase is beginning.
OnInformation,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:46 PM,12/14/2007 7:44:46 PM,1074016265,0x,Cleanup phase is beginning.
OnInformation,*******,Package,{62F748F0-EFC6-4D7A-9873-B31144C54873},{6AA40F9D-7E89-4D68-B2CD-31A43EA6B80D},12/14/2007 7:44:46 PM,12/14/2007 7:44:46 PM,1074016267,0x,"component "Dimension Processing" (112)" wrote 756860 rows.


Apparently this translates to the error DTS_E_ADDROWTOBUFFERFAILED.

This incremental load is trying to add several hundred thousand rows and it appears to get to around 10K or so and then fail. I can't seem to find any KB articles related to this problem. Does anyone have a clue what may be happening here?

I do have an incremental load of the same type that is working but I am continually getting the following warning when the load is running:

"The buffer manager detected that the system was low on virtual memory, but was unable to swap out any buffers. 6 buffers were considered and 6 were locked. Either not enough memory is available to the pipeline because not enough is installed, other processes are using it, or too many buffers are locked."


My server has 8GB of memory and is running WS 2003 x64 and SSAS 2005 x64

TIA

View 4 Replies View Related

Loading Fact Using Dimension W/ Ragged Hierarchy (geography)

Jul 27, 2007

I'm loading a fact table that has several geographic attributes - some are at the state level, some are at the county level, and then some are drilled farther in that that. I understand the basic concept of the dimension with the ragged hierarchy, but unsure of how to load to the fact table using lookups based on these geographic units. For example, if my geographic dimension contains 200 records for the state of Wyoming, basically a record for each fine-grain place (i.e. city/town), then how do I go about doing a county lookup. Wyoming only has 23 counties, but because of the repetitive nature of the dimension attributes that are not at the finest grain, I'll get more records in the lookup than I need. This activity repeats of course while I move up the geographic scale to state, then country. How do I configure/fill my dimension to handle these differing scales of data?

View 2 Replies View Related

Analysis :: Add Dimension To Cube Dimension Without Any Relation In Dimension Usage

Oct 26, 2015

When i add a dimension to the cube dimension without any relation in my dimension usage to any measure group my units are going down.However when i remove the dimension from the cube am getting the correct values.

View 4 Replies View Related

Filtering Out 1 Time Dimension Based On The Selection In Another Time Dimension

May 19, 2008

Hi!

Need some help building a query that does the following :

I have 2 Time Dimensions ; Time (Transdate) and ClosedDate (ClosedDate)

In my report/query, if [Time].CurrentMember = [Time].[YMD].[YMD].[2006].[200610].[20061031] I want to FILTER out all ClosedDate < [ClosedDate].[YMD].[YMD].[2006].[200610].[20061031]

Both Time Dimensions are Year -> Month -> Day and have the same Members.

I have every option available, using calculated Members and/or Measures to do this.

The report I'm creating is Aging of Receivables : Balance / 30 days / 60 days / etc.. But for the Aging, I need to filter like explained above.

Appreciate all help!

Regards,
Stian Bakke

View 3 Replies View Related

The T-SQL Practice

Jun 2, 2004

Hello, everyone:

Does any one know the good T-SQL practice topic web site? It is better to include solutions. Thanks a lot.

ZYT

View 2 Replies View Related

Best Practice For 'dbo'

Sep 3, 2006

When setting up databases for end users, what's the best practice regarding who's the dbo for each individual database - the user itself or a sysadmin?

Does it really have any importance at all who the owner (as defined by 'dbo') is ?

View 5 Replies View Related

What Is Better Practice

May 7, 2007

(Terms)
TermID, Term
1----- Abc
2----- Arcico
3----- Tunic

and
(RelatedTerms)
TermID, RelatedTermID
1 ------ 3
1------ 2
2------ 4


I want to get the following results

1.- a list of all the terms that start with A%
2.- a list of all the related terms … that belong to terms that start with A%


For number 1 - I am doing a select on Terms table with where term like A%.

For number 2 – I am joining both tables and then once again doing a where term like A%.


Would it be more efficient to take the first results and put them in a table variable, and then just do a join with the second table RelatedTerms.TermID = Terms .TermID

The number of records that generally comeback are between 500 to 1000 records that

What would you consider is a better approach ? or maybe there is an even better way ?

View 4 Replies View Related

Best Practice?

Aug 14, 2007

Wasn't sure where to ask this question.

Was wondering what everyone is doing in regards to server/db protection?

Do you run your DB's on independent servers?

Do you run your DB's on clustered servers?

Do you run your DB's on redundant haardware w/ a 3rd party
application for bit-to-bit data replication?

Is anyone placing the DB's on a SAN's?

We're at a crossroads, and looking for a good direction to ensure the DB's are up.

Any insight welcome.
Thanks,
Kerry

View 3 Replies View Related

How I Practice Sql

Dec 11, 2007

hello

i'm a newbie for sql , but i want to learn sql on my own , is there any way that i can learn sql , do i have to download sample database from the internet, do i need to have my own server to play with. Hopefully someone show some lights on this.

regards
sutha

View 2 Replies View Related

Need Some Practice

Apr 10, 2006

Please point me to a web resource from where I can study:1) writing complex queries such as those involving HAVING, mult-levelnested queries, GROUP BY, T-SQL functions2) Joins - a lot of practice3) Stored Procedures, transactions, cursors and triggers - I need someheavy-duty practiceWhere can I get some good practice of the above? Also, please recommenda good SQL Server/T-SQL book in the light of the above requirement.

View 3 Replies View Related

C# CLR Best Practice Example

Jan 11, 2006

Folks - had a look around Google and no surprises, but never found what i was looking for.

I want to see a real work best practice C# Stored Procedure for Sql 2005 (express is what i am using, but don't mind the Sql edition).

Almost everything i see is a "select * from table" which to be honest was my first stored proc many years ago - everything since has been fairly detailed.

I ask as i am sceptical, after years of trying to STOP building Sql queries in code (as it's hellish!) that the CLR technique really makes any kind of a diffence.

If someone has found that it HAS i'd love to hear about it. The thought of:

SqlCommand cmd = new SqlCommand ( "My Whole Stored Proc as Text" );

... doesn't appeal, never mind the potential for debugging syntactical issues and so on.

I was excited by this, until it became something i had to do in a real situation and then i got a little worried. Should i be?

View 7 Replies View Related

I Have A Practice!

Nov 2, 2006

create a table and name it Salary Information. Add an Employee Name and Salary column to the table. Create a column in the Employee table and name it Salary. Create a trigger that updates the Salary table with the employees's name and salary each time u insert data into the Salary column of the Employee table.

I tried but it didnt work!
Pls, help me!

View 4 Replies View Related

Best Practice Of Using .ndf

Mar 12, 2008

Hi, my database is growing over 1Gb, and I only have one .mdf to keep them all. Should I use a secondary data file for my data? Can I do that now? Thanks.

View 5 Replies View Related

New Database: Best Practice

Jan 24, 2007

Good Morning,
I work for a company that has sees alot of people come and go. The one thing I have noticed is that people use their admin accounts to log into SQL and create sp, views and databases.When the user leaves I am stuck with all these objects that are owned by somone no longer working for the company.
So my question to you guys is: What is the best practice to use in creating new objects?
Thanks for your guru-ness!

View 3 Replies View Related

Best Practice For Lookup

Jul 24, 2007

say i have a customer.aspx that allows a user to enter in customer data.
 on customer.aspx, i have dropdownSalesRep which allows the user to associate a sales rep with the customerbut some customers come to directly, and not thru a sales rep, so I want the user to be able to specify "none"
 Is it best to have a dummy record in my SalesReps table called "none" with an ID of say "999", or is there some other better way to deal with this?
 
 

View 5 Replies View Related

Best Practice For SQL Connections And Asp.Net

Aug 7, 2007

Hi.
We have developed as quite simple ASP.Net webpage that fetches a number of information from a SQL 2005 database. We are having some problems though, becuase of a firewall that is beetween the webserver and the SQL server, and I think this is because of bad code from my part. I'm not that experiensed yet, so I'm sure that there is much to learn.
Usualy when I do a query against a SQL database, I do something like this:
Function GO_FormatRecordBy(ByVal intRecordBy As Integer)        Dim dbQueryString As String        Dim dbCommand As OleDbCommand        Dim dbQueryResult As OleDbDataReader        dbQueryString = "SELECT Name FROM tblRegistrators WHERE tblRegistratorsID = '" & intRecordBy & "'"        dbCommand = New OleDbCommand(dbQueryString, dbConn)        dbConn.Open()        dbQueryResult = dbCommand.ExecuteReader(CommandBehavior.CloseConnection)        dbQueryResult.Read()       dbConn.Close()        dbCommand = Nothing        Return dbQueryResult("Name")    End Function 
Now, lets say that I have a DataList that I populate with Integer values, and I want to "resolve" the from another table, then i do a function like the one above. I guess that this means that I open and close quite alot of connections against the database server when I have a large tabel. Is there any better way of doing this? Chould one open a database connection globaly in lets say the ASA fil? Whould that be a better aproch?
When I added the CommandBehavior.CloseConnection to the ExecuteReader statment, I noticed that it was a bit faster, and I think there was fewer connections in the database, so maby there is more to the "closing connections" then I usualy do.
Any tips on this?
 Best reagrds,Johan Christensson

View 6 Replies View Related

Help Me With A Bit Of A 'practice' Issue...

Dec 16, 2003

Ok.

I recently started developing a web site for a client using storefront.net and ms sql server.

the db schema of storefront.net has autonumbers as the PKs for the products table (even though the products table contains an additional field for product_number.)

So here's my dilemma if you care to read:

I typically develop local, deploy remote (after testing). I have a local SQL server, and then the remote SQL server.

When I'm developing for this project, I'll insert data such as products to the products table (sometimes several times while i'm working out routines to import data to the products table.) this has the effect of creating a unique ID for each product based upon SQL auto-incrementing INTs.

This StoreFront.net (SF.NET) has another table that is a lookup table. For each part number, it has a corresponding categoryID number.

Now, if i have product_ID 1234, and I set the category ID to say 10 and get it working on my local box, every thing is fine.

Here's where the problem comes in: When I use DTS to transfer the database during remote deployment, each product is inserted into the remote DBs products table and gets a NEW product id. Same with the categories.

This has the effect of breaking the relationships. (SF.NET has no ref integrity nor relationships defined in the db.) let's say my product_id 1234 gets put into the remote copy, it'll get a new product_ID (PK). let's say it's now 5775. now my category ID will also get a new value. so my data is now not related.


I don't know how to handle this situation. The unique IDs generated on my local sql will nearly almost always be different from those generated on the remote db.

How do i handle this situatoin is my question? advice, guys?

View 3 Replies View Related

Auction Best Practice

Jun 28, 2005

Hi to everyone!   I've to create a little auction system that runs on web. Before starting developing, I'll would like to be sure to use the best practice...The main aspect is to avoid conflicts on database updating with bids, i.e. if a user places his bid I've to be absolutely sure that his bid is the highest at the moment of updating database. If not, I have to refuse it...So I ask you: using transaction is the best way for assuring the non-conflicts? And may I have to be careful of some other aspect in ASP.NET pages? Or there's no problem of conflicts at page level?Thank you very much in advance for any suggestion, and If anyone has some other thing to say about possible problem on auctions I'll be glad to hear him!!!! ;-)

View 6 Replies View Related

TSQL Practice

Jun 4, 2001

Following works fine in QA
master..xp_cmdshell 'osql -Usa -E -h -w250 -Q"set nocount on exec unallocated" -dTest -oc:est.txt' ,no_output

but does not from within a sp
err -
Server: Msg 170, Level 15, State 1, Procedure test, Line 2
Line 2: Incorrect syntax near 'master'.

Thank you.
Ivan

View 1 Replies View Related

Best Practice Question

Aug 20, 2004

I have around 10 databases currently residing on different platforms which make-up for roughly a terrabyte of information. I would like to migrate all of these DBs over so that they are all managed under one instance of SQL server 2K. In my view this streamlines things a lot and reduces costs of licensing/hardware.

However, is managing all of these databases on one clustered instance of SQL 2k the best approach from a performace stand point? Would it be better to seperate each database onto its own machine? I am under the impression that given enough hardware (processors, RAM) using just one instance of SQL 2k enterprise should be enough to perform the mangement of this data. Is this correct? Is there an optimal model?

Money is always a concern but in this case, performance is the main objective. The size of the data managed will be growing significantly so the system should be scalable.

My background is as a developer so I may not have provided enough to give a good answer. Please ask questions if you need more detail. I am looking for suggestions on the best way to handle this.

Specifically I would like to know the preferred architecture as well as any suggested hardware.

Thanks in advance

View 3 Replies View Related

Best Design Practice?

Dec 11, 2004

I'm building a database that has maybe four unique tables Student,
Advertiser, Employee, maybe Account. Three of the four table (Student,
Advertiser, Employee) have something in common in which they all contain
fields such as emailAddress, password, role, isAccountActive, etc. which
allow them to access their respected data. However, is it best practice to
build a fourth table which contain Account information or should I just
include that information in their respected tables?

My thinking is that if you have a fourth table such as Account then you can
manage all accounts (Student, Advertiser, Employee) from one table, but as
the database gets more in-depth you have to build more and more complex
stored procedure to do simply task such as update, delete, select, etc.

View 11 Replies View Related

Best Practice For This SP Scenario !

Oct 10, 2005

Hello All ..
This is the scenario I'm having :
-- I'm a beginner so bear the way I'm putting it ... sorry !

* I have a database with tables
- company: CompanyID, CompanyName
- Person: PersonID, PersonName, CompanyID (fk)
- Supplier: SupplierID, SupplierCode, SupplierName, CompanyID (fk)

In the Stored Procedures associated (insertCompany, insertPerson, insertSupplier), I want to check the existance of SupplierID .. which should be the 'Output' ...

There could be different ways to do it like:
1) - In the supplier stored procedure I can read the ID (SELECT) and :

if it exists (I save the existing SupplierID - to 'return' it at the end).
if it doesn't (I insert the Company, the Person and save the new SupplierID - to 'return' it at the end)
------------------------------------
2) - Other way is by doing multiple stored procedures,
. one SP that checks,
. another SP that do inserts
. and a main SP that calls the check SP and gets the values and base the results according to conditions (if - else)

3) it could be done (maybe) using Functions in SQL SERVER...

There should be some reasons why I need to go for one of the methods or another method !
I want to know the best practice for this scenario in terms of performance and other issues - consider a similar big scenario ..... !!!

I'll appreciate your help ...
Thanks in Advance . ! .

View 1 Replies View Related

Best Practice Advice

Jan 16, 2008

Need the following question addressed, as it keeps coming up in our development meetings and has been creating a divide. Pease voice your opinion.

To keep it simple, we have Table1 which identifies several questions that are revised on a regular basis. One of it's columns is called "Revision Status". Within revision Status, we would like to identify the possible status of a question such as:

New Questions;
Revised;
Resubmit;
Inactive;
Active;

...as well as several more.

I'm of the mind to have these in a seperate table identified with a unique ID... call it StatusTable.

Such as:

1 New Questions;
2 Revised;
3 Resubmit;
4 Inactive;
5 Active;

However others feel, just use the "Revision Status" column and simply use the numbers "WITHOUT" a table or description. The developer documentation will tell the developer which number equals the description. ie the following would be found in the Revision Status column.

1
2
3
4
5


My mind says the above is ilogical. I would rather join and say in my statement:

WHERE StatusTable.Status = 'Inactive'

Where the other way would be

Where [Revision Status] = 4


I hope i'm not being thick-headed.

Please advise.

View 10 Replies View Related

Best Practice Advice Please

Jul 20, 2007



I'm looking for advice on the best way to stop stored procedures and CLR assemblies from being

copied from their originally installed server to a different server, for the same company or even copied

to another company.



Are there established ways for achieving this level of protection.



Also, I was hoping that encrypting stored procedures would be a 100% reliable way to stop

malicious copying of the code. But I have read that this is not the case. Any advice in this

area would also be appreciated.



Thanks

Steve

View 4 Replies View Related

SQL Server Best Practice With IIS

Aug 9, 2006

Hi Everyone

We are building a new site using ASP and SQL as the backend.
Any idea where I can find some information about best practices to coonect the 2.
I was thinking about IIS on a DMZ with port 80 only open and the SQL inside the internal network and open port 1443 between them.

Any ideas will be welcomed.

Thank you

Oren Levy

View 2 Replies View Related

Best Practice To Do An Update

Mar 27, 2008

What is the best way to do this:

I have a table (TableA) say below with 6 colunms

CPubID --------CCTable ---------CCField--------------- DWTable----- DWField
101 XYZ Type NULL NULL
102 XYZ ClaimAssocType NULL NULL
103 XYZ ID NULL NULL
104 XYZ ID NULL NULL
105 XYZ PublicID NULL NULL
106 XYZ PublicID NULL NULL
107 XYZ PublicID NULL NULL
109 XYZ PublicID NULL NULL
201 XYZ PublicID NULL NULL
301 XYZ PublicID NULL NULL

and a Table (TableB) with three Colunm

CPubID --------DWTable ---------DWField---------------
101 T1 F1
102 T1 F2
103 T3 F1
104 T1 F3
105 XT1 F4
106 T1 F5
107 T2 F8
109 T1 F9
201 T1 G1
301 T2 G2
001 T1 F1
002 T1 F2
903 T3 F1
904 T1 F3
905 XT1 F4
706 T1 F5
307 T2 F8
409 T1 F9
801 T1 G1
301 T2 G2


I want to move data of TableB to TableA depending upon the CPubID and TableA should finally look as below

CPubID --------CCTable ---------CCField--------------- DWTable----- DWField
101 XYZ Type T1 F1
102 XYZ ClaimAssocType T1 F2
103 XYZ ID T3 F1
104 XYZ ID T1 F3
105 XYZ PublicID XT1 F4
106 XYZ PublicID T1 F5
107 XYZ PublicID T2 F8
109 XYZ PublicID T1 F9
201 XYZ PublicID T1 G1
301 XYZ PublicID T2 G2


It needs to update the TableA with respective data in Table B






View 3 Replies View Related

Best Practice To An Update

May 5, 2008



Please advice on the best way to this


Create Table TEST

(

[ColID] [varchar](50) COLLATE SQL_Latin1_General_CP1_CI_AS NULL,

[SubColID] [varchar](50) COLLATE SQL_Latin1_General_CP1_CI_AS NULL,

[Table1] [varchar](55) COLLATE SQL_Latin1_General_CP1_CI_AS NULL,

[Field1] [varchar](55) COLLATE SQL_Latin1_General_CP1_CI_AS NULL,

[Key1] [int] NULL,

[Key2] [int] NULL

)

INSERT INTO TEST (ColID, SubColID, Table1, Field1, Key1, Key2)

VALUES ('101', '1001', 'Payment', 'Amount', '','')

INSERT INTO TEST (ColID, SubColID, Table1, Field1, Key1, Key2)

VALUES ('101', '1001', 'Payment', 'ID', '','')

INSERT INTO TEST (ColID, SubColID, Table1, Field1, Key1, Key2)

VALUES ('101', '1001', 'Payment', 'CheckPublicID', '121','1221')

INSERT INTO TEST (ColID, SubColID, Table1, Field1, Key1, Key2)

VALUES ('101', '1002', 'Payment', 'Amount', '','')

INSERT INTO TEST (ColID, SubColID, Table1, Field1, Key1, Key2)

VALUES ('101', '1002', 'Payment', 'ID', '','')

INSERT INTO TEST (ColID, SubColID, Table1, Field1, Key1, Key2)

VALUES ('101', '1002', 'Payment', 'CheckPublicID', '131','1321')



select * from test



101 1001 Payment Amount 0 0
101 1001 Payment ID 0 0
101 1001 Payment CheckPublicID 121 1221
101 1002 Payment Amount 0 0
101 1002 Payment ID 0 0
101 1002 Payment CheckPublicID 131 1321



I want to replace the 0's and populate with respective keys as below

101 1001 Payment Amount 121 1221
101 1001 Payment ID 121 1221
101 1001 Payment CheckPublicID 121 1221
101 1002 Payment Amount 131 1321
101 1002 Payment ID 131 1321
101 1002 Payment CheckPublicID 131 1321

View 6 Replies View Related

Best Practice Guidelines

Jun 28, 2006

Hi All,

We have an application requirement for a database supporting field service engineers, which calls for a central SQL Server databse, and laptops with the same database replicated onto SQL Express. I'm resposible for designing the database for this, physical and logical. I've designed and built many a database, but never had to use replication before.

I've read through BOL, and understand how the merge replication process works, and I have no problem designing the database assuming it were to run on a single server.

What I am trying to find are whitepapers, or equivalent, on "best design and implementation practice", and especialy common mistakes to avoid.

I know that the windows programmers responsible for the UI will not completely abstract the database from the code (no matter how desirable that is or how often I tell them!), and I really don't want to find I have to change the physical tables or replication logic after they've coded most of the UI .

Many thanks in advance

Richard R

View 3 Replies View Related

Best Practice Question

Jul 30, 2007

I'm very sorry if this has been covered before but I can't seem to find and answer with the way I've been searching.

The question is what would be concidered best practice when I have a row of data that I need to add or update but I don't know if the row exists yet. Would it be best to call SELECT for the record and INSERT if no data is found and UPDATE if the record is found, or to attempt the INSERT and then attempt an UPDATE if the insert fails?

View 7 Replies View Related

How I Can Get SQL Table To Practice?

Jan 23, 2008



Hi,
To all, How I can get SQL table to practice?
Regards,
Sherazi

View 1 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved