Slow Changing Dimension Help

Apr 5, 2006

Hi there

I have a dimension table with effective dating; I'm loading historical transactional data and want to associate the correct surrogate key from the dimension with the fact table transaction. My dimension table has a start date but no end date, the end date is assumed by the start date of the next record with the same id. So I want the surrogate key which is the most current but whose date is not before the transaction date. I know I need a subquery but not too sure how to write it.

Thanks in advance.

Derek

View 1 Replies


ADVERTISEMENT

Slow Changing Dimension - A Question

Nov 30, 2006

Greetings SQL friends,

I am using the Ralph Kimball approach of having DIM_Effective_date, DIM_Expirey_date and Current_Flag.

When I use the SCD transformation it works very well for populating the Effective and Expiry dates but how do I make it insert an 'O' for the Current Flag (for expired records) and a 'C' for non expired dimension records?

Your advice would be much appreciated.




View 1 Replies View Related

Add Record Update Date In Slow Changing Dimension

Dec 6, 2007

Hi All,

I would like to know whether it is possible to add and updated date column in a slow changing dimension table using the slow changing dimension data flow transformation.

I would like to keep track of what record is updated in the dimension table based on the data being processed.

Thanks for you help and information


Regards,
Fadzli

View 1 Replies View Related

Validating Start And End Dates In Slow Changing Dimension Table

May 4, 2007

Has anyone written or come across a routine that validates start and enddate in a slow changing dimension? (eg verifies there is no overlap in any of the records).



thanks in advance

View 3 Replies View Related

Dates In Slow Changing Dimension Dataflow Transformations Component

Mar 31, 2008

I€™m trying to populate a table with fields of date type [DT_DATE] using the Slow Changing Dimension Transformation component. When I add the date fields to the component it would not build the stream. The wizard fails and tells me the date fields are not of the same type. The fields in the destination table are of type €œdate€? and the input columns are of type [DT_DATE]. Am I missing something?

View 3 Replies View Related

SSAS Dimension Browsing Is Slow

May 19, 2008



Hello,

I've run into a problem building a new olap cube. It's taking 15 minutes or more when pulling on a dimension before I've even pulled on any measures. This happens in Excel as well as Management Studio. When it does display it seems to only show dimension members that it has data for. In addition, running profiler seems to indicate that it queries each of my partitions when it's doing this. I know that in my last cube if you pulled on a dimension, no members would show up yet and it would be fast. However I can't seem to find the property that's telling it not to query the partitions to determine what members to show. Any ideas?

~Grant

View 6 Replies View Related

About Slowly Changing Dimension

Oct 25, 2007

I have one question regarding Slowly Changing Dimension component in SSIS. Does SCD also delete records in warehouse if they does not exist in source anymore, or does SCD only insert new and update existing records? Can someone explain me a little bit more about inferred members? Thanks.

View 1 Replies View Related

SQL 2012 :: Fast Changing Dimension

May 18, 2015

I have a table that needs to be incorporated into the data warehouse.The table has the following schema.

CREATE TABLE [dbo].[Consignment](
[Id] [int] IDENTITY(1,1),
[BooingID] INT
[BookingDate] [datetime] NULL,
[CarrierServiceName] [nvarchar](255) NULL,
[CarrierServiceCode] [nvarchar](255) NULL,

[code]...

This Table has the same granularity as the fact table as it’s one row per booking.However due to the nature of the data I would not want to incorporate this into the fact table.The Originating and Destination addresses are populated for each booking and are required for reporting.

Question:Should this be moved into a fast changing Dimension table.? or would there be a better way to incorporate this data.

View 1 Replies View Related

Type I Slowly Changing Dimension

May 16, 2008

I have a Type I SCD situation, ie, insert if new (by checking the business ID) and update if any attributes for a given business ID has changed.

The way I usually do this, (and I believe this is how most people do it), is I use a LOOKUP TASK to determine if the business ID exsist in the target table. If it doesn't then I insert. If the business ID exists, then I bring back the associated attributes and use a CONDITIONAL SPLIT TASK to compare if any of the incoming attributes are different. If there are changes, then update.

In doing this comparision, I often run into situations where I end up comparing a NULL value to something, which does not result in FALSE, but a NULL result. To get around this, I first check for NULLs and convert them into something valid before I do the comparision, but this results in a messy comparison expression, especially if I have to compare a lot of attributes.

So, how do you guys handle this?

As an alternative, I am looking into the SLOWLY CHANGING DIMENSION TASK, which I also have some questions on, but I would like to first address the above. Thank you.


View 11 Replies View Related

Package With Slowly Changing Dimension.

Jan 15, 2008

I have a package using Slowly changing dimension in the data flow task. It works fine if the number of records are less but for a large file the package fails with the "Violation of Primar Key" error even though there are no duplicate records in the table.
for eg i have a table with employee database with a composite primary key comprsing of Name and Employee Id. I need to do an UPSERT depending on the Name and Employee Id combination. I have a file with 100,000 records and when i try to execute the package it gives an error 'cannot insert duplicate data' even though the combination does not exist in the database.

Please help.

Aashna Behl

View 3 Replies View Related

Slowly Changing Dimension With 600,000 Rows

Jan 17, 2006

Hi,

We have been using tasks generated from the SCD wizard. We have smaller dimensions (< 30,000 rows) that work well. Our Product Dimension package is giving us performance problems (taking 7 hours to do 600,000 rows when 80,000 records are updated; the rest new inserts). It is similar to the smaller dimensions. Several columns are type 1 and are doing update statements; several are type 2 doing updates and inserts. The package had a complicated view as the initial task, but we have since modified to use a SQL command with variable and now the initial read appears quick, but is chunking in 10,000 record increments and taking the 7 hours (never let finish previously). So the package is pretty basic now (reading a source, a small derive and data conversion, a small lookup (cached 30,000 records) for a description, then the SCD). Before I start replacing what the SCD generates with stored procedures, anyone have any suggestions as to what might be the issue? We believe we have increased the number of type 2 columns and the SCD definately has more to do than just an insert or update, but 7 hours for 600,000 records seems excessive. Interestingly, the source task never turns green. Previously when we had a Merge Join it completed the read and bottlenecked at a sort and a Merge Join. Now that has been removed and simplified, and all tasks remain yellow with the 10,000 (actually 9,990 I think) chunks appearing at the source, then the SCD before the next chunk appears to be read. On the general release (not the beta). Thanks in advance!

View 3 Replies View Related

Slowly Changing Dimension [58] Error

Sep 21, 2006

i am using SCD to insert or update .my source and destination table are Oracle and i am using Orcale OLEDB provider . i am getting the following error while executing the package.what could be the solution

[Slowly Changing Dimension [58]] Error: An OLE DB error has occurred. Error code: 0x80040E5D. An OLE DB record is available. Source: "Microsoft OLE DB Provider for Oracle" Hresult: 0x80040E5D Description: "Parameter name is unrecognized.".


Thanks

Jegan

View 3 Replies View Related

Parent ID In A Slowly Changing Dimension

Dec 7, 2006

Hi There,

Just wondering if any of you implemented a (Kimball type 2) dimension structure, in which a ParentID column exists which points to a record from the same dimension table, using a SCD objects in SSIS. The ParentID column would have to be "Historical".

The challange here is that you would need to go through the table twice somehow, because if I would do a lookup of the parent record in the first run, I wouldn't be sure if I got the right parent record.

Thnx, Jeroen.

View 13 Replies View Related

Slowly Changing Dimension Wizard

Jun 6, 2007

I have a Company Dimension table that consists of various sources. One source will provide me address information, another source will provide industry info, etc. I created a historical load package that will pull all of this together so that I have all the necessary data related to a company in one record. All is well.



Since my company data is coming from various sources, how can I tell the SCD to update certain fields but not others for a type 2 change. In essence, I would like to "pull forward" the data that was in the original database row and then update it with only the changes coming from the proper source data. For example, if an address changed I will get the new address from the source but will not have the industry info. I would like to create the new record with the new address but also keep the industry data in tact. Is this possible?



Currently I will get the new record with the new address but will have null values for the industry data.







Thanks



View 1 Replies View Related

Slowly Changing Dimension Freezes

Apr 9, 2008

I am trying to move data from a transactional database to a data warehouse using a slowly changing dimension. The transactional data comes from a view in SQL server that takes <60 seconds to run and returns about 60k rows. The warehouse table is currently 80k rows long (and growing), and contains 7 historical (type 2) dimensions. When I execute the package in BIDS the DataFlow Task begins to execute, and shows that between 20k and 30k rows have been pulled from the data source into the SCD Transform in the first hour before it simply stops doing anything. This is not to say execution stops; it continues. There is no error thrown. No warning given. System resources are 98% free. The database is not being hit at all. And yet, I have let the package sit 'still' as it were for over 8 hours, and nothing ever happens.

Here is a copy of one log:

starttime endtime message
4/8/2008 9:36 4/8/2008 9:36 Execute phase is beginning.

4/8/2008 9:36 4/8/2008 9:36 PrimeOutput will be called on a component. : 1715 : Union All
4/8/2008 9:36 4/8/2008 9:36 A component has returned from its PrimeOutput call. : 1715 : Union All
4/8/2008 9:36 4/8/2008 9:36 PrimeOutput will be called on a component. : 2912 : Staged Queues
4/8/2008 9:36 4/8/2008 9:36 Rows were provided to a data flow component as input. : : 2970 : DataReader Output : 70 : Slowly Changing Dimension : 81 : Slowly Changing Dimension Input : 9947
4/8/2008 9:37 4/8/2008 9:37 A component has returned from its PrimeOutput call. : 2912 : Staged Queues
4/8/2008 9:37 4/8/2008 9:37 A component has returned from its PrimeOutput call. : 2912 : Staged Queues
4/8/2008 9:59 4/8/2008 9:59 Rows were provided to a data flow component as input. : : 1718 : New Output : 1715 : Union All : 1716 : Union All Input 1 : 3825
4/8/2008 9:59 4/8/2008 9:59 Rows were provided to a data flow component as input. : : 1688 : Historical Attribute Inserts Output : 1682 : Get End Date : 1683 : Derived Column Input : 645
4/8/2008 9:59 4/8/2008 9:59 Rows were provided to a data flow component as input. : : 1702 : Derived Column Output : 1692 : Update End Date : 1697 : OLE DB Command Input : 645
4/8/2008 10:01 4/8/2008 10:01 Rows were provided to a data flow component as input. : : 1759 : OLE DB Command Output : 1715 : Union All : 1758 : Union All Input 2 : 645
4/8/2008 10:01 4/8/2008 10:01 Rows were provided to a data flow component as input. : : 2970 : DataReader Output : 70 : Slowly Changing Dimension : 81 : Slowly Changing Dimension Input : 9947
4/8/2008 10:24 4/8/2008 10:24 Rows were provided to a data flow component as input. : : 1718 : New Output : 1715 : Union All : 1716 : Union All Input 1 : 3859
4/8/2008 10:24 4/8/2008 10:24 Rows were provided to a data flow component as input. : : 1688 : Historical Attribute Inserts Output : 1682 : Get End Date : 1683 : Derived Column Input : 641
4/8/2008 10:24 4/8/2008 10:24 Rows were provided to a data flow component as input. : : 1702 : Derived Column Output : 1692 : Update End Date : 1697 : OLE DB Command Input : 641
4/8/2008 10:26 4/8/2008 10:26 Rows were provided to a data flow component as input. : : 1759 : OLE DB Command Output : 1715 : Union All : 1758 : Union All Input 2 : 641
4/8/2008 10:26 4/8/2008 10:26 Rows were provided to a data flow component as input. : : 2970 : DataReader Output : 70 : Slowly Changing Dimension : 81 : Slowly Changing Dimension Input : 9947
4/8/2008 10:49 4/8/2008 10:49 Rows were provided to a data flow component as input. : : 1718 : New Output : 1715 : Union All : 1716 : Union All Input 1 : 3969
4/8/2008 10:49 4/8/2008 10:49 Rows were provided to a data flow component as input. : : 1688 : Historical Attribute Inserts Output : 1682 : Get End Date : 1683 : Derived Column Input : 662
4/8/2008 10:49 4/8/2008 10:49 Rows were provided to a data flow component as input. : : 1702 : Derived Column Output : 1692 : Update End Date : 1697 : OLE DB Command Input : 662
4/8/2008 10:49 4/8/2008 10:49 Rows were provided to a data flow component as input. : : 1793 : Union All Output 1 : 1787 : Get Start Date : 1788 : Derived Column Input : 9947
4/8/2008 10:49 4/8/2008 10:49 Rows were provided to a data flow component as input. : : 1814 : Derived Column Output : 1797 : Insert Destination : 1810 : OLE DB Destination Input : 9947
4/8/2008 15:34 4/8/2008 15:34 The pipeline received a request to cancel and is shutting down.
4/8/2008 15:34 4/8/2008 15:34 Thread "WorkThread1" received a shutdown signal and is terminating. The user requested a shutdown or an error in another thread is causing the pipeline to shutdown.
4/8/2008 15:34 4/8/2008 15:34 Thread "WorkThread1" received a shutdown signal and is terminating. The user requested a shutdown or an error in another thread is causing the pipeline to shutdown.
4/8/2008 15:34 4/8/2008 15:34 The pipeline received a request to cancel and is shutting down.
4/8/2008 15:34 4/8/2008 15:34 Thread "WorkThread1" has exited with error code 0xC0047039.
4/8/2008 15:34 4/8/2008 15:34 Thread "WorkThread1" has exited with error code 0xC0047039.


Notice the time difference between the last OnPipelineRowsSent event and the first OnError event (when I clicked the stop button): 5 hours! In all that time, SSIS did not log a single event, or use more than 2% of my processor or exceed 1GB page file or hit the database even once! I am assuming this means it is simply not doing anything. It is not failing, nor is it executing, it is just sitting there.

Has anyone experienced a similar problem? Does anyone know how I might troubleshoot this? Thanks in advance for any help, and let me know if I need to clarify. Also, I am new to SSIS, so if I am missing something obvious, go easy on me! Thanks.


Mitch Connors

View 6 Replies View Related

Slowly Changing Dimension.. Type 1.5

Aug 28, 2006



The warehouse I am writing packages for has sort of a "Type 1.5" design for most of its DIMs that I am trying to get to work with the slowly changing dimension object.

Basically it should behave like a type 1 with updates in place BUT send the old prior rows/values to an "archive" server to hold the historical data. Unlike a Type 2 this data will not be used for any processes - but it needs to be kept for historical reseach and auditing.

Any ideas to easily do with the SCD wizard? I thought using the wizard as a type 1 and then after the wizard is done attaching the "Historical Attribute Inserts Output" to the archive db/table would do the trick but that output from the SCD object never has data. I could manually do it with a lookup and so forth but I thought I'd check in here first to see if I am just overlooking something with the SCD object.

View 3 Replies View Related

Slowly Changing Dimension Component Help

Apr 1, 2008



Is there a way to change the data source for a SCD Component without having to go back to reinsert the matches for Source and Destination columns. Note the underlying data table hasn't changed, just the server the table resides on. Whenever I change the data source I am noticing that I have to painfully go back and match columns one by one.

Thanks
David

View 3 Replies View Related

Slowly Changing Dimension Question

Sep 1, 2005

Hi.

View 9 Replies View Related

Slow Moving Dimension And History Time Stamp

Nov 14, 2007

The following question might sound a bit stupid but I'm not a database expert so hopefully nobody minds me asking it.

Here's what I did:

1. I created an SSIS package that is supposed to import new data into my data warehouse as it becomes available.

2. Since I need to maintain some of the history I use the Slow Moving Dimensions part (set the history flag on input fields) but run into an error condition while running the package. The message basically says that I'm about to create a duplicate record which is not allowed.



Original Table1:

PK1 field1
PK2 field2
PK3 field3
field4
field5

-----------

Now I enhanced it like this:

Extended Table1:

PK1 field1
PK2 field2
PK3 field3
field4
field5

CreateDate (new)
NewDate (new)
ActiveFlag (new)
_____________________

Now on some records the package is supposed to archive history by populating the (new) fields. In order to keep the record unique (primary key constraint) thought, do I need to make the (new) fields primary keys as well?

So I guess I'm struggling with a more basic concept;)

I would appreciate if somebody could shed some light on this.

Thanks in advance.
Dirk

View 3 Replies View Related

Problem With Slowly Changing Dimension-transformation

Jun 7, 2006

Hi,

I have a problem with the SCD-transformation in SSIS. I have a variable that holds the batchid for the current batch and I want to add this variable to the datapipline in the Data Flow Task.

This is done by using a Derived Column, so far so good. The problem occurs in the Slowly Changing Dimension transformation where I do som evaluations of changed columns BUT I don´t want to do any evaluation of the batchid-variable because then all historical batchid will be updated.

I only want to update the batchid for row that have changed in the current batch.

Is it possible to do this in any way without adding the Derived Column after the SCD transformation??

Thank for any help!!

Patrick

View 3 Replies View Related

Slowly Changing Dimension Index Usage

May 22, 2007

We're using slowly changing dimensions to control a number of data tables in our system. Each table has five or six business keys, but the indexes of the tables are built so they're as efficient as possible (i.e. the fields with the highest diversity are listed first). How does the SCD wizard determine the order of the business key fields? Is there a way I can view or manipulate the statement the SCD task is using to make sure either (a) the indexes match the statement, or (b) the statement matches the indexes?

View 1 Replies View Related

Slowly Changing Dimension Wizard. Is It Bobbins?

Feb 12, 2007



On the face of it, the Slowly Changing Wizard seems a great idea, but is it really any good in the longer term? After you have selected your Type 1 and/or Type 2 changing columns ... then you have "customised" the generated data flow and carefully saved it ... how do you change the column definitions in the future?

If I want to add another column and make it Type 2 how do I go about it? If I run the "Wizard" doesn't this just destroy all my previous customisation work?

I have tried looking in the .dtsx xml file ... hmmm. I notice it's not really editable without some inside information. All those magic numbers in there ... I've fallen foul of them in the past with cut/paste or trying to INSERT a data flow task on the IDE ... luckily I back up my projects on a regular basis. I now have quite a large collection of Projects with the suffix "_corrupt". Are things going to get better?

Hasta SP2, eh?

View 8 Replies View Related

Error While Using A Slowly Changing Dimension Transformation

Jun 6, 2007

I have a SSIS package which contains a number of slowly changing dimension transformations. While the majority work I have one which gives me the following error 'Error: The variable "System::LocaleID" is already on the read list. A variable may only be added once to either the read lock list or the write lock list. '. This error only occurs if the destination table holds data. If I truncate the table and reload the data then the package complete successful. The only difference I can see between this dimension transformation and the other dimension transforms is that the one in question has 2 business keys while the rest have 1.



Can anyone shed light on this?



Thanks



View 3 Replies View Related

Slowly Changing Dimension Transformation Believes 'Å“' Is 'oe'?

Apr 1, 2008

Hi,

We're currently running into an interesting situation where it seems that a Slowly Changing Dimension Transformation believes that 'Å“' (ASCII #156) is the same as 'oe' (ASCII #111 + ASCII #101).

To make a long story short, one of our integration package updates some Product table based on the result of a Slowly Changing Dimension Transformation with three outputs: Unchanged, New and Changing Attribute Updates. Among the columns leading to a row redirection in the Changing Attribute Updates output is some French Description column defined in SQL Server as a varchar(60) (external column) and in the SSIS package as a (DT_STR, 60, 1252) (input column). Now, when the SCD Transformation compares the word 'coeur' (external column) with the word 'cœur' (input column) in this French Description column for a given row, the row is redirected to the Unchanged output (no other columns changed...) instead of to the Changing Attribute Updates output.

Is my example clear enough? Any idea what could explain this unfortunate result? Note that from a strict French point of view, 'Å“' instead of 'oe' is a typographical fantasy and that in my example above, the word 'coeur' ("heart") is really spelled 'c' + 'o' + 'e' + 'u' + 'r', but we're talking programmed comparison here, not linguistic, right?

Any comment will be appreciated.

Thanks,

AL

View 1 Replies View Related

How SQL Server Handles Slowly Changing Dimension

Apr 3, 2006

In our application we have created a SSIS package which extracts data from staging table and places the same in destination table. We have created a slowly changing dimension for the same. Slowly changing dimension uses a composite business key of two columns to decide whether it is a old record or a new record.

Problem : On execution of the package it copies duplicate records with same business keys instead of updating the same. Also the same does not happen for all records. For few records update works fine but for others it inserts a new duplicate record.



I will appreciate if anybody can guide me where I am doing something wrong.





Thank you

View 2 Replies View Related

Bug With Slowly Changing Dimension Task And Nvarchar(max)

Apr 18, 2006

Hello
I use SSIS to load a Unicode file into a single table
I Use a "slowly changing dimension" task to load the destination table and when i map a column (DT_WSTR) to a column with the datatype nvarchar(max) i have an error message that say that i can't map theses columns because there have not the same datatype.

I find a workaround : i map all my cols except the colums that must fill the cols with datatype nvarchar(max) , and after i modify manually the 2 subtask generated by the "slowly changing dimension" task (the insert and the update) and with this way i don't have error messages
It works fine but is it the good way?

It seems to be a bug from ssis?

Thanks in advance.

Nicolas Lievain

View 1 Replies View Related

Slowly Changing Dimension Transform And SCD Original ID

Jan 18, 2008

I have as simple IS package with a Flat File input and SCD Transform. The package is being used to load a Dimension Table, obviously.

The Table was designed in Analyical Services using the Dimension Wizard which very nicely adds End Date, Start Date, Status and Original ID to the Dimension Table design. I then Generate Relational Scheme... and Viola there is a Dim Table in my data warehouse database!

I next attempt to use the IS package with SCD Transform to load said Dim Table. The SCD Transform Wizard does a nice job of using the End and Start dates and Status fields in my new Dim Table, but I can not find any way to use the Original ID field?

I guess I was expecting that since the SCD Transform Wizard 'knows' how to use all the other fields in this Dim table why not the Original ID field? I know what the Original ID field is for - that's not the issue. I just can not figure out how to 'make' my new IS Data Flow with said SCD Transform use the Original ID field ?

Anyone ?

View 3 Replies View Related

Alternative For Slowly Changing Dimension (SCD) Object

Aug 19, 2007



Hi,

I think slowly changing dimension object is not a good choice to update my dimension. It's running slower than I expected. my dimension records has surrogate keys from a control table that SCD is looking up whenever it encounters a new record.

Any alternative I can use?

View 8 Replies View Related

SQL Server 2014 :: Columnstore And Slowly Changing Dimension

Sep 30, 2014

I have an existing SSIS that uses SCD on a rather large table. I am also migrating to SQL2014 Enterprise. My question is, is it possible to use SCD and Columnstore(CS) together? I know that I can drop the CS and add a business key then insert data. Then drop the business key and add the CS back.

View 2 Replies View Related

Slowly Changing Dimension Task CRASHES SSIS

Apr 5, 2006

We have successfully built a SSIS ETL implementation for a data mart.  Most of our dimension loads are using the slowly changing dimensions task.  We successfully built every package and had 1 complete successful load.

We have made a modification to one of the tables that now has 16 mil records in it.  I open the dimension's package and it takes 30 sec - 1 minute to validate.  When I go to the Data Flow tab and attempt to open the Slowly Changing Dimension task, SSIS will freeze up.  I will get no response for hours.  I my testing so far SSIS has yet to come back.

My CPU usage is below 5%.  devenv.exe memory usage is at 113 Mb. My computer is using 400mb / 1Gb of memory.   The status bar at the bottom says "ready". I have let it sit for up to 2 hours with no results.  When I kill the process in the Windows Task Manager it ends instantly.

I have also tried deleting the task and adding a new Slowly Changing Dimension Task.  When I select the table from the drop down I get the same results.  I have noticed this with all my million + record tables that use the slowly changing dimension task but not the smaller tables.
I am also seeing the problem when I am in "Work Ofline" mode.

I am accessing a SQL 2005 server over a VPN connection.  I am connecting to the server using a Native OLE DB OLE DB for SQL Server connection.  My computer is running Windows 2000 SP4,  Intel Pentium M1.86 Ghz, 1 Gb of RAM.

Can a Slowly Changing Dimension Task be used on large tables?  I can understand the process taking a while to run but I do not understand why I can not edit the task.

 

View 3 Replies View Related

Dimensional Modeling: Age As A Slowly Changing Dimension Attribute?

Oct 8, 2007



Hi all

Probably not the right forum - pointers would be appreciated - but I'll give it a try anyway:

I'm in the process of designing a relational database to be used in a BI scenario - ie. dimension and fact tables. The data will eventually be used to feed cubes in Analysis services, however end users will probably be allowed to run reports aginst views of the relational database.

I'm currently looking at the employee dimensions and my first try would designate AGE as a SCD Type 2 attribute. As a result every employee gets at least one new record every year as AGE increases. Given that BIRTHDATE is specified should I drop AGE from the tables and recreate it as a computed attribute in database views and/or cubes?

Regards, Steen

View 7 Replies View Related

Slowly Changing Dimension - Adding New Rows When No Change

Apr 16, 2007

Hi , I am using the Integration Services slowly changing dimension to move data from a SQL Server 2000 database table to a SQL Server 2005 table.

The problem is the package is not tracking changes, it is spending a lot of time doing lookups (i'm guessing this cost it's real slow), but ends up creating new records when there has not been a change.

I'm quite sure the business key is set up correctly (I'm using the PK from the source table).

The database I am transferring from has non Unicode data types (ie varchar and char) and the destination database has Unicode data types (ie nvarchar).

Also some of the fields in the dB are NULL - does this have an effect (ie one null doesn't equal another null)? Or shouldn't that matter?

View 4 Replies View Related

Slowly Changing Dimension:inferred Members Update

Jul 12, 2006



Hi,

Anybody who knows how inferred members update function in slowly changing dimension? I came across this when I started using some of the functions of the Slowly changing dimension object.

Thanks!

cherrie

View 4 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved