Data Warehouse Considerations - SCD Type 2 vs Overwrite Dimension Tables
How we handle data warehousing updates to dimension tables is crucial and this article covers Slowly Changing Dimensions versus overwriting tables.
2024-12-20
How we handle data warehousing updates to dimension tables is crucial and this article covers Slowly Changing Dimensions versus overwriting tables.
2024-12-20
How much load do real time updates place on a data warehouse? Steve has a few thoughts on the analysis of the Amazon Redshift dataset, Redset.
2024-10-07
165 reads
Learn how you can model days in a dimension that might need to be aggregated in different ways for your data warehouse operations.
2024-02-07
1,967 reads
This article gives an overview of Amazon Redshift, the cloud data warehouse in AWS.
2023-07-24
25,534 reads
In 2019 Canadian Broadcasting Corporation (CBC) news reported a massive data breach at the Desjardins Group, which is a Canadian financial service cooperative and the largest federation of credit unions in North America. The report indicated, a "malicious" employee copied sensitive personal information collected by Desjardins from their data warehouse. The data breach compromised the […]
2022-03-30
3,098 reads
Agile data warehousing can be challenging. Pairing the right methodologies and tools can help. Here is how my team met the challenge by using Data Vault methodology and BIML scripting.
2014-09-04
2,592 reads
2014-07-24
1,633 reads
In this article, Arshad Ali goes intp detail about how a data warehouse is different from operational data store and the different design methodologies for a data warehouse.
2013-07-03
6,374 reads
A primer on how to reduce network and source system load when reading a relational source into the data warehouse.
2012-07-30
5,529 reads
SQL Server 2008 introduced many new functional and performance improvements for data warehousing, and SQL Server 2008 R2 includes all these and more. This paper discusses how to use SQL Server 2008 R2 to get great performance as your data warehouse scales up. We present lessons learned during extensive internal data warehouse testing on a 64-core HP Integrity Superdome during the development of the SQL Server 2008 release, and via production experience with large-scale SQL Server customers. Our testing indicates that many customers can expect their performance to nearly double on the same hardware they are currently using, merely by upgrading to SQL Server 2008 R2 from SQL Server 2005 or earlier, and compressing their fact tables. We cover techniques to improve manageability and performance at high-scale, encompassing data loading (extract, transform, load), query processing, partitioning, index maintenance, indexed view (aggregate) management, and backup and restore.
2011-05-19
5,175 reads
By Brian Kelley
Core skills depends on the position, but the point is that ensuring you have...
By Steve Jones
zverism – n. the wish that people could suspend their civility and indulge in...
By Kevin3NF
The Source Control Dilemma Every DBA has been there. Trying to keep track of...
Hi I have a task in my SSIS package that moves files from source...
We are migrating our environments to Azure, it will be a mix of SQL...
Comments posted to this topic are about the item The Inefficiencies of Kubernetes
How do I calculate the amount of space needed for I/O buffers during a backup operation?
See possible answers