Activity
Mon
Wed
Fri
Sun
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
What is this?
Less
More

Memberships

Data Innovators Exchange

Public • 344 • Free

8 contributions to Data Innovators Exchange
Data Dreamland 2024
Like a few others in the community, I will be attending Scalefree’s Data Dreamland conference in Hannover September 23rd and 24th! I will be presenting on optimizing data architecture through a hybrid Data Fabric and Data Mesh approach. If you haven't yet, get your tickets here: https://lnkd.in/ebqPAXjw 🎙️Combining Data Fabric and Data Mesh: Advancing Data Management with Hybrid Architecture 📅 Tuesday, September 24th 🕦 11:30 AM
7
2
New comment Sep 16
Data Dreamland 2024
1 like • Sep 16
@Tim Kirschke Can't wait😄
Progress over Perfection: Why MVP beats POC
Here's the link to our webinar where Daniel Olsberg, from areto, and I discuss the importance of prioritizing progress over perfection in your data automation journey. Take a look, especially if you're interested in hearing the checkpoints of launching a Minimum Viable Product (MVP) over a Proof of Concept (POC), and learning how this approach can accelerate business value and efficiency. Take care 👋
4
1
New comment Sep 12
Progress over Perfection: Why MVP beats POC
To Hash or not to Hash 🧐
Hashing is a crucial part of Data Vault implementations. They help quickly identifying deltas, by not having to compare every single attribute of a satellite, but instead comparing the hashed value over all of these attributes. This helps to reduce the complexity of queries being written, since significantly fewer columns need to be fully specified. But now imagine a fully automated Raw Data Vault implementation, would you still generate Hashkeys and Hashdiffs? Since you don't write the loading scripts of satellites by yourself, what benefit do hash values bring to the Data Vault implementation? Wouldn't it be nicer to directly have business keys everywhere? You could argue that delta detection might be slower, when all columns need to be compared, but does anyone have experience if this is really the case? On modern databases, I would imagine this delta detection to not have an actual impact on overall performance. What's your opinion on skipping hashes? Let me know!
6
4
New comment Sep 2
To Hash or not to Hash 🧐
4 likes • Aug 29
On some platforms, like Snowflake, for instance, the hashing of BK's has a negative impact on performance. Automation tools need to provide both options.
MS Fabric
Where are you guys with Microsoft Fabric?
Poll
11 members have voted
7
5
New comment Aug 29
2 likes • Aug 28
@Lorenz Kindling, Yes, VaultSpeed released support for MS Fabric, so we are running our integration testing framework on it. This framework basically tests all the data vault patterns with data on a specific target technology. I like the vision of combining the lake(house) experience that runs on Spark with the warehouse experience that runs on T-SQL. Regardless of your background, both experiences can consume each other's work (and later, I heard that they will also be able to alter each other's work). That's a very powerful concept. Its engine also removes the need for tuning, which was more often needed in Synapse. Some capabilities are still lacking, but I'm confident they will improve over time. For example, harvesting metadata about FK's through the JDBC driver proves difficult. https://vaultspeed.com/resources/articles/release-5-7-2-introducing-vaultspeeds-beta-support-for-microsoft-fabric
0 likes • Aug 29
Hi @Richard Sklenařík, great question. We've supported Databricks for over 3 years now. I believe it's a fantastic platform and their vision to bring the warehouse experience to the lake is amazing. I've seen the platform grow in both speed and functionality and with the acquisition of Tabular they plan to increase the interoperability of their lakehouse even more. I've seen some great data vault stories on Databricks and I'm sure I'll see a lot more in the future. It will be very interesting to see how MS Fabric, Databricks, Snowflake will compare in the future.
Is Kimball Modelling Dead? Navigating Modern Data Management
What do you think? Brought to you by Vaultspeed this discussion delves into: Understanding of Kimball and Data Vault Methodologies their purposes, strengths, and weaknesses. How to adapt to Changes in Data Structures, and in particular how Data Vault can efficiently manage changes such as mergers, acquisitions, and expansions which might require rapid integration of new data types into existing models without compromising data integrity or historical data. A demonstration of VaultSpeed to illustrate how automation tools can simplify and accelerate the modelling process, showing how these tools can generate SQL and handle deployments. Practical Implementation and Transition Strategies on how to transition from a Kimball model to a Data Vault model using automation, the discussion provides practical strategies that you can apply directly to their work, especially when looking to enhance the resilience and flexibility of their data architecture. #vaultspeed #datamodelling #kimball #datavault #datawarehouse
4
6
New comment Aug 28
Is Kimball Modelling Dead? Navigating Modern Data Management
2 likes • Aug 28
I see it like this: Data Vault made Kimball a lot easier! So it's not dead, but perhaps the way we implement it has changed. Data Vault covers some functions like loading, historization, storage, ... that are just more difficult to maintain with Kimball. I've built (and maintained) Kimball the "old fashioned" way, and then later on top of data vault: I would never want to go back.
1-8 of 8
Jonas De Keuster
3
19points to level up
@jonas-de-keuster-8790
Jonas De Keuster is VP Product Marketing at VaultSpeed.

Active 55d ago
Joined Jul 11, 2024
powered by