Activity
Mon
Wed
Fri
Sun
Dec
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
What is this?
Less
More

Memberships

Data Innovators Exchange

Public • 322 • Free

22 contributions to Data Innovators Exchange
Data Warehouse Automation Challenge mit DSharp 7th of November
Tomorrow it is DSharps turn to tacle the Willibald-challenge organized by the Deutschsprachige Data Vault User Group (DDVUG) and @Michael Müller. Come to see how we did tomorrow Thursday the 7th at 17:00 (CET). https://www.linkedin.com/events/datawarehouseautomationchalleng7256429109080133632/
6
3
New comment 10d ago
1 like • 11d
Hi Konsta. Will be a recording from the webinar?
Data Innovator Community Meetups
To all Data Innovators, Where do you want to see the next Data innovator events? Due to the huge success of the community meetups in London and Munich we want to keep them coming. In these events we got very good feedback on the focus of the workshops, so we are going to keep the focus on them and stay with the 1 day format with: - Success story - 2 Workshops - Panel discussion The team currently is brainstorming locations and would love to hear your thoughts on this 🙂 So please let us know in the comments where you want a Data Innovators Event to take place and we will give our best to make it happen 🙂 Thank you for your input!
6
9
New comment Oct 10
1 like • Oct 7
Zürich
To Hash or not to Hash 🧐
Hashing is a crucial part of Data Vault implementations. They help quickly identifying deltas, by not having to compare every single attribute of a satellite, but instead comparing the hashed value over all of these attributes. This helps to reduce the complexity of queries being written, since significantly fewer columns need to be fully specified. But now imagine a fully automated Raw Data Vault implementation, would you still generate Hashkeys and Hashdiffs? Since you don't write the loading scripts of satellites by yourself, what benefit do hash values bring to the Data Vault implementation? Wouldn't it be nicer to directly have business keys everywhere? You could argue that delta detection might be slower, when all columns need to be compared, but does anyone have experience if this is really the case? On modern databases, I would imagine this delta detection to not have an actual impact on overall performance. What's your opinion on skipping hashes? Let me know!
6
4
New comment Sep 2
To Hash or not to Hash 🧐
2 likes • Sep 1
I had a discussion about hashing with Petr Beles at the last Data Dreamland. He sad, for example, that DV Builder is prepared to calculate hash diff but there is no database, which they support for now, which would be quicker in hash computing than in column by column comparison. So, DV Builder currently doesn't calculate and store hash diff. But they also make a performance test with every new version of target database and they are prepared to switch the hash diff for selected db and versions. Interesting.
What Was Your First Experience Working with Data? Some Fantasy Football experts here?
Hey everyone, I'm curious to know—what was your entry point into the world of data? For me, it all began with Fantasy Football. I wanted to create my own stat analytics, so I started querying player databases using an API. There’s so much to analyze in American Football, and with our fantasy draft kicking off this Sunday, I’m reminded of those early days. So, what was your first data project? How did you get started? Looking forward to hearing your stories!
8
9
New comment Sep 2
3 likes • Sep 1
I must say, I don't remember, what data it was. But I know that it was in summer 1990, my first part-time job - development of an educational program in dBase III to demonstrate basic functions of this database. It was on PC - XT with 10MB HDD and 640 kB of memory. No. Now I remember. The most first data I worked with, were final grades for the whole secondary school, I've studied. I was a second year student, fond of computers and the school received a brand new computing laboratory of PCs. Friend of my class teacher prepared a database and a program in dBase III for final grades recording, mandatory reports and school reports printing. But there was no one who was able to use it, much less maintain it. So, in June 1990, last two weeks of school before holidays, I was at school but not in class. I acted as a team leader of group of ten schoolmates. We sat in the computing laboratory and copied all the grades from paper into database. Then we printed all the school reports. As I remember, it was about 1.200 students. It was crazy time. The first years after "velvet revolution" in former Czechoslovakia were very special.
Snowflake vs. Databricks
Just finished watching a breakdown of Snowflake and Databricks. While the comparison itself isn't new, the focus on their origins and how that shaped their current offerings was insightful. Key takeaways: - Foundational Philosophies: Snowflake's roots in traditional data warehousing vs. Databricks' academic, notebook-centric beginnings are evident in their core strengths even today. - Architectural Choices: Snowflake's virtual warehouses & micro-partitions vs. Databricks' Spark clusters & Delta Lake highlight differing approaches to storage, compute & scalability. - The paths of the two platforms, Snowpark and the data lakehouse concept, reflect their initial goals. Share your thoughts! Which platform's philosophy resonates more with your data workflows?
Poll
7 members have voted
6
6
New comment 5d ago
Snowflake vs. Databricks
5 likes • Sep 1
I have, maybe, too simple point of view. If your data are more than 50% relational - go to Snowflake. Else - go to DataBricks. And don't tell me that you have 100% of data in semi-structured file extracts, although all your primary data sources are relational databases.
1-10 of 22
Richard Sklenařík
4
80points to level up
@richard-sklenarik-8201
DWH architect who built his first DWH in 1998 and went through different methodologies to finally find that there was one that finally work all along.

Active 2d ago
Joined Jul 9, 2024
Prague
powered by