Activity
Mon
Wed
Fri
Sun
Oct
Nov
Dec
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
What is this?
Less
More

Memberships

ADHSSpektrum ADHS Tribe

310 members • $15/m

Data Innovators Exchange

505 members • Free

Vegan leben 🐾💚

248 members • Free

BREATHWORK LERNEN

371 members • Free

Bewusste Beziehungen leben

310 members • $9/m

14 contributions to Data Innovators Exchange
Sh** in - Sh** out?
You can have the best data model and the best business logics in place, but if garbage data comes in, the output is also often useless, especially for unexpected garbage. What are your strategies to deal with that?
0 likes • Apr 19
I assume our near-future strategy is letting AI do a lot of work. Checking more contextual issues than technical ones. LinkedIn is crowded by "AI Agents" (to be honest, most of them are not agents, just simple processes, but that's another discussion), so dealing with real AI Agents, using a RAG approach with further domain-specific knowledge in the form of documents, wikis, structured data, ... Alternative ideas: - Implement an electrical shocker for business people entering the garbage. - Place at least 4 confirmation buttons after each single input field. - Blame the "garbage of the month" team. 🤪
Data Vault is dead!
I hear more and more people asking why you should do all the overhead of creating a Data Vault on modern data platforms. They often argue, that when having a persistent data lake and a modern data platform, which allows virtualizing everything on top of the lake, you don't need a Data Vault. What's your take on this?
1 like • Apr 19
Rephrase "Data Vault is dead!" to: "Materializing a Raw Vault dies but lives now in the Data Catalog and automated querying/processing!"
Agile with lnk tables
Hoping I can get some feedback on an issue we are really struggling with. We are trying to work through adding to our DataVault by adding small pieces at a time which works great with hubs and sats, but breaks down when you get to links. We have some idea of the full build but need to work on projects without having the entire picture. How do we handle adding new keys to an existing Link? Is it better just to add a new object? But then what about all the relationship history that was collected?
1 like • Nov '24
Hi from my side, too. We had that discussion in my CDV2P Bootcamp today. "Adding new keys" is changing granularity. If it's due to schema evolution or absorbing new sources, different keys or different amounts of keys -> separate links. Sometimes, having the same link elements, there may be a reason, too, to separate: Different driving keys in different source tables with identical keys is also a reason to separate. I hope that helps and supplements @Marc Winkelmann 's statement.
Is hashing good enough for anonymising data?
https://www.rnz.co.nz/news/business/527419/inland-revenue-giving-thousands-of-taxpayers-details-to-social-media-platforms-for-ad-campaigns?fbclid=IwY2xjawFLSwJleHRuA2FlbQIxMAABHfiQoZd2lKNuLPKWDo5IrGSrtYtTKNwWBrS0kfJBtccVTTWP9FPKrjY3zg_aem_jp9YxiznNYVeVO5Oo9wqfA
Is hashing good enough for anonymising data?
6 likes • Sep '24
If the hashing Algorithm is known, there is nearly no protection. Day of birth is about 100 (yrs) x 365 (days) values possible. Generating a 365k rows lookup table is a no-cost no-brainer. Same for city, zip code, social security number and more PII attributes. Or am I wrong with these assumptions?
What would you do if Data Vault would be forbidden?
Imagine a dystopian world, where you are not allowed to use Data Vault. How would you solve all the problems in your Data Warehouse?
4 likes • Sep '24
I‘d take the red pill and start implementing DV in the underground.
1-10 of 14
Volker Nürnberg
3
34points to level up
@volker-nurnberg-3425
CDV2P Trainer Data Vault Enthusiast ❤️🧑‍🧑‍🧒‍🧒 people (evolve people to power & agility) ❤️#️⃣ data (integrate data & business)

Active 15d ago
Joined Aug 17, 2024
ENFP
Mülheim an der Ruhr
Powered by