Close Menu
    Trending
    • Why PDF Extraction Still Feels LikeHack
    • GenAI Will Fuel People’s Jobs, Not Replace Them. Here’s Why
    • Millions of websites to get ‘game-changing’ AI bot blocker
    • I Worked Through Labor, My Wedding and Burnout — For What?
    • Cloudflare will now block AI bots from crawling its clients’ websites by default
    • 🚗 Predicting Car Purchase Amounts with Neural Networks in Keras (with Code & Dataset) | by Smruti Ranjan Nayak | Jul, 2025
    • Futurwise: Unlock 25% Off Futurwise Today
    • 3D Printer Breaks Kickstarter Record, Raises Over $46M
    AIBS News
    • Home
    • Artificial Intelligence
    • Machine Learning
    • AI Technology
    • Data Science
    • More
      • Technology
      • Business
    AIBS News
    Home»Artificial Intelligence»A Little More Conversation, A Little Less Action — A Case Against Premature Data Integration
    Artificial Intelligence

    A Little More Conversation, A Little Less Action — A Case Against Premature Data Integration

    Team_AIBS NewsBy Team_AIBS NewsMarch 29, 2025No Comments15 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    I discuss to [large] organisations that haven’t but correctly began with Knowledge Science (DS) and Machine Learning (ML), they typically inform me that they should run a knowledge integration undertaking first, as a result of “…all the information is scattered throughout the organisation, hidden in silos and packed away at odd codecs on obscure servers run by completely different departments.”

    Whereas it could be true that the information is tough to get at, operating a big information integration undertaking earlier than embarking on the ML half is well a foul concept. This, since you combine information with out figuring out its use — the probabilities that the information goes to be match for objective in some future ML use case is slim, at greatest.

    On this article, I focus on among the most vital drivers and pitfalls for this sort of integration tasks, and moderately recommend an strategy that focuses on optimising worth for cash within the integration efforts. The quick reply to the problem is [spoiler alert…] to combine information on a use-case-per-use-case foundation, working backwards from the use case to determine precisely the information you want.

    A need for clear and tidy information

    It’s simple to know the urge for doing information integration previous to beginning on the information science and machine studying challenges. Beneath, I checklist 4 drivers that I typically meet. The checklist isn’t exhaustive, however covers a very powerful motivations, as I see it. We’ll then undergo every driver, discussing their deserves, pitfalls and alternate options.

    1. Cracking out AI/ML use instances is tough, and much more so if you happen to don’t know what information is on the market, and of which high quality.
    2. Snooping out hidden-away information and integrating the information right into a platform looks as if a extra concrete and manageable downside to resolve.
    3. Many organisations have a tradition for not sharing information, and specializing in information sharing and integration first, helps to vary this.
    4. From historical past, we all know that many ML tasks grind to a halt resulting from information entry points, and tackling the organisational, political and technical challenges previous to the ML undertaking might assist take away these obstacles.

    There are in fact different drivers for information integration tasks, reminiscent of “single supply of fact”, “Buyer 360”, FOMO, and the fundamental urge to “do one thing now!”. Whereas vital drivers for information integration initiatives, I don’t see them as key for ML-projects, and subsequently is not going to focus on these any additional on this put up.

    1. Cracking out AI/ML use instances is tough,

    … and much more so if you happen to don’t know what information is on the market, and of which high quality. That is, actually, an actual Catch-22 downside: you may’t do machine studying with out the best information in place, however if you happen to don’t know what information you’ve got, figuring out the potentials of machine studying is basically inconceivable too. Certainly, it is likely one of the essential challenges in getting began with machine studying within the first place [See “Nobody puts AI in a corner!” for more on that]. However the issue isn’t solved most successfully by operating an preliminary information discovery and integration undertaking. It’s higher solved by an superior methodology, that’s nicely confirmed in use, and applies to so many alternative downside areas. It’s referred to as speaking collectively. Since this, to a big extent, is the reply to a number of of the driving urges, we will spend a number of strains on this subject now.

    The worth of getting folks speaking to one another can’t be overestimated. That is the one method to make a staff work, and to make groups throughout an organisation work collectively. It’s also a really environment friendly service of details about intricate particulars concerning information, merchandise, companies or different contraptions which can be made by one staff, however for use by another person. Evaluate “Speaking Collectively” to its antithesis on this context: Produce Complete Documentation. Producing self-contained documentation is tough and costly. For a dataset to be usable by a 3rd occasion solely by consulting the documentation, it needs to be full. It should doc the total context during which the information have to be seen; How was the information captured? What’s the producing course of? What transformation has been utilized to the information in its present type? What’s the interpretation of the completely different fields/columns, and the way do they relate? What are the information varieties and worth ranges, and the way ought to one cope with null values? Are there entry restrictions or utilization restrictions on the information? Privateness issues? The checklist goes on and on. And because the dataset modifications, the documentation should change too.

    Now, if the information is an unbiased, business information product that you just present to clients, complete documentation stands out as the method to go. If you’re OpenWeatherMap, you need your climate information APIs to be nicely documented — these are true information merchandise, and OpenWeatherMap has constructed a enterprise out of serving real-time and historic climate information by means of these APIs. Additionally, if you’re a big organisation and a staff finds that it spends a lot time speaking to those who it might certainly repay making complete documentation — then you definately do this. However most inside information merchandise have one or two inside shoppers to start with, after which, complete documentation doesn’t repay.

    On a basic observe, Speaking Collectively is definitely a key issue for succeeding with a transition to AI and Machine Studying altogether, as I write about in “Nobody puts AI in a corner!”. And, it’s a cornerstone of agile software program improvement. Keep in mind the Agile Manifesto? We worth people and interplay over complete documentation, it states. So there you’ve got it. Discuss Collectively.

    Additionally, not solely does documentation incur a price, however you’re operating the danger of accelerating the barrier for folks speaking collectively (“learn the $#@!!?% documentation”).

    Now, simply to be clear on one factor: I’m not towards documentation. Documentation is tremendous vital. However, as we focus on within the subsequent part, don’t waste time on writing documentation that’s not wanted.

    2. Snooping out hidden away information and integrating the information right into a platform appears as a way more concrete and manageable downside to clear up.

    Sure, it’s. Nevertheless, the draw back of doing this earlier than figuring out the ML use case, is that you just solely clear up the “integrating information in a platform” downside. You don’t clear up the “collect helpful information for the machine studying use case” downside, which is what you need to do. That is one other flip aspect of the Catch-22 from the earlier part: if you happen to don’t know the ML use case, then you definately don’t know what information it’s worthwhile to combine. Additionally, integrating information for its personal sake, with out the data-users being a part of the staff, requires superb documentation, which now we have already coated.

    To look deeper into why information integration with out the ML-use case in view is untimely, we are able to have a look at how [successful] machine studying tasks are run. At a excessive degree, the output of a machine studying undertaking is a type of oracle (the algorithm) that solutions questions for you. “What product ought to we advocate for this person?”, or “When is that this motor due for upkeep?”. If we follow the latter, the algorithm could be a perform mapping the motor in query to a date, particularly the due date for upkeep. If this service is supplied by means of an API, the enter will be {“motor-id” : 42} and the output will be {“newest upkeep” : “March ninth 2026”}. Now, this prediction is completed by some “system”, so a richer image of the answer could possibly be one thing alongside the strains of

    Picture by the writer.

    The important thing right here is that the motor-id is used to acquire additional details about that motor from the information mesh as a way to do a sturdy prediction. The required information set is illustrated by the function vector within the illustration. And precisely which information you want as a way to do this prediction is tough to know earlier than the ML undertaking is began. Certainly, the very precipice on which each ML undertaking balances, is whether or not the undertaking succeeds in determining precisely what data is required to reply the query nicely. And that is carried out by trial and error in the midst of the ML undertaking (we name it speculation testing and have extraction and experiments and different fancy issues, however it’s simply structured trial and error).

    When you combine your motor information into the platform with out these experiments, how are you going to know what information it’s worthwhile to combine? Certainly, you might combine every little thing, and hold updating the platform with all the information (and documentation) to the tip of time. However almost certainly, solely a small quantity of that information is required to resolve the prediction downside. Unused information is waste. Each the trouble invested in integrating and documenting the information, in addition to the storage and upkeep price forever to return. In response to the Pareto rule, you may count on roughly 20% of the information to offer 80% of the information worth. However it’s laborious to know which 20% that is previous to figuring out the ML use case, and previous to operating the experiments.

    That is additionally a warning towards simply “storing information for the sake of it”. I’ve seen many information hoarding initiatives, the place decrees have been handed from high administration about saving away all the information doable, as a result of information is the brand new oil/gold/money/forex/and so forth. For a concrete instance; a number of years again I met with an previous colleague, a product proprietor within the mechanical business, they usually had began accumulating all types of time sequence information about their equipment a while in the past. Someday, they got here up with a killer ML use case the place they wished to make the most of how distributed occasions throughout the economic plant had been associated. However, alas, after they checked out their time sequence information, they realised that the distributed machine situations didn’t have sufficiently synchronised clocks, resulting in non-correlatable time stamps, so the deliberate cross correlation between time sequence was not possible in any case. Bummer, that one, however a classical instance of what occurs while you don’t know the use case you’re gathering information for.

    3. Many organisations have a tradition for not sharing information, and specializing in information sharing and integration first, helps to vary this tradition.

    The primary a part of this sentence is true; there is no such thing as a doubt that many good initiatives are blocked resulting from cultural points within the organisation. Energy struggles, information possession, reluctance to share, siloing and so forth. The query is whether or not an organisation large information integration effort goes to vary this. If somebody is reluctant to share their information, having a creed from above stating that if you happen to share your information, the world goes to be a greater place might be too summary to vary that angle.

    Nevertheless, if you happen to work together with this group, embrace them within the work and present them how their information will help the organisation enhance, you’re more likely to win their hearts. As a result of attitudes are about emotions, and one of the best ways to cope with variations of this type is (consider it or not) to discuss collectively. The staff offering the information has a have to shine, too. And if they don’t seem to be being invited into the undertaking, they’ll really feel forgotten and ignored when honour and glory rains on the ML/product staff that delivered some new and fancy resolution to an extended standing downside.

    Keep in mind that the information feeding into the ML algorithms is part of the product stack — if you happen to don’t embrace the data-owning staff within the improvement, you aren’t operating full stack. (An vital cause why full stack groups are higher than many alternate options, is that inside groups, individuals are speaking collectively. And bringing all of the gamers within the worth chain into the [full stack] staff will get them speaking collectively.)

    I’ve been in quite a few organisations, and lots of instances have I run into collaboration issues resulting from cultural variations of this type. By no means have I seen such obstacles drop resulting from a decree from the C-suit degree. Center administration might purchase into it, however the rank-and-file workers principally simply give it a scornful look and keep it up as earlier than. Nevertheless, I’ve been in lots of groups the place we solved this downside by inviting the opposite occasion into the fold, and speaking about it, collectively.

    4. From historical past, we all know that many DS/ML tasks grind to a halt resulting from information entry points, and tackling the organisational, political and technical challenges previous to the ML undertaking might assist take away these obstacles.

    Whereas the paragraph on cultural change is about human behaviour, I place this one within the class of technical states of affairs. When information is built-in into the platform, it needs to be safely saved and straightforward to acquire and use in the best means. For a big organisation, having a technique and insurance policies for information integration is vital. However there’s a distinction between rigging an infrastructure for information integration along with a minimal of processes round this infrastructure, to that of scavenging by means of the enterprise and integrating a shit load of knowledge. Sure, you want the platform and the insurance policies, however you don’t combine information earlier than that you just want it. And, while you do that step-by-step, you may profit from iterative improvement of the information platform too.

    A fundamental platform infrastructure must also include the required insurance policies to make sure compliance to rules, privateness and different issues. Issues that include being an organisation that makes use of machine studying and synthetic intelligence to make choices, that trains on information which will or is probably not generated by people which will or might not have given their consent to completely different makes use of of that information.

    However to circle again to the primary driver, about not figuring out what information the ML tasks might get their palms on — you continue to want one thing to assist folks navigate the information residing in varied components of the organisation. And if we aren’t to run an integration undertaking first, what will we do? Set up a catalogue the place departments and groups are rewarded for including a block of textual content about what sorts of knowledge they’re sitting on. Only a temporary description of the information; what sort of information, what it’s about, who’re stewards of the information, and maybe with a guess to what it may be used for. Put this right into a textual content database or related construction, and make it searchable . Or, even higher, let the database again an AI-assistant that means that you can do correct semantic searches by means of the descriptions of the datasets. As time (and tasks) passes by, {the catalogue} will be prolonged with additional data and documentation as information is built-in into the platform and documentation is created. And if somebody queries a division concerning their dataset, chances are you’ll simply as nicely shove each the query and the reply into {the catalogue} database too.

    Such a database, containing principally free textual content, is a less expensive various to a readily built-in information platform with complete documentation. You simply want the completely different data-owning groups and departments to dump a few of their documentation into the database. They might even use generative AI to supply the documentation (permitting them to test off that OKR too 🙉🙈🙊).

    5. Summing up

    To sum up, within the context of ML-projects, the information integration efforts needs to be attacked by:

    1. Set up a knowledge platform/information mesh technique, along with the minimally required infrastructure and insurance policies.
    2. Create a listing of dataset descriptions that may be queried by utilizing free textual content search, as a low-cost information discovery software. Incentivise the completely different teams to populate the database by means of use of KPIs or different mechanisms.
    3. Combine information into the platform or mesh on a use case per use case foundation, working backwards from the use case and ML experiments, ensuring the built-in information is each needed and ample for its supposed use.
    4. Remedy cultural, cross departmental (or silo) obstacles by together with the related assets into the ML undertaking’s full stack staff, and…
    5. Discuss Collectively

    Good luck!

    Regards
    -daniel-



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleAI-Powered ECG Interpreter. Hello everyone, I’m excited to share my… | by Vishnusairachakatla | Mar, 2025
    Next Article Would You Try a ‘Severance’ Procedure for a $500K Salary?
    Team_AIBS News
    • Website

    Related Posts

    Artificial Intelligence

    Become a Better Data Scientist with These Prompt Engineering Tips and Tricks

    July 1, 2025
    Artificial Intelligence

    Lessons Learned After 6.5 Years Of Machine Learning

    July 1, 2025
    Artificial Intelligence

    Prescriptive Modeling Makes Causal Bets – Whether You Know it or Not!

    June 30, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Why PDF Extraction Still Feels LikeHack

    July 1, 2025

    I Tried Buying a Car Through Amazon: Here Are the Pros, Cons

    December 10, 2024

    Amazon and eBay to pay ‘fair share’ for e-waste recycling

    December 10, 2024

    Artificial Intelligence Concerns & Predictions For 2025

    December 10, 2024

    Barbara Corcoran: Entrepreneurs Must ‘Embrace Change’

    December 10, 2024
    Categories
    • AI Technology
    • Artificial Intelligence
    • Business
    • Data Science
    • Machine Learning
    • Technology
    Most Popular

    Chef Douglas Keene Is 86ing Toxic Kitchens Like in The Bear

    March 2, 2025

    Maine’s Floating Offshore Wind Setback: What’s Next?

    May 29, 2025

    Tesla Fires a Manager Who Criticized Elon Musk on Social Media

    February 28, 2025
    Our Picks

    Why PDF Extraction Still Feels LikeHack

    July 1, 2025

    GenAI Will Fuel People’s Jobs, Not Replace Them. Here’s Why

    July 1, 2025

    Millions of websites to get ‘game-changing’ AI bot blocker

    July 1, 2025
    Categories
    • AI Technology
    • Artificial Intelligence
    • Business
    • Data Science
    • Machine Learning
    • Technology
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2024 Aibsnews.comAll Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.