• Destide@feddit.uk
    link
    fedilink
    English
    arrow-up
    131
    arrow-down
    2
    ·
    1 year ago

    ❗❗❗❗❗❗❗❗❗❗❗URGENT❗❗❗❗PLeSE READ ASAP❗MY REQESTS ARE MORE IMPORATNT THAN YOUR TIME❗❗ CC: yourboss,your mum,your uni prof

    • derfl007@lemmy.wtf
      link
      fedilink
      arrow-up
      44
      ·
      1 year ago

      Behind every ❗️❗️❗️🚨🚨🚨URGENT🚨🚨🚨❗️❗️❗️ there is a person who’s about to miss a deadline and, instead of working on themselves to prevent that from happening in the future, makes it the developer’s deadline to miss

      • pomodoro_longbreak@sh.itjust.works
        link
        fedilink
        arrow-up
        21
        ·
        edit-2
        1 year ago

        Also that urgency is rooted in job insecurity, not even customer impact. They just don’t want to look bad.

        E: which I mean fair enough, me too, but still.

          • pomodoro_longbreak@sh.itjust.works
            link
            fedilink
            arrow-up
            6
            ·
            1 year ago

            Yeah but we’re all learning, so a certain amount of grace is called for.

            Anyway, not to counter my own point. There’s a line, is what I’m saying, and it’s blurry.

  • fosforus@sopuli.xyz
    link
    fedilink
    arrow-up
    79
    ·
    edit-2
    1 year ago

    Dump all your data in a data lake, throw a thousand GPUs in there too, and you’ll have your answer in 2 hours max.

  • Winged_Hussar@lemmy.world
    link
    fedilink
    English
    arrow-up
    70
    ·
    1 year ago

    And the data they want is the entire FY, is 3,000,000 records and they need every single data attribute making the file like 250 MBs. Then you put it in their SharePoint and they get mad they can’t just view it in the browser despite the giant “This file is too large to view online, download it” message.

      • perviouslyiner@lemm.ee
        link
        fedilink
        English
        arrow-up
        26
        ·
        1 year ago

        Newspaper: Hackers are announcing a trove of personal data leaked from [company] after a forwarded spreadsheet inadvertently contained more data than the sender realised.

  • Melatonin@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    46
    ·
    1 year ago

    Hey! I just started looking at SQL and this is the first SQL joke I’ve ever seen or at least ever gotten!

    So, congratulations me!

  • Dentarthurdent@lemmy.world
    link
    fedilink
    arrow-up
    42
    ·
    1 year ago

    Same feel as “how long is this going to take to pull?” Well I don’t know if part of what you’re asking for exists, how clean it is, and if can join the data you’re talking about, so anywhere from 5 minutes to never?

    • cybersandwich@lemmy.world
      link
      fedilink
      arrow-up
      13
      ·
      1 year ago

      That’s exactly how you should respond. I’ve been on the requester for some of these and if my team gave me that as a response I’d just say “let me know what you find out or when you know more.”

  • cybersandwich@lemmy.world
    link
    fedilink
    arrow-up
    36
    ·
    1 year ago

    How many widgets have we transferred to acme this year?

    Simple enough question right?

    But then when you look at the data, each region works with acme’s local offices differently. Some transfer using one method, some offices mark the transfer in the system as “other firm”. Oh, and we don’t even get a data feed from the north west region because they still haven’t upgraded their shit so I can request a spreadsheet but it’s in a different format than everything else.

    Then inevitably Acme has a different number of widgets that have been transfered. Because if a transfer gets kicked back or cancelled, it’s easier to just create a new transfer rather than go fix an old one because that process is laborious and requires tons of approvals so they just create a new transfer and send it over.

    But yea, 20 minutes should be enough time to get you that before your meeting with Acme.

  • NoneYa@lemm.ee
    link
    fedilink
    arrow-up
    33
    ·
    1 year ago

    I hate these requests so fucking much. I’ve learned a lot of SQL because of it but I’m sick of it. Especially sick of the users who ask for the same data over and over again.

    One guy asked me to run a report every first of the month and then he wouldn’t respond when I would send it so I stopped sending it. Additionally because he would request it AGAIN later in the month after I already sent it at the beginning of the month.

    Guess it’s too much to search your fucking emails before requesting a new report to be run. A report that I’ve told you countless times will slow down everything for everyone else who’s using the system.

    But tHis iS uRgENt aSAp to run a report asking for all data for the last 3 years.

    • dan@upvote.au
      link
      fedilink
      arrow-up
      11
      ·
      1 year ago

      Especially sick of the users who ask for the same data over and over again.

      Use something like Apache Airflow to automate it :)

  • pomodoro_longbreak@sh.itjust.works
    link
    fedilink
    arrow-up
    29
    ·
    edit-2
    1 year ago

    Man I don’t regret leaving this behind at my last job. You start out by doing someone a one-off like “sure I can pull the top 5 promotional GICs broken down by region for your blog article - I love supporting my co-workers!”

    Then requests become increasingly esoteric and arcane, and insistent.

    You try to build a simple FE to expose the data for them, but you can’t get the time approved so you either have to do it with OT or good ol’ time theft, and even then there’s no replacement for just writing SQL, so you’ll always be their silver bullet.

    • applebusch@lemmy.world
      link
      fedilink
      arrow-up
      16
      arrow-down
      1
      ·
      1 year ago

      At that point you teach them how to do it themselves. Isn’t there a way to give them an account that only has read access so they can’t inadvertently screw up the database?

      • pomodoro_longbreak@sh.itjust.works
        link
        fedilink
        arrow-up
        8
        ·
        1 year ago

        I like that idea, and it actually did work for our Marketing guy (Salesforce has a kind of SQL). Near the end there, I just had to debug a few of his harder errors, or double check a script that was going to be running on production.

        Never thought of it for Postres or Mysql, etc, but I suppose there’s got to be an easy enough way to get someone access

      • lazyslacker@sh.itjust.works
        link
        fedilink
        arrow-up
        3
        ·
        1 year ago

        In Oracle you’d just set up a user that has limited access and give them those credentials. Creating a few views that pulls in the data they want is a bonus.

  • PizzaMan@lemm.ee
    link
    fedilink
    arrow-up
    29
    ·
    edit-2
    1 year ago

    At work, I am currently dealing with a table that has no primary key, no foreign key, duplicate (almost) serial numbers, booleans stored as strings, and so on. It’s a nightmare of a table.

    Entity framework is acting like I’m on meth for using such a table.

    • (╯°□°)╯︵ ┻━┻@programming.dev
      link
      fedilink
      English
      arrow-up
      19
      ·
      1 year ago

      I’ve been there and you know what’s worse about it? When you fix it only you or a handful of people notice the astronomical labor you did.

      “It worked before why did you change it? You are just doing busywork”

      • PizzaMan@lemm.ee
        link
        fedilink
        arrow-up
        11
        ·
        1 year ago

        Yeah. Luckily the work I am doing is to fix some really bad work that the entire company has been complaining about. So once it’s fixed it will hopefully be a little bit more recognition than that. Plus my boss is pretty level headed.

        But who fucking knows? There is always the likelihood that people will say things along those lines. And it ain’t my job to fight them on that.

      • PizzaMan@lemm.ee
        link
        fedilink
        arrow-up
        11
        ·
        edit-2
        1 year ago

        No, we have worse. Dates sometimes stored as strings, sometimes as datetimes, and sometimes as integers. There is no consistency, logic, or forethought to the schema.

        It’s rough.

  • Crashumbc@lemmy.world
    link
    fedilink
    English
    arrow-up
    27
    ·
    1 year ago

    Worked on a enterprise medical database, had thousands of tables, and some of the most corrupt data possible. This triggers me :(

  • Ensign Rick@startrek.website
    link
    fedilink
    English
    arrow-up
    22
    ·
    1 year ago

    Me this morning. I’m gonna take a look at why this Jenkins pipeline is failing. This one job starts a dozen others. Half are failing. For different reasons. After starting rewriting a job that someone half assed. Realize the original error was caused by missing input but some are still valid. Still can’t figure out why my rewritten program is erroring. Get pulled away because another program did something weird… I completed nothing today but worked a ton.

    My day…

    • Gallardo994@sh.itjust.works
      link
      fedilink
      arrow-up
      27
      ·
      1 year ago

      If they existed for tons of random usecases. When was the last time you created views for “just in case someone asks” situations?

      • doctordevice@lemm.ee
        link
        fedilink
        arrow-up
        22
        ·
        edit-2
        1 year ago

        Basically scripts you can run on the fly to pull calculated data. You can (mostly) treat them like tables themselves if you create them on the server.

        So if you have repeat requests, you can save the view with maybe some broader parameters and then just SELECT * FROM [View_Schema].[My_View] WHERE [Year] = 2023 or whatever.

        It can really slow things down if your views start calling other views in since they’re not actually tables. If you’ve got a view that you find you want to be calling in a lot of other views, you can try to extract as much of it as you can that isn’t updated live into a calculated table that’s updated by a stored procedure. Then set the stored procedure to run at a frequency that best captures the changes (usually daily). It can make a huge difference in runtime at the cost of storage space.

        • dan@upvote.au
          link
          fedilink
          arrow-up
          5
          ·
          edit-2
          1 year ago

          It can really slow things down if your views start calling other views in since they’re not actually tables

          They can be in some cases! There’s a type of view called an “indexed” or “materialized” view where the view data is stored on disk like a regular table. It’s automatically recomputed whenever the source tables change. Doesn’t work well for tables that are very frequently updated, though.

          Having said that, if you’re doing a lot of data aggregation (especially if it’s a sproc that runs daily), you’d probably want to set up a separate OLAP database so that large analytical queries don’t slow down transactional queries. With open-source technologies, this is usually using Hive and Presto or Spark combined with Apache Airflow.

          Also, if you have data that’s usually aggregated by column, then a column-based database like Clickhouse is usually way faster than a regular row-based database. These store data per-column rather than per-row, so aggregating one column across millions or even billions of rows (eg average page load time for all hits ever recorded) is fast.

      • calabast@lemm.ee
        link
        fedilink
        arrow-up
        19
        ·
        edit-2
        1 year ago

        A view is a saved query that pretends it’s a table. It doesn’t actually store any data. So if you need to query 10 different tables, joining them together and filtering the results specific ways, a view would just be that saved query, so instead of “SELECT * FROM (a big mess of tables)” you can do “SELECT * FROM HandyView”

      • Restaldt@lemm.ee
        link
        fedilink
        arrow-up
        10
        ·
        edit-2
        1 year ago

        Predefined queries that you can interact with like another table more or less

  • Mango@lemmy.world
    link
    fedilink
    arrow-up
    10
    ·
    1 year ago

    Well if organizing things in pristine tables makes things quick, why can’t we have nice things?

    • Patches@sh.itjust.works
      link
      fedilink
      arrow-up
      24
      ·
      edit-2
      1 year ago

      Because Jen in accounting doesn’t believe in it, and Tom the CIO likes his data stored raw in TXT Amphibious Delineated. Then our biggest client prefers data as Jason so we swapped half of our database to that to speed things up.

      But the real problem is high turnover because we don’t pay anyone enough to work on things they are proud of. After 2 years we stop doing even 3% COL raises so they go elsewhere. So every 2-4 years each position gets a new opinionated asshole.

      • Alien Nathan Edward@lemm.ee
        link
        fedilink
        arrow-up
        9
        ·
        1 year ago

        our biggest client prefers data as Jason so we swapped half of our database to that

        the app I work with currently stores json as the only column in a sql table and it hurts me so very much. like watching someone pick up a screwdriver and try to bash a nail in with the handle.