❗❗❗❗❗❗❗❗❗❗❗URGENT❗❗❗❗PLeSE READ ASAP❗MY REQESTS ARE MORE IMPORATNT THAN YOUR TIME❗❗ CC: yourboss,your mum,your uni prof
Behind every ❗️❗️❗️🚨🚨🚨URGENT🚨🚨🚨❗️❗️❗️ there is a person who’s about to miss a deadline and, instead of working on themselves to prevent that from happening in the future, makes it the developer’s deadline to miss
Also that urgency is rooted in job insecurity, not even customer impact. They just don’t want to look bad.
E: which I mean fair enough, me too, but still.
But if you do your job properly, you don’t end up in this situation.
Yeah but we’re all learning, so a certain amount of grace is called for.
Anyway, not to counter my own point. There’s a line, is what I’m saying, and it’s blurry.
deleted by creator
Dump all your data in a data lake, throw a thousand GPUs in there too, and you’ll have your answer in 2 hours max.
Thanks, I’ll tell the client it’ll be ready in 2h.
Mr. Manager calling from a group
Hi everyone, I know we’re all busy, but I just wanted to align on this 2 hr estimate. Can we put our heads together and do this faster somehow?
Two thousand GPUs
And the data they want is the entire FY, is 3,000,000 records and they need every single data attribute making the file like 250 MBs. Then you put it in their SharePoint and they get mad they can’t just view it in the browser despite the giant “This file is too large to view online, download it” message.
“Just email it to me!”
Newspaper: Hackers are announcing a trove of personal data leaked from [company] after a forwarded spreadsheet inadvertently contained more data than the sender realised.
Hey! I just started looking at SQL and this is the first SQL joke I’ve ever seen or at least ever gotten!
So, congratulations me!
Same feel as “how long is this going to take to pull?” Well I don’t know if part of what you’re asking for exists, how clean it is, and if can join the data you’re talking about, so anywhere from 5 minutes to never?
That’s exactly how you should respond. I’ve been on the requester for some of these and if my team gave me that as a response I’d just say “let me know what you find out or when you know more.”
How many widgets have we transferred to acme this year?
Simple enough question right?
But then when you look at the data, each region works with acme’s local offices differently. Some transfer using one method, some offices mark the transfer in the system as “other firm”. Oh, and we don’t even get a data feed from the north west region because they still haven’t upgraded their shit so I can request a spreadsheet but it’s in a different format than everything else.
Then inevitably Acme has a different number of widgets that have been transfered. Because if a transfer gets kicked back or cancelled, it’s easier to just create a new transfer rather than go fix an old one because that process is laborious and requires tons of approvals so they just create a new transfer and send it over.
But yea, 20 minutes should be enough time to get you that before your meeting with Acme.
I hate these requests so fucking much. I’ve learned a lot of SQL because of it but I’m sick of it. Especially sick of the users who ask for the same data over and over again.
One guy asked me to run a report every first of the month and then he wouldn’t respond when I would send it so I stopped sending it. Additionally because he would request it AGAIN later in the month after I already sent it at the beginning of the month.
Guess it’s too much to search your fucking emails before requesting a new report to be run. A report that I’ve told you countless times will slow down everything for everyone else who’s using the system.
But tHis iS uRgENt aSAp to run a report asking for all data for the last 3 years.
Especially sick of the users who ask for the same data over and over again.
Use something like Apache Airflow to automate it :)
If it’s regular, I recommend
cron
+mailutils
. Have the cron job call a script with a variable sleep in it if you want to make it look more manual.
Man I don’t regret leaving this behind at my last job. You start out by doing someone a one-off like “sure I can pull the top 5 promotional GICs broken down by region for your blog article - I love supporting my co-workers!”
Then requests become increasingly esoteric and arcane, and insistent.
You try to build a simple FE to expose the data for them, but you can’t get the time approved so you either have to do it with OT or good ol’ time theft, and even then there’s no replacement for just writing SQL, so you’ll always be their silver bullet.
At that point you teach them how to do it themselves. Isn’t there a way to give them an account that only has read access so they can’t inadvertently screw up the database?
I like that idea, and it actually did work for our Marketing guy (Salesforce has a kind of SQL). Near the end there, I just had to debug a few of his harder errors, or double check a script that was going to be running on production.
Never thought of it for Postres or Mysql, etc, but I suppose there’s got to be an easy enough way to get someone access
phpmysqladmin 😆
In Oracle you’d just set up a user that has limited access and give them those credentials. Creating a few views that pulls in the data they want is a bonus.
At work, I am currently dealing with a table that has no primary key, no foreign key, duplicate (almost) serial numbers, booleans stored as strings, and so on. It’s a nightmare of a table.
Entity framework is acting like I’m on meth for using such a table.
I’ve been there and you know what’s worse about it? When you fix it only you or a handful of people notice the astronomical labor you did.
“It worked before why did you change it? You are just doing busywork”
Yeah. Luckily the work I am doing is to fix some really bad work that the entire company has been complaining about. So once it’s fixed it will hopefully be a little bit more recognition than that. Plus my boss is pretty level headed.
But who fucking knows? There is always the likelihood that people will say things along those lines. And it ain’t my job to fight them on that.
How about a date stored as an integer?
Edit: and I’m not taking about a timestamp
No, we have worse. Dates sometimes stored as strings, sometimes as datetimes, and sometimes as integers. There is no consistency, logic, or forethought to the schema.
It’s rough.
Worked on a enterprise medical database, had thousands of tables, and some of the most corrupt data possible. This triggers me :(
Me this morning. I’m gonna take a look at why this Jenkins pipeline is failing. This one job starts a dozen others. Half are failing. For different reasons. After starting rewriting a job that someone half assed. Realize the original error was caused by missing input but some are still valid. Still can’t figure out why my rewritten program is erroring. Get pulled away because another program did something weird… I completed nothing today but worked a ton.
My day…
You have perfectly described my day almost everyday.
Only 3h? What kind sql magician are you?!
It’s OKAY to say no.
I’ve gotta get better at this…
Somebody tell this dude about views.
If they existed for tons of random usecases. When was the last time you created views for “just in case someone asks” situations?
So my work is archaic and doesn’t even use SQL. What are views?
Basically scripts you can run on the fly to pull calculated data. You can (mostly) treat them like tables themselves if you create them on the server.
So if you have repeat requests, you can save the view with maybe some broader parameters and then just SELECT * FROM [View_Schema].[My_View] WHERE [Year] = 2023 or whatever.
It can really slow things down if your views start calling other views in since they’re not actually tables. If you’ve got a view that you find you want to be calling in a lot of other views, you can try to extract as much of it as you can that isn’t updated live into a calculated table that’s updated by a stored procedure. Then set the stored procedure to run at a frequency that best captures the changes (usually daily). It can make a huge difference in runtime at the cost of storage space.
It can really slow things down if your views start calling other views in since they’re not actually tables
They can be in some cases! There’s a type of view called an “indexed” or “materialized” view where the view data is stored on disk like a regular table. It’s automatically recomputed whenever the source tables change. Doesn’t work well for tables that are very frequently updated, though.
Having said that, if you’re doing a lot of data aggregation (especially if it’s a sproc that runs daily), you’d probably want to set up a separate OLAP database so that large analytical queries don’t slow down transactional queries. With open-source technologies, this is usually using Hive and Presto or Spark combined with Apache Airflow.
Also, if you have data that’s usually aggregated by column, then a column-based database like Clickhouse is usually way faster than a regular row-based database. These store data per-column rather than per-row, so aggregating one column across millions or even billions of rows (eg average page load time for all hits ever recorded) is fast.
A view is a saved query that pretends it’s a table. It doesn’t actually store any data. So if you need to query 10 different tables, joining them together and filtering the results specific ways, a view would just be that saved query, so instead of “SELECT * FROM (a big mess of tables)” you can do “SELECT * FROM HandyView”
Predefined queries that you can interact with like another table more or less
deleted by creator
Well if organizing things in pristine tables makes things quick, why can’t we have nice things?
Because Jen in accounting doesn’t believe in it, and Tom the CIO likes his data stored raw in TXT Amphibious Delineated. Then our biggest client prefers data as Jason so we swapped half of our database to that to speed things up.
But the real problem is high turnover because we don’t pay anyone enough to work on things they are proud of. After 2 years we stop doing even 3% COL raises so they go elsewhere. So every 2-4 years each position gets a new opinionated asshole.
our biggest client prefers data as Jason so we swapped half of our database to that
the app I work with currently stores json as the only column in a sql table and it hurts me so very much. like watching someone pick up a screwdriver and try to bash a nail in with the handle.