Gaywallet (they/it)

I’m gay

  • 21 Posts
  • 35 Comments
Joined 3 years ago
cake
Cake day: January 28th, 2022

help-circle

  • Any information humanity has ever preserved in any format is worthless

    It’s like this person only just discovered science, lol. Has this person never realized that bias is a thing? There’s a reason we learn to cite our sources, because people need the context of what bias is being shown. Entire civilizations have been erased by people who conquered them, do you really think they didn’t re-write the history of who these people are? Has this person never followed scientific advancement, where people test and validate that results can be reproduced?

    Humans are absolutely gonna human. The author is right to realize that a single source holds a lot less factual accuracy than many sources, but it’s catastrophizing to call it worthless and it ignores how additional information can add to or detract from a particular claim- so long as we examine the biases present in the creation of said information resources.


  • This isn’t just about GPT, of note in the article, one example:

    The AI assistant conducted a Breast Imaging Reporting and Data System (BI-RADS) assessment on each scan. Researchers knew beforehand which mammograms had cancer but set up the AI to provide an incorrect answer for a subset of the scans. When the AI provided an incorrect result, researchers found inexperienced and moderately experienced radiologists dropped their cancer-detecting accuracy from around 80% to about 22%. Very experienced radiologists’ accuracy dropped from nearly 80% to 45%.

    In this case, researchers manually spoiled the results of a non-generative AI designed to highlight areas of interest. Being presented with incorrect information reduced the accuracy of the radiologist. This kind of bias/issue is important to highlight and is of critical importance when we talk about when and how to ethically introduce any form of computerized assistance in healthcare.






  • It’s FUCKING OBVIOUS

    What is obvious to you is not always obvious to others. There are already countless examples of AI being used to do things like sort through applicants for jobs, who gets audited for child protective services, and who can get a visa for a country.

    But it’s also more insidious than that, because the far reaching implications of this bias often cannot be predicted. For example, excluding all gender data from training ended up making sexism worse in this real world example of financial lending assisted by AI and the same was true for apple’s credit card and we even have full-blown articles showing how the removal of data can actually reinforce bias indicating that it’s not just what material is used to train the model but what data is not used or explicitly removed.

    This is so much more complicated than “this is obvious” and there’s a lot of signs pointing towards the need for regulation around AI and ML models being used in places it really matters, such as decision making, until we understand it a lot better.








  • This really does not surprise me one bit. But also, nobody using these tools really cares. It reduces the amount of applications they need to review, which is often all they care about. Can’t wait for the inevitable company to pop up which will do the AI equivalent of SEO stacking your resume so you can get a job.

    Also, perhaps more importantly, this is just going to undo fifty years of antiracism and antisexism work. The biggest problem with AI is that it’s trained on a bigoted system and when it’s used to gatekeep said system, it just creates additional inequality.









  • So I’m not familiar with why that particular judge was criticized, but you’ll note that the entirety of arguments and discussion surround the brief and response letters which were submitted to the supreme court. It’s possible that Cannon was stepping outside what was presented, which would warrant criticism. The supreme court never steps outside what is presented, except when it’s necessary to understand what is being presented - for example they may call upon other legal text or rulings in order to fully frame what is and is not in scope with regards to the presented case or to understand precisely what a particular lawyer is arguing for or against.

    But I also think that the fact that the supreme court is the final court which gets to have say on a matter lends them to pontificate in depth about some issues that other courts may not be given latitude to do the same because it may address issues which are currently working their way through the court system or may be called upon as a matter of jurisprudence in the future. The entire text of every decision they make can have consequences on lower courts and understanding the limits and the spirit of law they weigh in on and as of such it’s important to fully understand the exact claims being made and appropriately scope where the response lies and whether certain issues could or should be weighed in upon during that case.