• 0 Posts
  • 62 Comments
Joined 2 months ago
cake
Cake day: July 28th, 2025

help-circle

  • In my city there was a photo radar set up at a construction site that issued 10 tickets to the same vehicle over a period of 2 weeks all before the driver recieved the first ticket in the mail. How does that improve road safety (assuming that actually the intent of photo radar)?

    So, one unique example where it took 2 weeks to have an effect is enough to say that they don’t work? While there are studies available through a simple word search that show the opposite? I’m confused here, but perhaps I misunderstood that 2nd paragraph.










  • I don’t know, real world data maybe? Your one, or 2, or even 10 experiences are very insignificant statistically speaking. And of course it’s not a rare story, people who talk online about a product are most usually people with a bad experience, complaining about it, it kinda introduces a bias that you have to ignore. So you go for things like failure rates, which you can find online.

    By the way, it’s almost never actually a fault from AMD or Nvidia, but the actual manufacturer of the card.

    Edit: Not that I care about Internet points, but downvoting without a rebuttal is… Not very convincing







  • I don’t believe in the “whoops we made a mistake, we don’t know what the average Canadian actually wants”, their business is in knowing what we want and how to sell it to us. That checklist you made as an example? Wouldn’t let them control exactly what they want to sell.

    I know that when something can be attributed for both malice and incompetence we should most often choose the latter, but I believe it less and less when it comes to corporations (as opposed to fallible people). Also some products which have nothing at all to do with Canada have been labeled with that maple leaf. It’s not a question of “which part of this do they want to be Canadian again?” when nothing about the product is.

    In the end, choosing to not fine the corpos is simply a message telling them to continue with the misleading labels as it allows them to better control what they want to sell without any fear of repercussions.



  • I agree with the part about unintended use, yes an LLM is not and should never act as a therapist. However, concerning your example with search engines, they will catch the suicide keyword and put help sources before any search result. Google does it, DDG also. I believe ChatGPT will start with such resources also on the first mention, but as OpenAI themselves say, the safety features degrade with the length of the conversation.

    About this specific case, I need to find out more, but other comments on this thread say that not only the kid was in therapy, suggesting that the parents were not passive about it, but also that ChatGPT actually encouraged the kid to hide what he was going through. Considering what I was able to hide from my parents when I was a teenager, without such a tool available, I can only imagine how much harder it would be to notice the depth of what this kid was going through.

    In the end I strongly believe that the company should put much stronger safety features, and if they are unable to do so correctly, then my belief is that the product should just not be available to the public. People will misuse tools, especially a tool touted as AI when it is actually a glorified autocomplete.

    (Yes, I know that AI is a much larger term that also encompasses LLMs, but the actual limitations of LLMs are not well enough known by the public, and not communicated enough by the companies to the end users)