• 0 Posts
  • 27 Comments
Joined 1 year ago
cake
Cake day: June 5th, 2023

help-circle
  • The device wouldn’t necessarily have to be constantly streaming the audio to a central server. If it’s capable of hearing wake up words like “Ok Google” it’s capable of listening for other phrases and having onboard processing to relay back the results much more compressed. Whether or not this is common practice is another matter, and yes the algorithms are scary good even without eavesdropping.






  • That’s fair. I think fundamentally a false positive/negative isn’t that much different. Pretty much all tests—especially those dealing with real world conditions—are heuristic, as are all LLMs by necessity of the design. Hallucination is a pretty specific term given to AI as an attempt to assign agency to a system that doesn’t actually have any (by implying it’s crazy and making stuff up instead of a black box with deterministic inputs and outputs spitting out something factually wrong but with a similar format to what is trained on). I feel like the nature of any tool where “you can’t trust this to be entirely accurate” should have an umbrella term that encompasses both types of providing inaccurate info under certain conditions.

    I suppose the difference is that AI is a lot more likely to randomly go off, whereas a blood test is likelier to provide repeated false positives for the same person with their unique biology? There’s also the fact that most medical tests represent a true/false dichotomy or lookup table, whereas an LLM is given the entire bounds of language.

    Would an AI clustering algorithm (say, K-means for instance) giving an inaccurate diagnosis be a false positive/negative or a hallucination? These models can be programmed on a sliding scale and I feel like there’s definitely an area where the line could get pretty blurry.


  • I mean, AI is used in fraud detection pretty often; when it hits a false positive (which happens frequently on a population-level basis), is that not a hallucination of some sort? Obviously LLMs can go off the rails much further because it’s readable text, but any machine learning model will occasionally spit out really bad guesses almost any person could have done better with. (To be fair, humans are highly capable of really bad guesses too).









  • One of our cats has a problem with marking personal items left on the floor, even though she’s fixed. We think it’s partially because she has bad eyesight and so marks stuff so she always knows where it is. It’s usually, but not exclusively, something I or my SO wear or have worn recently:

    ✔️ My purse
    ✔️ My jacket. Probably three times
    ✔️ SO’s backpack
    ✔️ My backpack that was sitting in the closet unused for several months
    ✔️ Articles of clothing that didn’t make it into a laundry basket
    ✔️ The top of the cat tower when we moved it into the living room (trying to assert dominance I guess)
    ✔️ Rugs (I dunno under what circumstances, she hasn’t marked their replacements. She will re-mark previously marked and washed rugs though)
    ✔️ Towels that don’t get put up

    The solution has more or less been to stop keeping anything important on the floor.

    The other cat likes to occasionally pull towels onto the floor so he can use them as temporary bedding or snuggle up against them. You can probably guess what sometimes happens after he leaves them there.