• 0 Posts
  • 44 Comments
Joined 1 year ago
cake
Cake day: June 12th, 2023

help-circle






  • Yea, this bubble is mostly LLM, but also deepfakes and other generative image algorithms. They are all ML. LLM has some fame because people can’t seem to realise that it’s crap. They definitely passed the Turing test, while still being pretty much useless.

    There are many other useless ML algorithms. Just because you don’t like something doesn’t mean it doesn’t belong. ML has some good stuff and some bad stuff. The statement “ML works” doesn’t mean anything. It’s like saying “math works”.

    There have been many AI bubbles in the past as well, as well as slumps. Look up the term AI winter. Most AI algorithms turn out not really working except for a few niche applications. You are probably referring to these few as “ML works”. Most AI projects fail, but some prevail. This goes for all tech though. So… tech works.

    What Microsoft is doing is they are trying to cast a wide net to see if they hit one of the few actual good applications for LLMs. Most of them will fail but there might be one or two really successful products. Good for them to have that kind of capital to just haphazardly try new features everywhere.


  • Lifter@discuss.tchncs.detoFuck AI@lemmy.worldIt isn't worth it
    link
    fedilink
    arrow-up
    1
    arrow-down
    1
    ·
    5 months ago

    Companies hve always simplified smart things and called it AI. AI is hotter than ever now, not only LLM.

    And again ML is a subset of AI, LLM is a subset of ML. With these definitions, everything is AI. Look up the definition of AI. It’s just a collection of techniques to do “smarter” things with AI. It includes all of the above, e.g. “If this then that” but also more advanced mathematics, like statistical methods and ML. LLM is one of those statistical models.



  • Lifter@discuss.tchncs.detoFuck AI@lemmy.worldIt isn't worth it
    link
    fedilink
    arrow-up
    4
    arrow-down
    5
    ·
    5 months ago

    You seem to be arguing against another stawman. OP didn’t say they only dislike LLM the sub is even “Fuck AI”. And this thread is talking about AI in general.

    Machine Learning is a subset of AI and has always been. Also, LLM is a subset of Machine Learning. You are trying to split hairs, or at least do a “That’s not a real Scotsman” out of the above post.










  • No thanks. It’s way more fun to be part of the decision process. If a manager can anticipate all of the requirements and quirks of the project before it even starts, it’s probably going to be a really boring, vanilla project at which point it’s probably just better to but the software.ä somewhere else.

    Creating something new is an art in itself. Why would you not want to be a part of that?

    Also: Isn’t it cheating to compare the two approaches when one of them is defined as having all the planning “outside” of the project scope? I would bet that the statistics in this report disregard ll those projects that died in the planning phase, leaving only the almost completed, easy project to succeed at a high rate.

    It would be interesting to also compare the time/resources spent before each project died. My hunch is that for failed agile project, less total investment has been made before killing it off, as compared to front loading all of that project planning before the decision is made not to continue.

    Complementary to this, I also think that Agile can have a tendency to keep alive projects that should have failed on the planning stage. “We do things not because they are easy, but we thought they would be easy”. Underestimating happens for all project but for Agile, there should be a higher tendency to keep going because “we’re almost done”, forever.