Skip to content
  • Kategorien
  • Aktuell
  • Tags
  • Beliebt
  • World
  • Benutzer
  • Gruppen
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Standard: (Kein Skin)
  • Kein Skin
Einklappen

other.li Forum

  1. Übersicht
  2. Uncategorized
  3. My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing.

My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing.

Geplant Angeheftet Gesperrt Verschoben Uncategorized
79 Beiträge 53 Kommentatoren 1 Aufrufe
  • Älteste zuerst
  • Neuste zuerst
  • Meiste Stimmen
Antworten
  • In einem neuen Thema antworten
Anmelden zum Antworten
Dieses Thema wurde gelöscht. Nur Nutzer mit entsprechenden Rechten können es sehen.
  • ? Gast

    My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

    LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

    In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

    But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

    If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

    ? Offline
    ? Offline
    Gast
    schrieb am zuletzt editiert von
    #14

    @EmilyEnough very interesting observation, thanks a lot. I haven’t perceived it that way - I used to work a lot with probabilistic models, simulated annealing and genetic algorithms and in that case the computer works entirely deterministic but the result is always different. So I lost my expectation to a deterministic result long time ago 😉

    1 Antwort Letzte Antwort
    0
    • ? Gast

      My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

      LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

      In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

      But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

      If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

      ? Offline
      ? Offline
      Gast
      schrieb am zuletzt editiert von
      #15

      @EmilyEnough Very astute and exactly my experience too - I went into computing for the same kinds of reasons and as you say LLMs break that. Thank you for expressing it so clearly.

      1 Antwort Letzte Antwort
      0
      • ? Gast

        My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

        LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

        In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

        But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

        If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

        ? Offline
        ? Offline
        Gast
        schrieb am zuletzt editiert von
        #16

        @EmilyEnough I think you're absolutely correct on this. Yet another reason why we need to find a way to irrevocably destroy this abomination.

        But also it's not just the style of "communication" that these algorithms are pretending to do, it's that you cannot trust that their output is even correct because they have no understanding of what they are "saying". They could be "hallucinating" complete nonsense but they'll output it in an authoritative way and may even make up references that don't exist. They're 100% bullshit generators (it's even been scientifically proven).

        ? 1 Antwort Letzte Antwort
        0
        • monkee@chaos.socialM monkee@chaos.social shared this topic on
        • ? Gast

          My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

          LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

          In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

          But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

          If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

          monkee@chaos.socialM This user is from outside of this forum
          monkee@chaos.socialM This user is from outside of this forum
          monkee@chaos.social
          schrieb am zuletzt editiert von
          #17

          @EmilyEnough

          "They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem."

          OMG - That's perfect. Maybe also explains why everyone loves them that much. 🤨

          1 Antwort Letzte Antwort
          0
          • monkee@other.liM monkee@other.li shared this topic on
          • ? Gast

            My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

            LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

            In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

            But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

            If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

            ? Offline
            ? Offline
            Gast
            schrieb am zuletzt editiert von
            #18

            @EmilyEnough 🏆🏆🏆

            1 Antwort Letzte Antwort
            0
            • ? Gast

              My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

              LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

              In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

              But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

              If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

              ? Offline
              ? Offline
              Gast
              schrieb am zuletzt editiert von
              #19

              @EmilyEnough Well said. This could never have been LLM-generated. 🙂👍

              1 Antwort Letzte Antwort
              0
              • ? Gast

                My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                ? Offline
                ? Offline
                Gast
                schrieb am zuletzt editiert von
                #20

                @EmilyEnough On a slight side note, have you seen this...

                https://mastodon.social/@Climatehistories/116217634545241753

                1 Antwort Letzte Antwort
                0
                • ? Gast

                  My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                  LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                  In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                  But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                  If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                  ? Offline
                  ? Offline
                  Gast
                  schrieb am zuletzt editiert von
                  #21

                  @EmilyEnough this is a very justified rant

                  But the thought of computers being too autistic so people had to turn them neurotypical by adding llms is just so funny

                  ? 1 Antwort Letzte Antwort
                  0
                  • ? Gast

                    @EmilyEnough Wow, I have thought a lot about how coding LLMs are antithetical to my own OCD tendencies that want everything to be built and formatted in a very specific way (i.e. the right way), but had not considered how terrible the interface would be for folks who prefer not to have to process information conversationally.

                    I would love to read an entire book or series of articles about how LLMs as an interface enforce neurotypical modes of communication on neurodiverse people.

                    ? Offline
                    ? Offline
                    Gast
                    schrieb am zuletzt editiert von
                    #22

                    @mikemccaffrey @EmilyEnough The "you can write natural language queries" idea has always gotten a response from me of "why the fuck would I want to do that?" Standard search engine queries and stuff are so much easier.

                    ? ? 2 Antworten Letzte Antwort
                    0
                    • ? Gast

                      My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                      LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                      In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                      But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                      If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                      ? Offline
                      ? Offline
                      Gast
                      schrieb am zuletzt editiert von
                      #23

                      @EmilyEnough Had an interesting chat with the senior director at my office recently. He pointed out that as far as he can see, he already uses natural language to explain what he wants from software. This is just faster.
                      It was a perspective I hadn't considered before, but the more I think about it the more I think it's deeply insulting.

                      ? ? 2 Antworten Letzte Antwort
                      0
                      • ? Gast

                        @EmilyEnough Had an interesting chat with the senior director at my office recently. He pointed out that as far as he can see, he already uses natural language to explain what he wants from software. This is just faster.
                        It was a perspective I hadn't considered before, but the more I think about it the more I think it's deeply insulting.

                        ? Offline
                        ? Offline
                        Gast
                        schrieb zuletzt editiert von
                        #24

                        @rupert he is telling you flat out that he plans on replacing the expensive translation layer (you) asap. By and large that’s how the entire capital class sees this technology, as a way to eliminate expensive human labor without doing any actual work themselves.

                        ? 1 Antwort Letzte Antwort
                        0
                        • ? Gast

                          @rupert he is telling you flat out that he plans on replacing the expensive translation layer (you) asap. By and large that’s how the entire capital class sees this technology, as a way to eliminate expensive human labor without doing any actual work themselves.

                          ? Offline
                          ? Offline
                          Gast
                          schrieb zuletzt editiert von
                          #25

                          @EmilyEnough I knew the plan. I just couldn't understand why he thought it would work.

                          1 Antwort Letzte Antwort
                          0
                          • ? Gast

                            My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                            LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                            In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                            But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                            If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                            ? Offline
                            ? Offline
                            Gast
                            schrieb zuletzt editiert von
                            #26

                            @EmilyEnough As your fellow ND professional computer toucher, I'm 100% with you - the unpredictability drives me batty. If I want a RNG I'll call one - what I intend to be deterministic should be, verifiably, repeatably. Lipsticked pig LLMs have snuck into what I have to do for work and beating one's head against that BS is a good way to eventually flame the fuck out of tech. Corporate controlled computing was a mistake.

                            1 Antwort Letzte Antwort
                            0
                            • ? Gast

                              @EmilyEnough Had an interesting chat with the senior director at my office recently. He pointed out that as far as he can see, he already uses natural language to explain what he wants from software. This is just faster.
                              It was a perspective I hadn't considered before, but the more I think about it the more I think it's deeply insulting.

                              ? Offline
                              ? Offline
                              Gast
                              schrieb zuletzt editiert von
                              #27

                              @rupert @EmilyEnough

                              As a system architect, this is also what I do. The thing is, I absolutely depend on the people who do the implementation having good judgement. They need to fill in the gaps (if there were no gaps, I would have an implementation already) but also tell me if there are real problems with some of the ideas. This is why the first thing I do with a design is have it reviewed by people who will implement it. If they tell me ‘actually, this thing you forgot to consider is where our critical path is’ then that often leads to a complete redesign, or at least to significant change. The LLM will just produce something. With an ‘agentic’ loop and some automated testing, it will produce something that passes my tests. But it won’t tell me I’m solving the wrong problem.

                              I don’t have a problem with constrained nondeterminism in general. There are loads of places where this is fine. The place I used machine learning in my PhD was in prefetching. Get it right and everything is faster. Get it wrong and you haven’t lost much. This kind of asymmetry is great for ML-based probabilistic approaches: the benefit of a correct answer massively outweighs the cost of an incorrect one. The other place it works well is if you have a way of immediately validating the output. I supervised a student using some machine-learning techniques to find better orderings of passes for LLVM. They were tuning for code size (in a student project, this was easier than performance, which requires more testing). You run the old and new versions, one is smaller. That gives you an immediate signal and so using non-deterministic state-space exploration is great. You (probably) won’t get the optimal solution but you will get a good one, for far less effort than trying to reason about the behaviour of the interactions between dozens of transforms.

                              It’s not clear to me that LLMs for programming have either of these properties.

                              ? 1 Antwort Letzte Antwort
                              0
                              • ? Gast

                                @EmilyEnough this is a very justified rant

                                But the thought of computers being too autistic so people had to turn them neurotypical by adding llms is just so funny

                                ? Offline
                                ? Offline
                                Gast
                                schrieb zuletzt editiert von
                                #28

                                @Chase @EmilyEnough yeah the concept of a neurotypical computer is forever living in my head rent free

                                1 Antwort Letzte Antwort
                                0
                                • ? Gast

                                  My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                                  LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                                  In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                                  But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                                  If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                                  ? Offline
                                  ? Offline
                                  Gast
                                  schrieb zuletzt editiert von
                                  #29

                                  @EmilyEnough I completely agree. This rant inspired a tangential thought. There’s a article “ChatGPT is Bullshit” that talks a lot how LLMs are bullshit generators. It starts with Harry Frankfurt’s famous essay “On Bullshit,” which defines bullshit as distinct from lying. As I recall, a lie requires 2 things: some reference to the truth (you can’t lie without knowing that what you’re saying isn’t true); and some intent. It argues that a liar needs intent and a bullshitter doesn’t care.

                                  It’s clear that LLMs have no reference to something like truth. That’s easy. But intent? The article makes a decent case that LLMs have a built in intent: deception. Pretending to be human is their intent. They “intend” to write words that are very human like. So do they have intent? Maybe. It’s part of why all the best uses of LLMs are around fraud.

                                  I thought this might be an interesting slight pivot off the idea that they don’t have intent. You’re right: they don’t have it like a human, who presumably has some point; some reason for writing what they write. But maybe there is a latent intent.

                                  https://link.springer.com/article/10.1007/s10676-024-09775-5

                                  1 Antwort Letzte Antwort
                                  0
                                  • ? Gast

                                    My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                                    LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                                    In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                                    But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                                    If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                                    ? Offline
                                    ? Offline
                                    Gast
                                    schrieb zuletzt editiert von
                                    #30

                                    @EmilyEnough Yeah, very telling that the people most excited about LLM seem to be middle managers and C-levels: people adept at the "waffling about" conversations.

                                    1 Antwort Letzte Antwort
                                    0
                                    • ? Gast

                                      @EmilyEnough Wow, I have thought a lot about how coding LLMs are antithetical to my own OCD tendencies that want everything to be built and formatted in a very specific way (i.e. the right way), but had not considered how terrible the interface would be for folks who prefer not to have to process information conversationally.

                                      I would love to read an entire book or series of articles about how LLMs as an interface enforce neurotypical modes of communication on neurodiverse people.

                                      ? Offline
                                      ? Offline
                                      Gast
                                      schrieb zuletzt editiert von
                                      #31

                                      @mikemccaffrey Neurotypicality is just one of many biases that LLMs amplify. It also amplifies the latent racism, sexism, ableism, Western ideologies that dominate English language writing online, etc.

                                      But until I read this post by @EmilyEnough , I didn’t realise what a neurodivergent torture device LLMs are. I think not enough has been written on that subject yet. My adult son is neurodivergent and an awesome programmer. He also hates LLMs with a passion. I’m now seeing how this all comes together.

                                      1 Antwort Letzte Antwort
                                      0
                                      • ? Gast

                                        My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                                        LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                                        In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                                        But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                                        If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                                        ? Offline
                                        ? Offline
                                        Gast
                                        schrieb zuletzt editiert von
                                        #32

                                        @EmilyEnough thank you, I can absolutely relate to that! ❤️

                                        the struggle that coworkers / managers don't see ambiguity or inaccuracy in requirements that they wanted me to write software for seems to be the same lack of understanding when talking with the same people about software produced by LLMs. they seem to favor "something but faster" over "correct thing" and when pointed out, the "solution" seems to be to generate multiple iteration until finally reaching a "good enough" version. this is absolutely not how I understand my profession.

                                        1 Antwort Letzte Antwort
                                        0
                                        • ? Gast

                                          My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                                          LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                                          In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                                          But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                                          If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                                          ? Offline
                                          ? Offline
                                          Gast
                                          schrieb zuletzt editiert von
                                          #33

                                          @EmilyEnough Another thing is that it seems to hijack the thinking autonomy of a lot of people. People defer to an LLM instead of putting the struggle and effort into researching and learning. I'm not anti-convenience, but when we don't need to think about things anymore, the brain's thinking facilities just atrophy.

                                          ? 1 Antwort Letzte Antwort
                                          0
                                          Antworten
                                          • In einem neuen Thema antworten
                                          Anmelden zum Antworten
                                          • Älteste zuerst
                                          • Neuste zuerst
                                          • Meiste Stimmen


                                          • Anmelden

                                          • Anmelden oder registrieren, um zu suchen
                                          • Erster Beitrag
                                            Letzter Beitrag
                                          0
                                          • Kategorien
                                          • Aktuell
                                          • Tags
                                          • Beliebt
                                          • World
                                          • Benutzer
                                          • Gruppen