• Aceticon@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      2
      ·
      21 days ago

      Yeah, if you’re supposedly in AI/ML and don’t recognize a (stupidly simplified) diagram for a Neural Network, you don’t really make stuff with it, you’re just another user (probably a “prompt engineer”).

      Even people creating Machine Learning solutions with other techniques would recognize that as representing a Neural Network.

      That should be as recognizable to a professional in that domain as a long string of 0s and 1s would be recognizable as binary to a programmer - even if you’re not working with it at that level, you recognized such building blocks of your trade.

      • NotANumber@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        21 days ago

        To be more specific this is an MLP (Multi-Layer Perceptron). Neural Network is a catch all term that includes other things such as Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), Diffusion models and of course Transformers.

        What you are arguing online is some variant of a Generative Pre-trained Transformer, which do have MLP or MoE layers but that’s only one part of what they are. They also have multi-headed attention mechanisms and embedding + unembedding vectors.

        I know all this and wouldn’t call myself a machine learning expert. I just use the things. Though I did once train a simple MLP like the one in the picture. I think it’s quite bad calling yourself a machine learning expert and not knowing all of this stuff and more.

        • Holytimes@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          1
          ·
          20 days ago

          I can’t help but read MLP as my little pony and now I’m picturing you training a series of marshmallow horses to pretend to be human for the profits of our corporate overlords on social media.