• bstix@feddit.dk
    link
    fedilink
    arrow-up
    2
    ·
    6 months ago

    I think LLMs are better for more fluffy stuff, like writing speeches etc.

    Excel solutions are often very specific. A vague question like separating a date can be solved in many ways, using a variety of formulas, the text-to-column wizard, VBA, import queries or even just formatting, all depending on what you really need, what the input is and what locality is used and other things.

    The text-to-column method is great, because it transforms whatever the input is into a date type, making it possible to treat it as and make calculations as an actual date. It’s not always the right solution though, for instance if the input is ambiguous.

    It’s fine that he learned to use this method, but I wonder what he’d ask the LMM in a case where it isn’t the right solution and what it’ll come up with then. He didn’t actually learn to separate a date from the input. He learned to use the text import wizard.

    In my experience it’s preferable to learn these things on a more basic level if only just to be able to search more specifically for the right answer, because there is a specific answer. Having a language model run through a bunch of solutions and presenting the most popular one might just be a waste of time and leading you into a wild goose chase.

    • AFK BRB Chocolate@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      6 months ago

      You might have missed where I said it explained both the text to columns wizard and a formula. He used the formula, which is what he was looking for. He’s a top notch software developer, he just doesn’t use Excel much.

      But I agree with your broader point. I keep having to remind people that the “LM” part is for “language model.” It’s not figuring anything out, it’s distilling what an answer should look like. A great example is to ask one for a mathematical proof that isn’t commonly found online - maybe something novel. In all likelihood, it’s going to give you one, and it will probably look like the right kind of stuff, but it will also probably be wrong. It doesn’t know math (it doesn’t know anything), it just has a model of what a response should look like.

      That being said, they’re pretty good for a number of things. One great example is lesson plans. From what I understand, most teachers now give an LLM the coursework and ask it to generate a lesson plan. Apparently they do an excellent job and save many hours of work. Anything that involves summarizing information is good, especially as that constrains the training data.