I promise this question is asked in good faith. I do not currently see the point of generative AI and I want to understand why there’s hype. There are ethical concerns but we’ll ignore ethics for the question.

In creative works like writing or art, it feels soulless and poor quality. In programming at best it’s a shortcut to avoid deeper learning, at worst it spits out garbage code that you spend more time debugging than if you had just written it by yourself.

When I see AI ads directed towards individuals the selling point is convenience. But I would feel robbed of the human experience using AI in place of human interaction.

So what’s the point of it all?

  • communism@lemmy.ml
    link
    fedilink
    arrow-up
    4
    ·
    9 hours ago

    I use LLMs for search results when conventional search engines aren’t providing relevant results, and then I can fact-check whatever answers the LLMs give me. Especially using them to ask questions that are easy to verify, like mathematical questions where I can check the validity of the answers. Or similarly programming questions where I can read through the solution, check the documentation for any functions used, and make sure the output is logical, and make any tweaks if the LLM gives a nearly-correct answer. I always ask LLMs to cite their sources so I can check those too.

    I also sometimes use LLMs for formatting, like when I copy text off a PDF and the spacing is all funky.

    I don’t use LLMs for this, but I imagine that they would be a better replacement for previous automated translation tools. Translation seems to be one of the most obvious applications since LLMs are just language pattern recognition at the end of the day. Obviously for anything important they need to be checked by a human, but they would e.g. allow for people to participate in online communities where they don’t speak the community’s language.