I have been using ChatGPT because it was the big name early on and I have never really looked into any alternatives. With the rapid growth of AI assisted services, I am curious to hear what others are using.

  • @[email protected]
    link
    fedilink
    93 months ago

    Asked several to write a c implementation of some basic networking stuff.

    ChatGPT: needed to refine my input, got reasonable output. Complete answers, just compile and run.

    Google: the output was just a few snippets, nothing to be used as-is.

    MSFT: terrible output, and -no suprise here- the compiled code crashed with null pointer references etc. The worst answers ever.

    For simple problems (programming low-level microcontrollers), my go to will be ChatGPT everytime.

    Google should get it’s act together, Microsoft can exit the stage.

    • @Hawk
      link
      23 months ago

      Phi3 is pretty good for the size of the model!

      Also subs of the Microsoft libraries used to train models are quite good.

      Oh and copilot, whether you like it or not, it’s quite a technical achievement in terms of response time and accuracy.

  • Rhaedas
    link
    fedilink
    73 months ago

    I’m sure many don’t have the hardware to run local, but for most things that will probably work just as well as the full models, plus you can modify them and experiment. Start with Ollama as the base to run them, and see what works best. I tend to primarily use the edited uncensored versions of llama3 like the Neural Daredevil variations.

    But just remember at any model’s base, even the biggest and best, they are at the core a predictor. This works great for some uses, not so well for others. Don’t use a screwdriver for a hammer…at least not until they merge them to be able to do both well.

  • @[email protected]
    link
    fedilink
    23 months ago

    I’ve been using ChatGPT, specialized ones on Huggingface, and a bunch of local ones using ollama. A colleague who is into this deep says Claude is giving him best results.

    Thing is, depends on the task. For coding, I’ve found all suck. ChatGPT gets you up to a point, then puts out completely wrong stuff. Gemini, Microsoft, and CodeWhisperer put out half-baked rubbish. If you don’t already know the domain, it will be frustrating finding the bugs.

    For images, I’ve tried DALL-E for placeholder graphics. Problem is, if you change a single prompt element to refine the output, it will generate completely different images with no way to go back. Same with Adobe generators. Folks have recommended Stability for related images. Will be trying that next.

    Most LLMs are just barely acceptable. Good for casual messing around, but I wouldn’t bet the business on any of them. Once the novelty wears off, and the CFOs tally up the costs, my prediction is a lot of these are going away.

  • HubertManne
    link
    fedilink
    13 months ago

    like any other technology whatever is standard or convenient. Not super wild about them in relation to how the rest of the internet and technology has been going.

  • @[email protected]
    link
    fedilink
    English
    13 months ago

    I do a lot of incredibly specific VHDL and 45GS02 asm, so the answer is none.

    Even if I didn’t do obscure things with obscure languages, answer’d still be none, because I’d rather spend a few hours learning what the code does and how to use it, instead of “just hope the output runs” while not knowing what and why it’s trying to do what it’s doing.

  • @Hawk
    link
    13 months ago

    Codestral and Yi:34b, are pretty good.