While not in multilingual SWE benchmarks (7-9 languages other than python), this is best model in the world at J programming/understanding, while being fast, with succinct well structured thinking, as well. The multi SWE benchmarks are chart topping, and actual performance is good, meets expectations, for me, compared to GPT which tends to disappoint relative to scores compared to others.

For my use, this is the greatest AI breakthrough leap forward so far, considering speed/cost/quality of results.

    • humanspiral@lemmy.caOP
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      12 days ago

      cost of RAM is an issue, but q3 will fit in 128gb. q8 in 256. 512 is my line for “reasonably small” because single desktop computers under 1500w power circuits (with monitors) can satisfy requirements, at far lower cost than a single 140gb HBM card, where multiples fail to meet power budget.

  • humanspiral@lemmy.caOP
    link
    fedilink
    English
    arrow-up
    2
    ·
    12 days ago

    I retract my praise on this model’s ability to generate a full code file. It’s ok at thinking on one small thing at a time. Similar problems to other models on bigger tasks.