DeepSeek is an AI assistant which appears to have fared very well in tests against some more established AI models developed in the US, causing alarm in some areas over not just how advanced it is, but how quickly and cost effectively it was produced.

[…]

Individual companies from within the American stock markets have been even harder-hit by sell-offs in pre-market trading, with Microsoft down more than six per cent, Amazon more than five per cent lower and Nvidia down more than 12 per cent.

  • reksas@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    1 day ago

    i hope someone will make decent model that isnt controlled by china or america. But at least this one managed to deal decent hit to those greedy fuckers.

  • Naia@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    23
    arrow-down
    2
    ·
    2 days ago

    Been playing around with local LLMs lately, and even with it’s issues, Deepseek certainly seems to just generally work better than other models I’ve tried. It’s similar hit or miss when not given any context beyond the prompt, but with context it certainly seems to both outperform larger models and organize information better. And watching the r1 model work is impressive.

    Honestly, regardless of what someone might think of China and various issues there, I think this is showing how much the approach to AI in the west has been hamstrung by people looking for a quick buck.

    In the US, it’s a bunch of assholes basically only wanting to replace workers with AI they don’t have to pay, regardless of the work needed. They are shoehorning LLMs into everything even when it doesn’t make sense to. It’s all done strictly as a for-profit enterprise by exploiting user data and they boot-strapped by training on creative works they had no rights to.

    I can only imagine how much of a demoralizing effect that can have on the actual researchers and other people who are capable of developing this technology. It’s not being created to make anyone’s lives better, it’s being created specifically to line the pockets of obscenely wealthy people. Because of this, people passionate about the tech might decide not to go into the field and limit the ability to innovate.

    And then there’s the “want results now” where rather than take the time to find a better way to build and train these models they are just throwing processing power at it. “needs more CUDA” has been the mindset and in the western AI community you are basically laughed at if you can’t or don’t want to use Nvidia for anything neural net related.

    Then you have Deepseek which seems to be developed by a group of passionate researchers who actually want to discover what is possible and more efficient ways to do things. Compounded by sanctions preventing them from using CUDA, restrictions in resources have always been a major cause for a lot of technical innovations. There may be a bit of “own the west” there, sure, but that isn’t opposed to the research.

    LLMs are just another tool for people to use, and I don’t fault a hammer that is used incorrectly or to harm someone else. This tech isn’t going away, but there is certainly a bubble in the west as companies put blind trust in LLMs with no real oversight. There needs to be regulation on how these things are used for profit and what they are trained on from a privacy and ownership perspective.

  • FartsWithAnAccent@fedia.io
    link
    fedilink
    arrow-up
    100
    arrow-down
    2
    ·
    3 days ago

    Does that mean this stupid fucking bubble finally popped? Cramming AI into everything is getting real old real fast.

  • aesthelete@lemmy.world
    link
    fedilink
    English
    arrow-up
    35
    arrow-down
    1
    ·
    3 days ago

    Drew Carey.jpg Welcome to American capitalism, where the valuations are made up and the company financials don’t matter.

    • mosiacmango@lemm.ee
      link
      fedilink
      English
      arrow-up
      66
      ·
      edit-2
      2 days ago

      Knocking 1 trillion dollars out of a global rivals stock market for one.

      For two, making huge, huge headlines that drive huge, huge investment for your future, locked up models. That’s why facebook released llama.

      I think the first is a bonus, and the later is the reason. Deepseeks parent company is some crypto related thing which was stockpiling GPUs and opted to pivot to AI in 2023. Seems to have paid off now.

    • queermunist she/her@lemmy.ml
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      1
      ·
      edit-2
      3 days ago

      They’re outsourcing development of their platform onto independents who will work for free to advance the project, which then improves the value of their platform. It’s the same design philosophy behind the Android Open Source Project.

    • evasive_chimpanzee@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      3 days ago

      It depends on what type of licensing. One way it could be beneficial to them (and this is me purely speculating with no checking) is that any work done from outside of their company on their code base is basically free labor. Yeah, they’ll lose some potential revenue from people running their own instances of the code, but most people will use their app.

  • shoulderoforion@fedia.io
    link
    fedilink
    arrow-up
    42
    arrow-down
    3
    ·
    3 days ago

    My Alexa turns on my TV and lights, it tells me the time and the date, it tells me how many grams a half teaspoon of fresh ginger should be. I have no other use of AI. I hope everyone has a nice time with it, and remembers to hydrate. Goodbye.

    • ignirtoq@fedia.io
      link
      fedilink
      arrow-up
      30
      ·
      3 days ago

      Are you sure the answer you’re getting from AI about the weight of ginger is right? Before AI I would trust the answer from a smart speaker. Now I don’t trust anything any AI produces that should be fact-based. (Turning on lights and TV I would trust because I can see the results myself.)

      • shoulderoforion@fedia.io
        link
        fedilink
        arrow-up
        10
        arrow-down
        1
        ·
        3 days ago

        Amazon Alexa isn’t AI yet, it’s still just a smart speaker, and I don’t remember an instance it’s fucked up an answer to a quantity/weight question so badly i had to go back and research what it should have been

      • Lemminary@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        edit-2
        3 days ago

        Funnily enough, I’ve had more trouble using Gemini than the previous assistant for simple tasks like setting up a countdown. At least it loads faster, I guess.

        • barsoap@lemm.ee
          link
          fedilink
          English
          arrow-up
          11
          arrow-down
          1
          ·
          2 days ago

          It depends on the density of the ingredient, as well as the packing density, e.g. coarse vs. fine salt makes quite a difference.

          Which is why it’s silly to use volume in cooking which is why Americans are doing it.

          • raef@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            1 day ago

            Science sure, but cooking, just leave people alone. Success is evident in results and people can achieve good results with cups and spoons. It’s not a science. There’s going to be more variation in whole ingredients like eggs, temperatures, etc, than a couple of grams here and there

            • barsoap@lemm.ee
              link
              fedilink
              English
              arrow-up
              2
              ·
              edit-2
              1 day ago

              When you’re baking bread you want 1% of flour weight salt, plus minus a bit. For a quite standard bread made with 500g flour that’s 5g, being off by “a couple of grams” ranges from none at all to twice as much. With a cheap kitchen scale there’s no issue landing at 4.5-5.5g which is adequate. It’s the rest of the ingredients you can and should adjust as needed but I’m still going to measure out 300g of water because that’s the low end of what I want to put in.

              But that’s not actually the main issue, the issue is convenience up to plain possibility: The thing I actually weigh the most often is tagliatelle, 166g, a third of a pack, doesn’t need to be gram-accurate just ballpark. Try measuring differently-sized nests of tagliatelle by volume, I dare you. Spaghetti you can eyeball, but not that stuff.

              • raef@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                1 day ago

                I’ve cooked and baked all my life. I know all about the baker’s ratio. I still measure the salt in my palm. I will never weigh pasta. I don’t imagine a world where that’s that important to me.

                I think 1% is a bit low, tbh

          • dependencyinjection@discuss.tchncs.de
            link
            fedilink
            English
            arrow-up
            3
            ·
            2 days ago

            Thanks. This makes perfect sense and I agree I think recipes should use weight. I don’t know what a cup of flour is but I do know how to weight out 200g.

  • Imgonnatrythis@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    27
    arrow-down
    2
    ·
    3 days ago

    Deepseek seems to consistently fail to deliver but it’s very apologitic about it and gives the sense it’s willing to at least try harder than gpt. Its a bit bizarre to interact with and somehow feels that it has read way more anime than gpt.

    From Deepseek :

    🔗 New Wizard Cat Image Link:
    https://i.ibb.co/Cvj8ZfG/wizard-cat-leather-2.png

    If this still doesn’t work, here are your options:

    1. I can describe the image in vivid detail (so you can imagine it!).
    2. Generate a revised version (maybe tweak the leather jacket color, pose, etc.).
    3. Try a different hosting link (though reliability varies).

    Let me know what you’d prefer! 😺✨

    (Note: Some platforms block auto-generated image links—if all else fails, I’ll craft a word-painting!)

    • hotspur@lemmy.ml
      link
      fedilink
      English
      arrow-up
      15
      ·
      3 days ago

      Haha this is so amusing. I’ll take that though over the blind confidence you get out of so many other products I guess.

      • catloaf@lemm.ee
        link
        fedilink
        English
        arrow-up
        12
        ·
        3 days ago

        Well, it blindly and confidently generated a link to an image that doesn’t exist.

        • hotspur@lemmy.ml
          link
          fedilink
          English
          arrow-up
          2
          ·
          2 days ago

          Haha ok I missed that part. It doesn’t do image gen does it? I think just released a different model that does that.

  • Hello_there@fedia.io
    link
    fedilink
    arrow-up
    5
    arrow-down
    34
    ·
    3 days ago

    Just sue them for copyright infringement. All the other AI models are facing them. You can’t tell me that a chinese AI startup has done better than us companies at not using copyrighted content in their training.

    • hotspur@lemmy.ml
      link
      fedilink
      English
      arrow-up
      17
      arrow-down
      1
      ·
      3 days ago

      Well yeah, almost certainly. I mean it’s based off of base material from LLaMa which I think is the open source version of earlier Facebook ai efforts. So it definitely used copyright material for training. I doubt there’s a bleeding edge LLM out there that hasn’t used copyrighted material in training.

      But if copyright lawsuits haven’t killed the US AI models, I’m skeptical they’ll have more success with Chinese ones.

    • shoulderoforion@fedia.io
      link
      fedilink
      arrow-up
      16
      arrow-down
      2
      ·
      3 days ago

      Can’t very well sue a Chinese company, else every industry China has ripped off starting with General Electric fucking Nuclear Power plant designs. Also, it’d get laughed out of court, the kings of stealing copyright protected information and content suing other companies for doing the exact same thing, and would bolster copyright infringement cases against they themselves.

      • Tony Bark@pawb.socialOP
        link
        fedilink
        English
        arrow-up
        13
        ·
        edit-2
        3 days ago

        Yeah. At this point, China is just following in the footsteps of American companies. While the criticism is more than deserved, we’d still need to address our own country’s problems.

    • net00@lemm.ee
      link
      fedilink
      English
      arrow-up
      3
      ·
      2 days ago

      They can sue DeepSeek as much as we can sue openAI for scraping human generated content… this time though we got the model for ourselves, and not to squeeze profit

    • Miaou@jlai.lu
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      2 days ago

      Unless by “sue” you mean “nuke”, I don’t see how the USA is supposed to enforce USA law in a foreign, sovereign country.