To be clear, its not that twitter is too fucked up for nestle to work with, they absolutely would if they thought it would benefit them. Its that twitter has become so toxic that they see advertising there as a net negative.
To be clear, its not that twitter is too fucked up for nestle to work with, they absolutely would if they thought it would benefit them. Its that twitter has become so toxic that they see advertising there as a net negative.
No, that was me running the model on my own machine not using deepseek’s hosted one. What they were doing was justifying blatent politcal censorship by saying anyone could spend millions of dollars themselves to follow their method and make your own model.
You’ll notice how they stopped replying to my posts and started replying to others once it became untenable to pretend it wasnt censorship.
Yes I’m aware, I was saying that the method is the same.
I dont doubt that a part of it is that they are Chinese, but I think a big part of it is that they are willing to undercut the current players by 10x on price. That has scared the crap out of the “broligarchy” (great term) who are used to everything being cozy and not competing on price with each other, only as a method to drive non-tech companies out of markets.
They see what deepseek is doing as equivalent of what Amazon did in online sales or uber in taxis, an agressive underpricing in order to drive competion out the market.
Absolutely, the big American tech firms have gotten fat and lazy from their monopolies, actual competition will come as a shock to them.
600B is comparable to things like Llama 3, but r1 is competing with openAI’s o1 model as a chain of thought model. How big that is is classified but its thought that chatGPT4 was already in the trillions and that o1 was a big step beyond that.
Maybe they should have been clearer than saying people were joking about it doing something that it actually does if they wanted to make a point.
In fairness that is also exactly what chatgpt, claude and the rest do for their online versions too when you hit their limits (usually around sex). IIRC they work by having a second LLM monitor the output and send a cancel signal if they think its gone over the line.
Oh I hadnt realised uncensored version had started coming out yet, I definitely wil look into it once quantised versions drop.
No it’s not a feature of ollama, thats the innovation of the “chain of thought” models like OpenAI’s o1 and now this deepseek model, it narrates an internal dialogue first in order to try and create more consistent answers. It isnt perfect but it helps it do things like logical reasoning at the cost of taking a lot longer to get to the answer.
Oh, by the way, as to your theory of “maybe it just doesnt know about Tiananmen, its not an encyclopedia”…
Ok sure, as I said before I am grateful that they have done this and open sourced it. But it is still deliberately politically censored, and no “Just train your own bro” is not a reasonable reply to that.
https://www.analyticsvidhya.com/blog/2024/12/deepseek-v3/
Huh I guess 6 million USD is not millions eh? The innovation is it’s comparatively cheap to train, compared to the billions OpenAI et al are spending (and that is with acquiring thousands of H800s not included in the cost).
Edit: just realised that was for the wrong model! but r1 was trained in the same budget https://x.com/GavinSBaker/status/1883891311473782995?mx=2
Wow ok, you really dont know what you’re talking about huh?
No I dont have thousands of almost top of the line graphics cards to retain an LLM from scratch, nor the millions of dollars to pay for electricity.
I’m sure someone will and I’m glad this has been open sourced, its a great boon. But that’s still no excuse to sweep under the rug blatant censorship of topics the CCP dont want to be talked about.
??? you dont use training data when running models, that’s what is used in training them.
It’s not a joke, it wont:
Because you used Russia and USSR interchangeably in your argument
It worked well against Russia and the Soviet Union, but it’s been my claim for years now, that this won’t work against China anymore. Because China is way more advanced now than the Soviet Union ever was by comparison for the time
…
China has 10 times the people Russia has
You first say what worked again Russia and the USSR (which is a bit weird, like writing new york and the USA) wont work against China, then say its because China has 10 times the pop of Russia.
I dont fundamentally disagree with you, but it should be pointed out that the USSR and its proxies in the Warsaw pact had about 0.4 billion people in 1980 to the PRC’s 1 billion. So it wasnt anywhere near 10x the size. The important difference IMO is that china has opened up economically, whereas the USSR stayed closed, and are acting out Lenin’s maxim of “the capitalists will sell us the rope to hang them”.
An alliance with the DPRK is preferable to one with the USA
most rational tankie
Sure but i can run the decensored quants of those distils on my pc, I dont need to even open the article to know that openai isnt going to allow me to do that and so isnt really relevant.