I think the funniest part of this computer memory shortage is that all of it is just being bought up by OpenAI to give the illusion of a growing company, who is then immediately shelving it in warehouses and never using it. When this industry crashes the amount of brand new GPUs flooding the secondary market is going to be nuts
Timeline
Post
Remote status
Context
8
@Shadowman311 >illusion of a growing company
They were just trying to starve competitors of ram. Everyone was already aware of their tenuous situation. They have first mover advantage and are perceived as the “brand name” LLM. They’re trying to maintain this status and advance it by kneecapping everyone else. It won’t work. As long as the Chinese keep getting the results of these models for free no advancement made by spending billions will matter in the face of a competitor 6 months behind getting it for free.
They were just trying to starve competitors of ram. Everyone was already aware of their tenuous situation. They have first mover advantage and are perceived as the “brand name” LLM. They’re trying to maintain this status and advance it by kneecapping everyone else. It won’t work. As long as the Chinese keep getting the results of these models for free no advancement made by spending billions will matter in the face of a competitor 6 months behind getting it for free.
@john_darksoul >getting the results of these models for free
>getting it for free
What do you mean here exactly?
>getting it for free
What do you mean here exactly?
@WandererUber They’re getting all the weights through espionage no? I’m sure they’re still spending on hardware, but if a new version comes out with more parameters they get those models by stealing them. The new Chinese open source model that just came out straight up answered as Claude.
@john_darksoul >no?
no. I don't think so.
>The new Chinese open source model that just came out straight up answered as Claude.
which is that? They fine tune new models on agent outputs. They used to train chatGPT on GPT starting from "What follows is a conversation between a user and an AI Agent" and then they had a bunch of Nigerians also write synthetic responses for them.
The fact that the new model sometimes thinks it's Claude, doesn't mean they have the weights. There's no evidence they do. And you couldn't hide if you did, when your model is open source. Anthropic would immediately notice that.
I think you just fell for Anti-China FUD on this one
no. I don't think so.
>The new Chinese open source model that just came out straight up answered as Claude.
which is that? They fine tune new models on agent outputs. They used to train chatGPT on GPT starting from "What follows is a conversation between a user and an AI Agent" and then they had a bunch of Nigerians also write synthetic responses for them.
The fact that the new model sometimes thinks it's Claude, doesn't mean they have the weights. There's no evidence they do. And you couldn't hide if you did, when your model is open source. Anthropic would immediately notice that.
I think you just fell for Anti-China FUD on this one
@WandererUber >no. I don’t think so.
I do. It doesn’t matter if any of the companies know. What would they do about it?
I do. It doesn’t matter if any of the companies know. What would they do about it?
@john_darksoul the Chinese open source models have contributors listed and papers published. This would either be a serious legal incident or an international one, considering the temperature of the rhetoric already.
Since you don't have any evidence, I think we'll just leave it at that.
Since you don't have any evidence, I think we'll just leave it at that.
@WandererUber I think you might be Chinese
Replies
0Fetching replies…