ChorizoPicozo
Member
the china show said the deep seek claims are fake news

Lol wtf is the china show?the china show said the deep seek claims are fake news![]()
We will certainly one big societal upheaval sooner or later but AI will for sure be embedded into the fabric of society - delivery, dogs, waiters, workers. Factories will produce factories via automated pipelines, AI will do the research and so on. Wars will be waged using machines and so on. However I don't expect things like that to be everywhere - certain countries and areas will stay relatively backward, albeit with mobile phones and digital currency. Hell, in my own home country hospitals still don't use computers much - still rely on pens and paper.So what do you believe the long term solution (Year 2040+) will be when AGI is reached in A.I.? And then combining that with robotics that will be physically working in your average store in the 2030s?
How you managed to miss the point is pretty astonishing.
We are going to string up the oligarchs, put their heads on a pike, ban AI, and execute anyone who tries to restart it.I agree UBI can't work for the reasons you presented. But it's the only thing we will have LOL! What other solution will there be? If many won't implement HUGE regulations to protect human societies?
I have no idea what young folks will do for work in 5-10 years.
It "knows" all about Tiananmen Square, it simply won't speak about it.Just need that code to be forked so it can learn about Tiananmen Square
The document includes what models were used as a reference, they include a destiled version of Llama.Fair enough, It’s already known it’s trained by using other models and known it was created for speed optimization. You can ask it that and it will tell you that itself. It won’t tell you what other ai models were used but will refer you to a faq. But what is its scalability with the more complex routines. Are we saying that the optimization was so bad in the other models that a bunch of recent graduates that were 1 to 2 years out of school were able to produce something better. I think what we are seeing is that a.i. models will all be specialized in routine training and usage. Also what validation have we on how much was really used to train and operate.
Haven't read too deeply into it but assuming it's fully open source that should be more than possible right?It "knows" all about Tiananmen Square, it simply won't speak about it.
The censorship on these models happen at the behavioral level in reinforcement-based post training, not at the level of the massive corpus of training text.
Unlearning the behavioral modes it has been given via additional reinforcement learning would be easy.
We will create a new purpose which may end up being akin to the matrix. In the short term it will be painful but literally we will be in the pods and eat the bugs as they say. You may not want to live a virtual life but each new generation will be more and more willing to do it. Kinda like our generation is the Amish and the next generation embraces the "tech world".Not just young people. Anybody working, regardless of age and experience could lose their jobs.
My role as a data engineer could easily be replaced by AI in a few years. All that time I spend monthly trawling through data, identifying trends, writing reports, presenting data and action plans etc could be done by AI in a fraction of the time. I can see it happening soon as well. What was the point of all that strudying, all that time learning excel, SQL and Python. All those hours spent working on my orator skills to present the data and push back against key stakeholders. Years and years wasted to be replaced by a machine.
![]()
UBI isn't the answer as people will only be given another money to meet their basic needs. People need a purpose in life and a majority don't want to sit around and not be useful. Oh, but this will give people the chance to be more creative and artistic. They can take the UBI money, then spend the time writing a book, or painting landscapes to sell for money. Well tough shit, because AI will also take that away from us as well. You know that novel you've been writing for 10 years? Well a LLM can crank out the same novel in a day.
I think if enough people are impacted then we'll start to see major push back from the public to fight for their jobs and they're humanity.
Oh my we just found GAFs secret billionaire and market guru. Are you here out of curiosity to observe and laugh at our miserable peasant lives?If a person lost their life savings over one stock, they did something seriously wrong. $NVDA has been a great play for me in the past 18 months (6 figures in realized gains and more than that in unrealized gains). I'm happy to buy the dip here as I needed to replenish the shares I've sold off anyway.
Not just young people. Anybody working, regardless of age and experience could lose their jobs.
My role as a data engineer could easily be replaced by AI in a few years. All that time I spend monthly trawling through data, identifying trends, writing reports, presenting data and action plans etc could be done by AI in a fraction of the time. I can see it happening soon as well. What was the point of all that strudying, all that time learning excel, SQL and Python. All those hours spent working on my orator skills to present the data and push back against key stakeholders. Years and years wasted to be replaced by a machine.
![]()
UBI isn't the answer as people will only be given another money to meet their basic needs. People need a purpose in life and a majority don't want to sit around and not be useful. Oh, but this will give people the chance to be more creative and artistic. They can take the UBI money, then spend the time writing a book, or painting landscapes to sell for money. Well tough shit, because AI will also take that away from us as well. You know that novel you've been writing for 10 years? Well a LLM can crank out the same novel in a day.
I think if enough people are impacted then we'll start to see major push back from the public to fight for their jobs and they're humanity.
if the chinese AI claims are legit, it means less hardware if required to produce current top-tier AI... which means fewer nvidia GPU sales.why it affects nvidia? shouldnt it affect openai if anyone at all?
The CEO on Anthropic said this to the Wall Street Journal just last week!!!!
He's right............we'll need to have a real conversation as a human race that most aren't ready to have.
I think that only really Silicon Valley nerds worry about their lives having no meaning if people don't think they are the most intelligent form of life in the universe.He's right............we'll need to have a real conversation as a human race that most aren't ready to have.
If there was ever an industry that needed heavy regulation then it is AI. AI should be tools that make life better for humanity, not replace humanity.
China won't regulate so USA won't regulate
The dangers of not being first to singularity is more scary to them than any ethical dilemmas.
One of the reasons (probably not the only one) is pricing.Not understanding the connection. Doesn't DeepSeek use Nvidia GPUs as well behind the scenes?
I'm deeply skeptical as well.the china show said the deep seek claims are fake news![]()
Guess I'm going to ask my VP of machine learning WTF he is doing not switching to R1 already. If we can spin up our own servers, we can make it FedRAMP compliant too (though I'm probably going to have to research that more)!One of the reasons (probably not the only one) is pricing.
OpenAI wants 60 bucks per 1 million output tokens generated with o1.
DeepSeekR1, which generates equivalent quality, wants 2,19.
At this price the profit margin goes down the drain. OpenAi wasnt profitable before, at these prices its questionable whether all that investment in hardware by Nvidia was worth it. Big US companies like Meta, Amazon, Google and MS have already spent that money hoping they can dictate the pricing for years to come and make it a profitable business.
R1 is open source, so everyone can spin up their servers and hence prices might go down even further (though R1 might be supported by the Chinese state for all we know, just to break the oligopoly on US Ai companies).
So yeah Nvidia can sell to a broader audience (small and medium companies can now quiet easily run very powerful distilled models locally on their premise and don't need Cloud providers like google or amazon), but Nvidia wont be able to dictate pricing as much, because when the margin with AI services goes down, no one can pay these sky high fantasy prices anymore.
Basically R1 has become the new ground level, since its open. There is no need to use anything lesser than it (or v3 for that matter). In order for US companies to keep charging high prices for their API's they need to heavily outperform with their newer models. but news of the past few weeks looked like the new models might underperform. But we will see.
the china show said the deep seek claims are fake news![]()
Nobody knows and don't listen to anybody that says otherwise. Do you like gambling? Do it. Don't? Keep investing in a good, diversified, long-term portfolio and expect 8-10-12% returns at the end of the road if everything goes well.Is this a “buy the dip” situation? I only invest in broad market ETFs and crypto.
why it affects nvidia? shouldnt it affect openai if anyone at all?
Man it all sucks. These ceos and companies getting rich with fake fucking non existing money and we have to work like idiots.
Stock is essentially vibe check gambling for rich people
I agree with most of this, but I don’t think this is correct:One of the reasons (probably not the only one) is pricing.
OpenAI wants 60 bucks per 1 million output tokens generated with o1.
DeepSeekR1, which generates equivalent quality, wants 2,19.
At this price the profit margin goes down the drain. OpenAi wasnt profitable before, at these prices its questionable whether all that investment in hardware by Nvidia was worth it. Big US companies like Meta, Amazon, Google and MS have already spent that money hoping they can dictate the pricing for years to come and make it a profitable business.
R1 is open source, so everyone can spin up their servers and hence prices might go down even further (though R1 might be supported by the Chinese state for all we know, just to break the oligopoly on US Ai companies).
So yeah Nvidia can sell to a broader audience (small and medium companies can now quiet easily run very powerful distilled models locally on their premise and don't need Cloud providers like google or amazon), but Nvidia wont be able to dictate pricing as much, because when the margin with AI services goes down, no one can pay these sky high fantasy prices anymore.
Basically R1 has become the new ground level, since its open. There is no need to use anything lesser than it (or v3 for that matter). In order for US companies to keep charging high prices for their API's they need to heavily outperform with their newer models. but news of the past few weeks looked like the new models might underperform. But we will see.
if the chinese AI claims are legit, it means less hardware if required to produce current top-tier AI... which means fewer nvidia GPU sales.
I think if enough people are impacted then we'll start to see major push back from the public to fight for their jobs and they're humanity.
China won't regulate so USA won't regulate
The dangers of not being first to singularity is more scary to them than any ethical dilemmas.
Oh my we just found GAFs secret billionaire and market guru. Are you here out of curiosity to observe and laugh at our miserable peasant lives?
That's a dangerous game.
Once we hit the singularity, there is then no going back. Pandora's box would have truly been opened, but without careful planning and worldwide cooperation, it could easily be the last invention we ever make.
I think we're delving into sci-fi territory here lolThat's a dangerous game.
Once we hit the singularity, there is then no going back. Pandora's box would have truly been opened, but without careful planning and worldwide cooperation, it could easily be the last invention we ever make.
the china show said the deep seek claims are fake news![]()
Seems legit guys. Engineers are testing it out in the US on our own servers as we speak. Here is directly from my VP of machine learning. This is literally what he gets paid for. I'm quoting verbatim. Not posting a screenshot as I'm not comfortable with it on the internet.if the chinese AI claims are legit, it means less hardware if required to produce current top-tier AI... which means fewer nvidia GPU sales.
but even if the chinese AI claims are legit (dont know yet), future top-tier AI may require tons of hardware... or it'll run on a potato, who knows.
timing's odd though... US announces $500 trillion on AI investment, then china shows it can be done cheaply. china trying to save US some money? dont think so.
The market is almost never rational in the short term. I have no stake in $NVDA (except through ETFs), but i think more efficient/capable models are great for $NVDA. The bottleneck is still capability/usefulness of the models. If they become better/cheaper, demand will surge further. Besides that CoT adds the necessity - and usefulness- of extra compute at inference time.
I think we're delving into sci-fi territory here lol
I think we're delving into sci-fi territory here lol
Not really.
Reaching the singularity means a loss of control. Once we hit this milestone, it means a superintelligent AI system could take over the process of invention and innovation.
This used to be sci-fi, but now it's a very real possibility.
Jesus Christ. Talk about missing the timeline of AI.
So cute
Yup. LLM is the end game. That's it. AI stops there. Was a good run. OpenAI and Stargate was just aiming for youtube AI videos.
Best - inference - no - matter - the - vendor
DeepSeek-V3 is dependent as much as any previous model before it on inference speed.
AMD always claims higher bandwidth than Nvidia. The opposite of your DeepSeek claim you made that "no longer need to make sure you have access to the massive bandwidth Nvidia". I discover that you do not know what you're talking about.
your maths are something else
Less efficient for what GHG? Which tasks are you trying to program? Or benchmark?
Unsloth AI last year raised 2.2x inference in LLM with 70% less RAM usage by using converting sequential code paths into parallel ones using Triton kernels.
Is your whole premise of Nvidia getting 85% of market share because of Cuda? Look at the big tech firms. None are using cuda. You're missing the whole reason why Nvidia is selling like it does. Total cost of ownership. Not Cuda. Even with AMD getting better theoretical number on papers, their racks and implementation at these big AI farm levels is total shit.
Key Findings
- Comparing on paper FLOP/s and HBM Bandwidth/Capacity is akin to comparing cameras by merely examining megapixel count. The only way to tell the actual performance is to run benchmarking.
- Nvidia’s Out of the Box Performance & Experience is amazing, and we did not run into any Nvidia specific bugs during our benchmarks. Nvidia tasked a single engineer to us for technical support, but we didn’t run into any Nvidia software bugs as such we didn’t need much support.
- AMD’s Out of the Box Experience is very difficult to work with and can require considerable patience and elbow grease to move towards a usable state. On most of our benchmarks, Public AMD stable releases of AMD PyTorch is still broken and we needed workarounds.
- If we weren’t supported by multiple teams of AMD engineers triaging and fixing bugs in AMD software that we ran into, AMD’s results would have been much lower than Nvidia’s.
- We ran unofficial MLPerf Training GPT-3 175B on 256 H100 in collaboration with Sustainable Metal Cloud to test the effects of different VBoost setting
- For AMD, Real World Performance on public stable released software is nowhere close to its on paper marketed TFLOP/s. Nvidia’s real world performance also undershoots its marketing TFLOP/s, but not by nearly as much.
- The MI300X has a lower total cost of ownership (TCO) compared to the H100/H200, but training performance per TCO is worse on the MI300X on public stable releases of AMD software. This changes if one uses custom development builds of AMD software.
- Training performance is weaker, as demonstrated by the MI300X ‘s matrix multiplication micro-benchmarks, and AMD public release software on single-node training throughput still lags that of Nvidia’s H100 and H200.
- MI300X performance is held back by AMD software. AMD MI300X software on BF16 development branches have better performance but has not yet merged into the main branch of AMD’s internal repos. By the time it gets merged into the main branch and into the PyTorch stable release, Nvidia Blackwell will have already been available to everyone.
- AMD’s training performance is also held back as the MI300X does not deliver strong scale out performance. This is due to its weaker ROCm Compute Communication Library (RCCL) and AMD’s lower degree of vertical integration with networking and switching hardware compared to Nvidia’s strong integration of its Nvidia Collective Communications Library (NCCL), InfiniBand/Spectrum-X network fabric and switches.
- Many of AMD AI Libraries are forks of NVIDIA AI Libraries, leading to suboptimal outcomes and compatibility issues.
- AMD customers tend to use hand crafted kernels only for inference, which means their performance outside of very narrow well defined use cases is poor, and their flexibility to rapidly shifting workloads is non-existent.
That's nothing new and has nothing to do with inference speed or computational parallelism. They went to the metal with near assembly language on GPU. What does that have to do with GPU vendor? There's a mountain pile of alternatives to Cuda.
Peoples pick Cuda / Pytorch not for the best performance, they pick it for ease of implementation and easy to code. Do you have any idea how it is to implement DeepSeek? No you don't. It doesn't remove the needs of broad and easy coding languages. This is equivalent to saying that we should never have made an API because assembly. Sure, nothing beats that, but good luck coding.
Yup
Also a business dimwit that has never heard of the Jevons paradox. Its quite cute.
Seems legit guys. Engineers are testing it out in the US on our own servers as we speak. Here is directly from my VP of machine learning. This is literally what he gets paid for. I'm quoting verbatim. Not posting a screenshot as I'm not comfortable with it on the internet.
Me: Hey have you looked into Deep seek R1 yet?
Him:yes. The model came out last week.
Me: Are the performance numbers and costs what they claim to be?
Him: Yes. Although I am surprised why the market is reacting to it this way today. By stock market standard, it is a week (eternity) old news. Why did it catch your interest?
Me: Seeing it hit mainstream news with the Nvidia price drop. Didn't know anything about it till then
Him: They spent $5 million to crater Nvidia by $600 B![]()
I just don't buy that AI will be able to infinitely code itself with no human intervention. Generally, I believe that the more complex the system is, the more fragile. Trip up in the wrong place will lead to cascade failures.Not really.
Reaching the singularity means a loss of control. Once we hit this milestone, it means a superintelligent AI system could take over the process of invention and innovation.
This used to be sci-fi, but now it's a very real possibility.
Yeah basic human reasoning has not been emulated yet. Like I said earlier, we are spending days teaching it to read basic flow charts. AI code generation is simply going to make development faster. It will be yet another tool in the toolkit and can mayyyy be replace a handful of junior engineers that show no growth potential. Replacing $300K devs with AI? Not something I'm worried about for the rest of my career. Can it happen in the distant future? Who the hell knows. May be. May be we will invent time travel too and be visited by our future selves. Or aliens. It's all... possible...I just don't buy that AI will be able to infinitely code itself with no human intervention. Generally, I believe that the more complex the system is, the more fragile. Trip up in the wrong place will lead to cascade failures.
The tool is still being assembled, and everyone wants to own a part of it.I always knew nvidia was inflated. good to see.
Everyone going all in on AI is equally moronic. AI is just a tool. its the products that matter.