Which part of the market has slept away, exactly ?
Everything you wrote is supposition and extrapolation. Nvidia has a chokehold on the entire market. All other players still exist in the small pockets that Nvidia doesn’t have enough production capacity to serve.
And their dev ecosystem is still so far ahead of anyone else. Which providers gets chosen to equip a 100k chips data center goes so far beyond the raw chip power.
A lot. As someone that has been responsible for trainings with up to 10K GPUs, things fail all the time. By all the time I don't mean every few weeks, I mean daily.
From disk failings, to GPU overheating, to infiniband optical connectors not being correctly fastened and disconnecting randomly, we have to send people to manually fix/debug things in the datacenter all the time.
If one GPU fails, you essentially lose the entire node (so 8 GPUs), so if your strategy is to just turn off whatever fails forever and not deal with it, it's gonna get very expensive very fast.
And thats in an environment where temperature is very well controlled and where you don't have to put your entire cluster through 4 Gs and insane vibrations during take off.
> Nvidia has been using its newfound liquid funds to train its own family of models
Nvidia has always had its own family of models, it's nothing new and not something you should read too much into IMHO. They use those as template other people can leverage and they are of course optimized for Nvidia hardware.
Nvidia has been training models in the Megatron family as well as many others since at least 2019 which was used as blueprint by many players. [1]
Nemotron-3-Nano-30B-A3B[0][1] is a very impressive local model. It is good with tool calling and works great with llama.cpp/Visual Studio Code/Roo Code for local development.
It doesn't get a ton of attention on /r/LocalLLaMA but it is worth trying out, even if you have a relatively modest machine.
Deep SSMs, including the entire S4 to Mamba saga, are a very interesting alternative to transformers. In some of my genomics use cases, Mamba has been easier to train and scale over large context windows, compared to transformers.
It was good for like, one month. Qwen3 30b dominated for half a year before that, and GLM-4.7 Flash 30b took over the crown soon after Nemotron 3 Nano came out. There was basically no time period for it to shine.
It is still good, even if not the new hotness. But I understand your point.
It isn't as though GLM-4.7 Flash is significantly better, and honestly, I have had poor experiences with it (and yes, always the latest llama.cpp and the updated GGUFs).
I find the Q8 runs a bit more than twice as fast as gpt-120b since I don’t have to offload as many MoE layers, but is just about as capable if not better.
Do they have a good multilingual embedding model? Ideally, with a decent context size like 16/32K. I think Qwen has one at 32K. Even the Gemma contexts are pretty small (8K).
> you don't need to make a video model. You probably don't need to decode the latents at all.
If you don't decode, how do you judge quality in a world where generative metrics are famously very hard and imprecise?
How do you go about integrating RLHF/RLAF in your pipeline if you don't decode, which is not something you can skip anymore to get SotA?
Just look at the companies that are explicitly aiming for robotics/simulation, they *are* doing video models.
You are giving some honestly really bad and dangerous info.
The HPV strains that cause cancer and the ones that cause genital warts are different. The strains that cause cancer do not cause warts.
So you can very much have HPV without genital warts.
And conversely, while having genital warts tells you you are infected with the low risk strains, it does not guarantee you that it is the only strain you are carrying.
Thus you cannot rely on the presence of genital warts to know if you are or are not infected with the high risk strains, they are completely uncorellated.
The cancer-causing strains cause no symptoms and can only be detected by getting tested for them.
You're putting words and assumptions in my mouth that I never said in my comment. My comment includes different facts about different strains, in one comment, which some people might misinterpret. Your reply is re-stating the facts in more detail, so that's fine, I am happy for anyone to clarify information. However, the assigning of bad faith and action to me just because you don't like the way I presented the facts, is pretty rude. If you want to get really specific, we should probably clarify to the readers these statements you made:
> The cancer-causing strains cause no symptoms and can only be detected by getting tested for them
Cancer-causing strains can still cause the following symptoms: persistent sore throat, lumps, pain when swallowing, earaches (one-sided), swollen lymph nodes in the neck (painless lump), painful/difficult urination or bowel movements, unusual lumps or sores, or unexplained weight loss, in addition to others I have not listed here. However, early cancers often do not present symptoms.
> and can only be detected by getting tested for them
There is no test that covers all strains. You would need to get penile brushing, urethral brushing, semen samples, and anal pap smear. So "getting tested" is not the only solution, and getting regular scans for cancer is the best detection method. Therefore there is more involved than you have indicated, making your own comment as ;really bad and dangerous' as mine.
Perhaps we should trust people to do their own research and ask their doctor, rather than only listen to randos on the internet?
False. Not all do. And more importantly, the ones that cause cancer do not!
> Once you are confirmed HPV positive (again, you won't be confirmed without getting genital warts)
Again false. You can be tested without genital warts and be positive to a strand of HPV that simply does not cause wart. You might have had (or heard about) a bad experience with a health professional that refused to test without warts, but the presence or absence of warts has absolutely nothing to do with the strands that matter.
> you need to inform your partners, as it causes cancer in both men and women (but mostly women).
False again. Since you were specifically talking about the strands of HPV causing warts, then it does not cause cancer. You can still inform them if you care about no propagating warts, but the fact that you have wart-strand HPV does not make you more at risk of getting/causing cancer than someone with no symptoms whatsoever.
Your comment clearly says that someone with cancer-causing HPV will have warts, thus someone reading this might feel confident they are not carrying a cancer-causing strand since they do not have warts, which is dangerous because again, it is 100% false. It might also needlessly worry someone that recently noticed genital warts on themselves into thinking they might have gotten/propagated a dangerous disease, while the wart causing strand are in fact harmless and are just unpleasant aesthetically.
So tl;dr, you should get vaccinated if you can, and if you want to be sure you do not have a cancer causing strand, you need to get tested for it, that's the only way. Warts or no warts is completly unrelated.
> Perhaps we should trust people to do their own research and ask their doctor, rather than only listen to randos on the internet?
You must be looking at non tech positions, most of their research/applied role go up to ~550k, and they do offer more than advertised for strong candidates. + hiring cash bonus + equity (which is a lot).
Uncharitable take.
His last public stance on this a few months ago when he released nanochat was that he didn’t use coding LLM for it, even though he tried, because they were not good enough and he was just losing time, so coded everything manually. Andrej is already set for life, and has moved into education where most of what he does is released for free.
The author is talking about progress within one game/set/race not over the long term. In lifting during your first rep you can notice that your stance was slightly too wide and adjust it for the other 4 reps. Just like in rowing if your first paddle is a bit out of sync you can fix it.
The point is those are activities with highly repetitive efforts and you can adjust after each one with feedback.
Golfing is not like this, if you miss your first swing, you can’t micro adjust for the second one, because it’s going to take place under completely different conditions where the feedback you just got does not apply usually.
That is not my read on the post, and your point doesn't really make sense. The objective of all these activities is measurable improvement. The practical discipline in each practice session for each activity is the bare minimum to improve and every activity has it. The point of the post is that some activities tend to see consistent, linear improvement over the long term where others tend to see droughts/plateaus.
Running is a great example because a dedicated runner, even a hobbyist, can expect to see 3-5% improvement in speed/endurance or whatever every season. But no runner expects to see improvement during a single run. But the same isn't true for activities like golf or language acquisition (my own example).
Running improvement isn't actually like this, except for beginners.
First, you actually do quite a bit of periodization at the season level, so you might have a long base block, followed by a more stamina/quality oriented block, then race specific sharpening, followed by taper and an A race, and then rest. Improvement is distinctly non-linear across these phases, and you'll actually start each season fairly far behind where you were at your peak.
There are also plateau effects, where you've basically adapted as much as you can to an existing stimulus and you need to find new ways of triggering adaptation.
Just for training and processing the existing context (pre fill phase). But when doing inference a token t has to be sampled before t+1 can so it’s still sequential
Which part of the market has slept away, exactly ? Everything you wrote is supposition and extrapolation. Nvidia has a chokehold on the entire market. All other players still exist in the small pockets that Nvidia doesn’t have enough production capacity to serve. And their dev ecosystem is still so far ahead of anyone else. Which providers gets chosen to equip a 100k chips data center goes so far beyond the raw chip power.
reply