Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Unique data set. And Elon. And with Elon, comes a great set of talent. From https://x.ai/about

> Our team is led by Elon Musk, CEO of Tesla and SpaceX. Collectively our team contributed some of the most widely used methods in the field, in particular the Adam optimizer, Batch Normalization, Layer Normalization, and the discovery of adversarial examples. We further introduced innovative techniques and analyses such as Transformer-XL, Autoformalization, the Memorizing Transformer, Batch Size Scaling, μTransfer, and SimCLR. We have worked on and led the development of some of the largest breakthroughs in the field including AlphaStar, AlphaCode, Inception, Minerva, GPT-3.5, and GPT-4.

They are already competitive despite the late start: https://x.ai/blog/grok-1.5v



Does it, though? That was probably true pre-X™, but it seems like the primary selection metric has gone from “competence” to “doesn’t ever contradict Elon”


>Collectively our team contributed some of the most widely used methods in the field,

So they hired some guys from other AI companies.


What unique dataset? Tweets?


Yup. They're going to have the greatest training set of trolls, shitposts, and propaganda.

The only universe they're going to end up understanding is the one inside Elon's head.


I, for one, eagerly await the new insights into the universe which will be unlocked by training an AI on dril's tweets.


But we know from Google that unless you can definitively solve the "is this sentence real or a joke" datasets like Twitter, Reddit etc are going to be more trouble than they are worth.

And Elon's recent polarising nature and the callous nature with which he disbanded the Tesla Supercharger team means that truly talented people aren't going to be as attracted to him as in his early days. They are only going to be there for the money.


The datasets should not be used for knowledge but to train a language model.

Using it for knowledge is bonkers.

Why not buy some educational textbook company and use 99.9% correct data? Oh and use RAG while you are at it so you can point to the origin of the information.

The real evolution still has to come though, we need to build a reasoning engine (Q*?) which will just use RAG for knowledge and language models to convert its thought into human language


How does one differentiate knowledge from the language model in an LLM? At least in a way that would provide a benefit?


You use formal verification for logic and rags for source data.

In other words - say you have a model that is semi-smart, often makes mistakes in logic, but sometimes gives valid answers. You use it to “brainstorm” physical equations and then use formal provers to weed out the correct answer.

Even if the llm is correct 0.001% of the time, it’s still better than the current algorithms which are essentially brute forcing.


I’m still confused as to the value of training on tweets though in that scenario?

If you need to effectively provide this whole secondary dataset to have better answers, what value do the tweets add to training other than perhaps sentiment analysis or response stylization?


I still fondly remember the story an OpenAI rep told about fine-tuning with company slack history. Given a question like "Can you do this and that please." the system answered (after being fine-tuned with said history) "Sure, I'll do it tomorrow." Teaches you to carefully select your training data.


>Twitter Supercharger team

interesting.


Unique? You mean tweets? Yeah sure

It's 6B down the drain. Saying grok 1.5 is competitive is a joke, if it was any good it would be ranked well in chatbot arena (https://chat.lmsys.org/). Elon is a master in hyping underperforming things and this is no exception.


No, there is no ranking for Grok. It’s not participating.

It would be hard to judge rate of improvement at this point, since the company has only been around for 1.25 years, and grok 1.5 is yet to be released for general access.


>> It’s not participating.

I wonder why


Well, grok 1.5 hasn't been released yet, except to very few private testers.


You really think investors like sequoia and a16z are dumb enough to fall for Elon hyping things up? They know who he is and They’ve seen him operate at levels basically no other entrepreneur can snd are betting on that


> You really think investors like sequoia and a16z are dumb enough to fall for Elon hyping things up?

a16z invested $350 in Adam Neumann's real estate venture - after WeWork. VCs will absolutely knowingly invest on hype if they think it's going to last long enough for them to cash out with great returns.


SBF


Elon’s created multiple 100B companies


This is the second 20B company he created. Unfortunately the other one is Twitter.


But that doesn’t mean investors can’t be stupid


I mean, he can try. The world already has a number of AI corporations headed up by totalitarian megalomaniacs though, the market may eventually reward some other course of action.


If there's one place Musk has proven his worth, it's entering a crowded market late and taking the same approach as the competition.

Ah, nevermind. He's just pissing away investor money. Must be fun!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: