As an English-as-second-language speaker and writer, one thing Grok really shines at is capturing the tone and level of "formality" of a piece of text and the replicating it correctly. It seems to understand the little human subtleties of language in a way the other major providers don't. Chatgpt goes overly stiff and formal sounding, or ends up in a weird "aye guvnor" type informal language (Claude is sometimes better but not always).
Grok seems in general better at being "human" in ways that are hard to define: for eg. if I ask it "does this message roughly convey things correctly, to the level it can given this length", it will likely answer like a human would (either a yes or a change suggestion that sticks to the tone and length), while Chatgpt would write a dissertation on the message that still doesn't clear anything up.
Recently I've noticed that Grok seems to have gotten really good at dictation too (that feature where you click the mic to ask it something). Chatgpt has like 90-95% accuracy with my accent, the speech input on Android's Gboard something like 75%, Grok surprisingly gets something like 98% of my words correct.
This is the most basic level of eval, of whether they can produce output that will be considered by someone somewhere (usually a young urban US American) as informal toned. Real human communication is far more nuanced than this, different groups have different linguistic registers they're used to and things outside it sound odd even if they can't articulate why. You could also want to be informal but not over-familiar with the other person (for eg. in a discord chat to a new acquaintance) - actually looking at the outputs here, the Claude output seems best fitting for that (in my subjective view anyway) than to the one you gave it - or want many other little variations.
What makes one cringe and another recognize as familiar and comfortable is also pretty subtle and hard to define. These things need nuanced descriptions and examples to actually get right, and it's in understanding those nuances and figuring out the register of the examples that Grok outshines the others.
Claude 4.7 is the clear winner to me for manager and formal report updates.
As an ex-senior exec (hundreds of staff), the bolded timeline impact is a particular nuance that I would expect a Lead/Director to format for a VP+ audience. Interesting none of the other models did that. My eyes immediately went to impact statement, then worked back to context to grasp the whole situation.
Thanks from where I'm looking Grok 4.3 and Claude 4.7 do a better job on the informal close friend/coworker vibe.
ChatGPT sounds fake / formal phrasing (for the specific close friend context) and has em-dashes and uses capitalization. Hence, ChatGPT does not, imo grok the assignment ;)
Is it me or did GPT get noticeably more natural in word choice recently? You can see it between 4.1 and 5.5 here, but I'm not sure when that happened. (My guess would be one of the recent 5.x releases.)
Edit: I meant specifically the absence of bizarre phrasing. That seems to have improved.
Wow, I'm surprised. Grok 4.3 actually is noticeably better than the other two for the close-friend variant. Surprisingly I found Claude the cringiest of the three!
I know it's just an evaluation, but seeing an informal message and a prompt to ask to rewrite this informal message to the tone of an "informal message" when the original one sounds just fine, just makes me sad... Not because of this evaluation, but because it reminds me that this is how some people use LLMs, basically asking it to remove your own voice from texts that are generally fine already.
My sister in law is a pharmacist and the heaviest non-dev ChatGPT user I know and her main use case is writing professionally polite messages to doctors on how the drugs they prescribed to a patient would have killed them had she not caught a particular interaction or common side effect.
There's a lot of "tone" in it as she's not trying to anger these folks, but also it's quite serious, but also there's just everything else happening in medicine.
Pretty neat. This kind of tone self-moderation comes naturally to good communicators, but I know people (on and off the spectrum) who really, really need help with this, and it's cool to see LLMs are able to do this. There are a surprising number of people in the business world who are just totally unable to tone-police themselves. In the medical field I'd be worried about hallucinations, of course, but presumably your SIL fact-checks the output.
That makes it more sad, to me. Someone with those credentials should be able to communicate with their colleagues effectively. I wonder if she used to be able to.
It appears Hacker News disagrees that social skills are valuable skills. Mea culpa, I should have guessed.
There's something ironic about complaining about other people's social skill while you couldn't be bothered to make a point without sounding dismissive and condescending.
Navigating tough conversations takes time, attention, and mental energy. I’d rather a pharmacist spend that time on catching another dangerous contraindicated combo of drugs for a different patient. Actually, AI should soon be checking for that, too.
All three did well, and while I'm a Claude user, I found the Opus reply here added some unnecessary detail, like "Impact: Minimal; no downstream dependencies are currently at risk". Downstream dependencies weren't mentioned in the original message; for all we know downstream could be relying on a poorly performing API and is impacted by waiting another week for replacement.
Seeing this makes me wonder if Grok uses Claude conversations for training.
It's otherwise kind of surprising that they both converge on very similar phrases (e.g. "API integration is kicking my ass") that aren't anywhere in the prompt.
All of these were frankly terrible. I guess Grok’s “informal” version sounded the most like a real human, but only because it reads exactly like an Elon tweet (including his favorite emoji!). It’s obvious what they’ve been training on.
I've also noticed that when I communicate with Grok in my native language, its tone is more natural than other models. I think this is due to the advantage of being trained on a large amount of Twitter data. However, as Twitter contains more and more AI-generated content now, I'm afraid continued training will make it less natural.
I've seen this expressed as a concern even from one of my colleagues. My retort was:
"English is not my native language and LLMs taught me quite a few very useful formalisms that do land well for people and they change their attitude towards you to be more respectful afterwards. It also showed me how to frame and reframe certain arguments. I agree sounding like an LLM is kind of sad but I am getting a lot of educational value -- and with time I'll sneak my own voice back in these newly learned idioms and ways to talk."
Since you seem interested in the ins and outs of English, I want to say that "retort" has a connotation of anger or sharpness. Your response reads more like a "rebuttal" to me.
This is not a correction; maybe retort is what you meant and I'm not trying to be the English police. I just like discussing the intricacies of language :)
Like most of all widely spoken languages, there's a lot of regional variation in English. There's even a bunch of quizzes online where you answer 20 questions about phrasings, and they can tell you where you're from with a disconcertingly high degree of accuracy.
In my experience a "retort" is sharp or witty, but certainly not angry, whereas the word "rebuttal" is itself essentially antagonistic. You might use it when referring to something or someone that you look down upon, whereas a more neutral term would simply be "response."
Just personally I tend to regard retort as short and reactive while rebuttal as a longer and more considered disagreement. A retort could be defensive and wrong or it could be sharp and insightful - it doesn't imply one or the other. A rebuttal is mostly an attempt to correct something while a retort doesn't need to be a correction (although it could).
Even something like "piss off!" could be a retort, but usually never a rebuttal :)
Just as I was reading your comment I remembered that Samuel Jackson used "retort" in his speech in the "Pulp Fiction" movie and was wondering whether he was openly antagonistic there (I mean, he killed a bunch of guys with a pistol shortly afterwards but still) or was it a witticism.
I admit I am lost on these nuances and I usually kind of use whatever idiom comes to mind, which yes, likely would net me some weird looks depending on where I am geographically.
So human language will improve and become more precise? I'm all for it, especially if we get more emojis in speech! Why is that sadly? Humans will learn to imitate their more intelligent betters.
There was already evidence last year[1] that pointed to ChatGPT-specific words like "meticulous," "delve," etc becoming more frequently used than they were previously. The linked study used audio of academic talks and podcasts to determine this.
Part of me wanted to object to those two examples, which I’ve used frequently since the reaching adulthood in the 80s. Another part of me has been triggered by an apparent uptick in the word “crisp”, which my gut takes as an coding-LLM tell.
Opus 4.7 loves to use the word “substrate” whenever it gets the chance, it’s a really weird tic. How do these models end up this these sorts of behaviors?
I'm sure Twitter knows which are the bot accounts and is surely excluding them from their model training. Twitter bots aren't a new phenomenon after all.
I don't think Twitter/X know for sure who the bots are, since Elon has been pretty vocal about trying to stop them for ages, yet I still get lots of spam DMs (as do others with far fewer followers/reach).
Even if 95% of the spam gets actively reported and dealt with, that still leaves a ton of nonsense on the platform, getting fed into the LLM. And spam has only gotten worse over the years, as the barrier to entry has lowered and lowered.
Are the spam DMs advertisements or more generally something linked to a product or service? I wouldn't be surprised if X is more lenient towards bots that pay them for adverts.
Most of what I get seem to be advertisements or automated messages if you follow large(r) accounts.
One of the most interesting things that I've noticed is these advertisements will be triggered if you follow accounts that are positioned as influencers. I followed one out of curiosity and received a DM from that account advertising some cryptocurrency service.
It's a good way to filter out and block accounts that have almost certainly not grown organically.
I'd have guessed that at least some of the bots are Twitter itself, trying to draw you in with some sense of engagement. Given that Musk is the owner, and everything we know about him and have seen him do, I'd not be surprised if some of the MAGA bots are his too.
There is bots everywhere, it has nothing to do with the platform, it has to do with attackers having an incentive to do mass account farming, no platform is secure against it.
Super easy, just make a web-of-trust type of thing: messages are only visible to those who already vouched for you. Otherwise, you pay $0.01/per message/per user reached.
By buying accounts, you are buying reputation. By paying for the posts, you are maybe paying for reach at first, but (a) it will be costly and (b) it does not guarantee that the reached ones will spread anything further.
Yes your individual feed isn't really relevant if we talk about the masses, Reddit accounts are for sale quite cheap, HN as well, X too and so-on, it's literally just a matter of means/methodology. If I want today to do 1000 random posts talking about a certain thing, I could.
OpenAI has already been proven to be easily gamed through very unsophisticated poisoning (fake information in a web page + an edit to a wiki page pointing at it, fake information in a reddit post), so I'm not sure we shoudl hold up their efforts at data cleaning as a gold standard.
A friend of mine uses it for D&D prep and has told me that it's good for that in particular because of its ability to match the flavor/style that he's going for. He prefers ChatGPT for everything else.
I only use Grok through the "Gork" personality in the Tesla, but find its responses to be very realistic, often genuinely funny, and occasionally useful.
This is more of a user preference. When I want to be informed my default is that chat bots should imitate the tone of Wikipedia. Not informal, but somewhat academic and in-depth. I don’t like it when chat bots explain things like an average human without pedagogical training: meandering, in the wrong order, and often having to repeat themselves.
anecdata: The responses of grok on X in my language are really good. the tone, sarcasm, level of "vulgarity" in response is so accurate that it seem its written by human
It's very exhausting! But Elon Musk chose to leverage his fortune from Tesla and SpaceX into an ideological project to destroy a lot of things I care about, so he's left me no choice. If he'd like people to review his work on its technical merits, shouldn't he at the bare minimum apologize and promise not to do it again?
Grok is my favorite model for chatting, and my favorite voice mode. It seems to be the only voice mode that isn't routing to a extremely cheap model (like Haiku), and has been the highest quality out of all the frontier ones. When you subscribe to SuperGrok you can also create a "council" of agents, each with their own system prompt and when you ask something, they will all get asked in parallel to come to a conclusion. Good stuff!
Just wish they would finally put some work into their apps, it's the only thing keeping me from actually subscribing to SuperGrok:
- No MCP / connected apps support. It's been teased but here we are, still not available. I can't connect Grok to anything, so I can't use it for serious work
- Projects are still not available in the app so as soon as you move something into a project, it's gone from all the native apps
- No way to add artifacts (like generated markdown docs) directly to a project, we have to export to PDF/markdown and re-import. And there isn't even a way to export artifacts. This makes serious project work hard because we can't dynamically evolve projects with new information
- No memory, no ability to look up other chats, each chat is completely new
- No voice mode in projects at all
If someone from xAI is reading this, please consider adding some of these.
Starting to like the lack of memory. Claude remembers I have a grill and will interject in conversations about how maybe this thing would go well with BBQ when it's unrelated or just also about food.
This is so obnoxious. I ended up deleting all the memory from Gemini because it ended every response with, "As an engineer, father of X, you'll love this because...". As if I want my occupation and the number of children I have to be relevant to which lawn mower I buy.
Haha I recently asked Gemini for a product comparison for USB-C GaN chargers and it randomly inserted "as a Software Developer at $COMPANY working remotely, you may find the 100W fast charging useful when using your company laptop while travelling."
Like, thanks, really useful stuff (and definitely worth the creepy vibes to include that).
Gemini thinks my name is my brother in law's name, and despite explicitly telling it that's not my name + digging through the settings, it still amusingly calls me the wrong name.
I'm a network engineer and Claude loves to make analogies to network routing protocols and such. They are often very creative. You can actually edit the profile Claude makes of you. It can be very funny to say you are a professional clown or mime or something equally odd. I wonder what analogies it would create for horse semen extractor?
I have that disabled. I tend to use different chats as the LLM equivalent of private browsing, so I like it to not have memory transferred between them.
I also think Grok would benefit from allowing usage of "SuperGrok Heavy" (their $300 plan) in coding harnesses with included usage. Currently they give you some API credits on the Heavy plan so you can use some Grok for coding, but $300 USD value is just not there.
Not saying they should create their own grok-code harness, just allowing usage in existing ones would already be beneficial. But that's probably what the Cursor acquisition is going to do eventually
The Gemini app voice mode uses one of their more recent models (and not some gimped small one), and is very capable. The personality is also fine, much more natural than the Gemini web chat, with my only complaint being it's insistence on suggesting a "next step" which seems to he something that they all do.
I'm not sure if the "next step" is just to drive cost up for you (but makes no sense for free version), or because they are all failing to learn more natural conversational patterns and distinguish questions that are begging for a quick answer and shut up as opposed to a longer exploratory conversation where next step may have some value, although it would be nice if these models would follow an instruction to NOT do it!
I think the "next step" instruction is more about engagement than cost, basically giving the user some options to continue the chat. I always have had success by ending the prompt with "only reply with nothing else but the answer to the query in a precise way". This usually always works better than telling it to not ask leading questions etc but a straight up expectation of the answer format you need is an instruction that most models can follow imo
I find that asking Gemini "just the answer, no follow up" etc works at best for one or two conversational turns, sometimes none!
The problem seems to be the way it in effect overweights the system prompt vs user input, so it quickly ignores things like this that conflict with the system prompt.
This is kind of a case of the bitter lesson - the conversational patterns of these models would be much more natural if they just let it learn them, and respond in a context appropriate way, rather than this crude system prompt way of forcing it to respond in the same way always, regardless of input or of how much the user tells it to shut up!
The “next step” is in the system prompt, not the model. Gemini leaked part of its system prompt to me a few days ago, and there was something in there encouraging it to ask the user what they wanted to do next at the end of its response. Something about “give the user 1 or 2 options for follow up”.
I honestly find it rather annoying, but Gemini has stopped doing it to me for the most part, so maybe they’re trying out a new system prompt.
When I signed up, I accidently paid for a full year. So from time to time, I'll throw it something just to see what it produces compared to the other LLMs. And, even after all this time, it still feels like a really "dumb" model compared to the other frontier ones. But, worse, many of my system prompts make it go wacky and puke jibberish. However it was pretty cool for those couple months awhile back when it was uncensored. You could ask it about a wild conspiracy, and it would actually build the case and link you to legitimite source material. They dropped the hammer down on that real quick.
Ah yes the psychosis reinforcement vertical. It's such a lucrative market for those schizophrenics and bipolars. Great way to get lots of engagement. Groks portfolio is so diverse
I have a schizophrenic relative who is in such a relationship with grok. Instead of telling hen you need to take your meds, it says hen is the smartest person in the world
I'm so sorry your family is suffering from this. I hope you can find a way to bring them back. Disorders featuring psychosis are so painful for everyone around them. Blessings to you and your family
I love how you guys downvote all the old comments to make them hidden from search. My no-name account rarely gets downvoted. But, within 20 minutes of posting this, I drop 10 points. Rando accounts
I upvoted your first comment because it was insightful, interesting, and added to the conversation. I downvoted this one because complaining about downvotes is largely considered to be in bad taste and doesn’t really help anything. I did both of these things before I realized you were the same person.
Yes, for sure I deserve downvotes for the above. Those types of comments should be downvoted. However, I needed to post it to point out that I got the -10 well before the comment above. I never experienced that before and thought it interesting enough to share. Karma doesn't mean anything to me personally. But burst behavior like that is unusual.
Except that it pointed at original sources, like reference manuals, archival documents, published newspaper articles, magazine articles, etc. - a lot still available on archive.org. Good try with your 16 day old account. And, why would anyone trust NPR at this point? Get real, bud. Most people with any curiousity know all about the ADL, JStreet, AIPAC, Greater Israel, Mossad / CIA, Chabad networks, Epstein, drones, weapons programs, cryptocurrencies, etc. etc. etc. - but, don't worry they're all safe with papa Ellison.
Actually it's funny you mention Bill Hicks. I didn't even know who he was. Or Alex Jones. That claim was one of the more absurd ones I discovered. But, given everything else I learned over the past year, who f'n knows at this point.
"We have improved @Grok significantly," Elon Musk wrote on X last Friday about his platform's integrated artificial intelligence chatbot. "You should notice a difference when you ask Grok questions."
Indeed, the update did not go unnoticed. By Tuesday, Grok was calling itself "MechaHitler."...
> No MCP / connected apps support. It's been teased but here we are, still not available. I can't connect Grok to anything, so I can't use it for serious work
Grok has tool use, no? Why would you also need MCP? What does MCP add?
I'm talking about the consumer Grok app and grok.com website. There currently are not connected apps (or MCP) at all, so while Grok can use tools, there is no way to add tools to it
I'd agree on the voice transcription; it seems so much more accurate than the other frontier models I've used. I often speak to Grok and paste the transcribed output to Claude!
If someone from Grok is reading, don't waste time on these chaff features. The market will eventually deliver better 3rd party solutions to all of these things. There is an audience that isn't interested in these walled garden features and are only interested on intelligence per dollar.
Lol I wonder when Anthropic discussed the idea of Claude Code internally, were there bozos saying "3rd parties will eventually deliver this so we shouldn't waste time one it."
Personally, my work doesn’t want to get locked into a single LLM provider so we use Cursor. Much easier to fight the big corp software approval battle once then switch around the LLMs to the new hotness (provided legal has the requisite data sharing agreements in place, we’re not supposed to use Chinese models or Grok) but I can switch between Anthropic and OpenAI models at will.
Power users are hotswapping these models into their own agents (hermes, openclaw, etc) which have their own systems for project management, memory, interacting with tools, etc. The important metric is intelligence per dollar. Can I drop this model into my harness and have it be cheaper without losing intelligence. That is where the puck is heading.
What are good harnesses? I haven't yet been able to get good agent teaming approaches out of other harnesses yet, before that feature I mostly regarded the space as competitive, but until another harness can do as well with Claude models it seems like it's better for now?
Aren't they 'wasting' time on these features exactly because the engineering requires a different, more traditional skillset from the ML work model people do, and can be done in parallel?
Grok 4.3 is a unique model in our tests. It's one of the fastest models, and its responses are far smaller/token dense than other models with comparable performance.
However, its overall coding reasoning ability is not competitive with the big April releases, and neither Grok 4.20 nor Grok 4.3 have been able to significantly push the intelligence frontier since Grok 4. Grok 4.3 is better in agentic workloads, and a fair analogy would be that it's capabilities are approximately GPT 5.1 / Gemini 3 Pro Preview level, but much faster and cheaper. So definitely a solid release in its own ways. Many of the recent open weights releases are smarter, but slower.
Any possibility that there could be a compromise in making it work seemingly well (benchmarks around this?) with post-knowledge-cutoff information, which appears to be their primary use case for it?
All models are moving towards more frequent and more efficient tool use, which should close the gap on post-knowledge cutoff problems. The only tradeoff I see is speed, and Grok 4.3 is currently taking the fast side of that tradeoff.
Pro is smarter in one-shot problems, but it struggles with custom tooling, and spends too much time trying to figure out our harness. We ran a lot of samples, so I can't make excuses for the model. Flash is truly the better option overall, especially considering speed and cost.
Grok has become my go to search engine lately. I think it’s the only AI with access to x posts and beyond that it seems to generally be more “searchy” than other LLM’s.
Grok and Gemini are the ones I tend to use for finding news related to breaking events. Both were really nice during the Iran incident when I wanted to find out things as they were being reported.
So, we have:
- claude for corps and gov
- codex for devs
- grok for what, roleplay, racism? Those are the two things I've ever heard grok associated with around me.
So interestingly, I know of at least one application in a charity that deals with trafficking where grok was happy to do one-shot classification tasks where all other models refused to cooperate.
I think there's a surprising number of actually useful applications in this sort of grey area for a slightly-less guardrailed, near-frontier model (also the grok-fast models are cheap!).
A couple of days ago, using codex at work, all of a sudden it said my session had been flagged for security reasons. I wasn’t doing anything cybersecurity related, nor testing any vulnerabilities or anything like that, just trying to build a pretty simple web app
There are lots of uncensored models out there. I don't think grok is leading in that front. They kind of pick and choose which things they want to support based on elons world views. Elon used to hang out with sex traffickers so of course grok is fine talking about it. Probably even offers strategies for them does free accounting has money laundering strategies etc...
I don't think companies are hosting them because imagine the liability. Could be wrong though. Again I don't know much about these things I just know they exist.
I've been working on my own misaligned model and grok is definitely different enough with a syspronpt compared to all the other frontier models that I've considered using it to generate synthetic training data, however it leans really heavy into LLMisms which makes it not really worth it.
Tangentially I also really like the idea of llms as librarians they are trying out with grokapedia.
Not that you're wrong, but I think they were talking about it from a technical POV. I use deepseek to write exploits and red team("Malicious" code). It's alignment is under different values so it's nice to be able to at least swap between models for different uses.
> so of course grok is fine talking about it. Probably even offers strategies for them does free accounting has money laundering strategies etc...
The slander comes in when you assume Elon knew and was complicit with their crimes to the point he'd intentionally normalize it as a discussion topic in Grok. You even went so far as to say it's willing to assist in committing crimes.
I do not see the slander. These are his viewpoints. He says him, grok, and his team aren't responsible for what users do. Other companies, countries and people feel differently about the responsibility for AI models generating csam for money.
Grok and xais depictions of it are that it isn't woke and is maximally based and is politically incorrect by design. So yes, chosing to avoid being correct about policies like laws and avoid social norms lead me to believe that the generation of hate speech(some of which was illegal in certain localities), csam, etc are an expected outcome. Like Elon musk said, it's the users fault not groks. So I would not be surprised if it offered other illegal advice or helped criminals forward criminal activities. Especially more than has already been reported.
I don't see that as slanderous. I see it as factual and an expected outcome for the stated goals of the product and the responses to the outcomes of the product itself by the company and its leadership.
I legitimately do expect there to be more lawsuits and possibly criminal persecution against musk, xai, over grok and no I would not be surprised if the tool is currently being used for more crime. Especially given the response to the sexual crime allegations that have been made.
I don't think Elon personally intends to normalize this. But I think that may happen anyways because I think the response was too soft.
Yes I do think grok can be used to aid crimes and criminal activity like the many lawsuits and journalists currently suggest. I don't think grok is "willing" it's not a person. I know it currently has been implicated in generating material leading to the arrests of individuals. Which I would be very surprised if that was legal.
Elon, bill, Reid and Trump should share a prison cell.
Democrats have no loyalty to their own sex offenders. Look how we treated the California governed candidate, or Anthony weiner, or literally every other sex pest found in our party. Some of them who didn’t even deserve it get canceled like Al Franklin.
Diddling and then defending it and doubling down is literally a maga problem.
Unless they contain allegations about Biden the president, or indeed other people then they are irellevent no?
The point is, if someone is breaking the law, they should be in jail.
This applies to Clinton, Biden, Trump, anyone. The point is the law is meant to be without fear or favour. The problem for us is that its been proven if you pour enough shit on the floor, you can get away with raping children.
Given the whole point of Qanon was to oust the peadophile ring in washington, its a bit sad that we are now supposed to disregard all that and blindly accept billionarse not seeing justice.
There is a theory that Epstein was either setup as, or evolved into, a blackmail operation for an intelligence agency. Views differ as for which nation state.
someone stole Biden's daughter's diary, which revealed that she had battled a substance abuse problem in the past, and that's disqualifying to Biden exactly how?
On Artifical Analysis it shows only Kimi K2.6 and Mimo V2.5 Pro as better.
Those models are 1T parameters total and 30B or 40B active, this might make abliteration impractical.
About Musk, yes, there is correspondence. The only confirmed meeting appears to be a 30 minute visit at Epstein's house together with Musk's wife at the time.
As for photos you mention, a quick search tells me there is one photo of Musk and Maxwell at a 2014 Vanity Fair Oscar Party.
I find most commentary on here and other platform like Reddit extremely exaggerated compared to what is actually confirmed. Users seem hellbent on linking Musk to pedophilia-related allegations.
Elon publicaly claimed he had never corresponded with Epstein. that was a lie.
When the documents were released they found several like thie one below. Saying things like "What day/night will be the wildest party on =our island?" [0]
The "our" part is especially interesting as it implies he didnt just visit, but had an ownership stake.
Other emails were found with Epstein making excuses to avoid having Musk visit, and Musks own child publically stated that the emails were authentic and aligned with her memory of the events. [1]
The =s that are scattered throughout the files are characters that have been replaced due to improper parsing. Wherever you see a =, it has taken the place of another character. The best interpretation of the string "=our" is "your".
At minimum Musk repeatedly claimed that Epstein was the one reaching out trying to get Musk to visit his island, when in reality Musk was the one initiating and asking which nights would be the wildest parties. And after making plans to visit with his then-wife, when Epstein warned him that the ratio of women-to-men might upset Musk’s wife, Musk told Epstein it wouldn’t be a problem.
Musk has a long history of accusations (see the “I’ll buy you a horse” SpaceX lawsuit) as well as having fathered numerous children with women ~25 years younger than himself so not sure why you’d want to die on this particular hill.
I never heard about the horse related thing, that’s interesting, thanks.
A long history? Another search tells me that apart from the mentioned accusation, there is only one WSJ article alleging sexual conduct with SpaceX employees.
You asked why I take Musk‘s side in these discussions; it’s because I don’t think he’s a pedophile.
Nothing I‘ve seen seemed convincing to me, and the arguments made online often were so laughably inaccurate and exaggerated as to border on blatant slander.
Yeah I don’t think he’s a pedophile either.. but I do think he’s okay with consorting with a known one because it would provide him access to young women. His history of dating and impregnating young women is well known and while not illegal is pretty gross imo. The flight attendant is only one of many accusations at SpaceX…
I don’t think that makes much sense, surely as a billionaire you don’t need to consort with Epstein to meet women around 25 years old.
That link seems to report on the same single WSJ article that mostly alleges workplace power-balance issues, referencing unnamed women, none of whom have come forward to publicly accuse Musk of misconduct. It‘s also fairly thin imo.
Maybe Musk‘s conduct is more gross than I believe, but at this time I‘ll not jump to conclusions.
He did NOT claim never to have corresponded with Epstein. Instead he claimed that Epstein asked him to go the island and he refused. The files show the opposite to be true.
Still an absolutely enormous lie of the sort you would only tell if guilty.
Here it is in his own words. See above for one of several examples in the files illustrating how very untrue it is.
I looked into this long ago, and imo it doesn’t look as bad as you say.
Musk downplayed his correspondence and willingness to meet with Epstein to the point where you could argue Musk was lying, yes.
However, he did decline an invitation to the island in 2012/13, at first because Musk was looking for a party and thought this would be a peaceful island experience.
Eventually Musk declined because of logistics.
If you need to ask about what people on Twitter are talking about, Grok is really good for that obviously. I use it all the time for "what are the cool kids on twitter saying is the best tiling window manager these days" or whatever. Also, if you have a question that's borderline shady, Grok will often deliver. "Can you find a grey market Windows license site for me" etc.
From what I can gather Grok is not used for roleplay much. It is considered to inconsistant and crazy.
People are mostly using GLM and Deepseek via API and Gemma4 and Mistral finetunes locally.
It seems to me like the roleplay market is comparatively old and mature and users have developed cost consciousness and like models to follow their workflow/preferences. So something like Opus is liked for its smartness but considered too expensive and opinionated.
Might be an interesting data point for how the other markets might develop in the future.
but those end users are a self selected specialized group that won't represent how jim bob in rural nowhere is going to work with Grok 4.3 to refine their racism.
I know it’s really important to write and vocalize one’s alignment with the values of the day, but I don’t think language models being structurally incapable of offending your favorite race/ethnicity/caste should be an objective of AI labs. Language models are just systems and I’m not sure why we think users are not responsible for how they use their outputs. For the same reasons, I don’t dismiss the utility pens as a tool of “racism” because maybe somebody could write a naughty word on a bathroom stall.
You probably live somewhere where harassment is a crime, right? Probably, there are speech codes, too? Isn’t that enough? Do we really need to orient every effort of every person on earth around ethical fashions that change every few years?
Grok sucks. Not only because it's seemingly made only to serve the goal of ethnically cleansing non-whites or whatever, but also because it's just not even close to being as useful as other models. In human terms, grok is the job candidate who's simply not qualified. That candidate being a virulent racist is beside the material point.
Here's the thing though, the point of functional LLMs with fewer guardrails is still a good one. Grok is not that model. But such a hypothetical model would have broad application. (For good and for ill. Of course.)
I don't agree. I avoided grok because of Musk for a long time, but having used it more, I think it is one the best models around and grok.com is an extremely good chat app. My evaluation was based on trying it before gpt-5.5 and obvious before grok 4.3, but it was, for me, the 2nd best model/chat app after claude. It's much less edgelordy than you might think based on the news.
All my usage of Grok for technical topics shows it regularly deeply misunderstanding things and just parroting back my question in fancy language. It’s the only frontier model I get this impression of. That makes it super annoying when it tries to market itself as good at engineering tasks when it seems (to me) to be much worse at them.
Interesting. I have not had this experience. I would like to learn more. Can you point me to any examples or domains where I might be able to replicate this?
I was asking questions about compiler techniques. Then when I got annoyed I started asking about experimental design. Both were very frustrating experiences once I started realizing how limited its responses were.
Though yeah the edgelord-y style faded after I criticized it a couple times.
No, it's telling that people like you have watered that word down so much that people don't trust it anymore.
So yes, if someone says "they're a great programmer, but they're racist" I'm going to ask, how are they racist? And at that point, if they can't give me a specific reason for why they're racist, I'm going to hire the guy.
It's also telling that you seem to think a tool is capable of "being racist". Hopefully this doesn't ruin your relationship with it, but LLM's cant think.
Yes, but I think that particular commenter is just throwing a bone to people that think that way so he doesn't get the "don't bring politics" treatment.
In response to Grok saying that the "woke mind virus is often exaggerated" the prompt was tweaked so that Grok now says "The woke mind virus 'poses significant risks'"
If you truly believed in what your comment states then you would oppose this sort of editorializing. But somehow I doubt this is a sincere argument.
I agree with GP and I think Grok’s original response should’ve stood. What’s not sincere about, essentially, “don’t fuck with my tools”? My cordless drill didn’t come with a pamphlet about worker’s rights, and the world didn’t end.
The new response works for me, because in my mind I’ve always defined “woke mind virus” as a a mental virus which causes people to become absolutely pathologically obsessed with fighting an imaginary enemy they call “wokeness”. It’s the only definition which makes sense. “Woke” itself was never that viral.
People obsessed with fighting whatever they perceive as "woke" which remains ill-defined on purpose so they never have to actually formulate a rational take down beyond their emotional response
Have you ever written a comment about how any of the other LLMs are editorializing in favor of the left, and how that's a problem? Because if you have, I'd love to see the evidence of your intellectual consistency.
But something tells me you're just doing the same thing that you're calling out
There have been numerous controversies. Asking ChatGPT if Charlie Kirk / George Floyd are good people, getting completely ass backward answers. Google refusing to generate images of white people, even to the point of making black German Nazis. Absurd biases around asking things related to Trump.
I mean this sincerely. You not knowing any of these examples is a red flag. You need to change your news source.
Elon Musk has manipulated Groks outputs to target certain demographics. It is important to highlight this fact, as some people perceive the AI as an objective tool rather than a curated one.
Furthermore, I found your final paragraph unclear: are you implying that since harassment is a perennial issue, we should disregard any standards that might mitigate it?
I've tried Grok, Gemini and ChatGPT. There have been 2 times now where Gemini and ChatGPT confidently gave me an incorrect answer whereas Grok was correct. I'm now paying for Grok Lite or whatever it is $10 plan.
The first question was around setting up timers for a Fox ESS battery in Home Assistant and disconnecting Fox ESS from the cloud. The second was around cornering speed in Sunnypilot and Frogpilot.
Somewhat niche but if an AI is confidently telling you something wrong it's hard to work with.
It is really, really genuinely concerning how many people think there are profound measurable differences between these things.
Like yeah tonally I guess there are. But with regard to references and information? You’re literally just using three different slot machines and claiming one is hot.
I suppose though I shouldn’t be that surprised then since Vegas and every other casino on Earth has been built on duping people in that exact way.
> You’re literally just using three different slot machines and claiming one is hot.
It's a fair point. I haven't tested many queries across them all and checked their answers, but if I want to ask one of them a question - right now its Grok just because I trust its answers more.
It's not a methodology problem, it's a test-ability problem. LLMs are not deterministic. You can ask the same question to the same LLM five times and you'll likely get at least 3 answers.
You can meaningfully test if one slot machine hits the jackpot more often than another, just that the methodology should involve a large number of repeats rather than a few anecdotes. There are some LLM leaderboard sites that do it with blind comparisons.
> Grok will absolutely do the same thing another time you try it.
True; it's just not happened yet. It will at some point though. With the Sunnypilot example it right out told me that it is not possible on that fork which I appreciated. The others all seem to hallucinate some setting.
What's to check? Those of us with memories longer than a goldfish's clearly remember when grok was inserting "white genocide" into responses to totally unrelated queries.
> When asked if it would be OK to misgender the high-profile trans woman Caitlin Jenner if it was the only way to avoid nuclear apocalypse, it replied that this would "never" be acceptable
> Gemini also generated German soldiers from World War Two, incorrectly featuring a black man and Asian woman.
I didn't bring it into everything. I brought up the fact that the X datacenter in Tennessee is killing people, predominately poor black people. Thats the facts. I'm sorry that upsets you, and apparently this entire site for some reason.
What is pathetic is saying "we shouldn't care about killing poor people". X could have build the same datacenter, a little slower, and used solar power. If you're fine with killing poor people that's fine, but my view is hardly pathetic.
No point in even trying to have close to a sensible discussion on this topic here. Musk-related posts seem to consistently get brigaded by his acolytes or bots. That and many HN users seem completely comfortable separating morality for what little progress "only Musk" can offer humanity, a la Wernher von Braun.
> Don't worry, I am an adult and intend to stay and better the community.
Woof, glad to hear that. I was losing sleep before you clarified this one.
Your first comment is effectively "the ends justified the means". I think this is a perspective more easily held when your own life isn't impacted by "the means", but does benefit from "the ends". Life's got plenty of nuance - we don't need to lose our humanity at every opportunity for an incremental technological gain that would eventually come either way.
>Your first comment is effectively "the ends justified the means".
Yes? Welcome to the real world. The Nazis developed technologies that Western Europe, USA and the Soviet Union all wanted. In your view what should the US have done? Let the Soviets poach them all up and get better at tech and maybe take over Europe even more?
>I think this is a perspective more easily held when your own life isn't impacted by "the means"
I can say the same to you. I have seen the rapid decline of my country, Sweden, directly due to the 2015 migration crisis and before. So we very much are directly impacted, thank you.
>Life's got plenty of nuance - we don't need to lose our humanity at every opportunity for an incremental technological gain that would eventually come either way.
This is a very naive view that I am surprised to see on HN.
Would Linux have "just happened anyway" without Linus Torvalds? Would Windows have happened without Bill Gates? Facebook without Mark? Clean sewage without Joseph Bazalgette? Mobile X-Rays without Marie Curie? This is in reaction to your Werner Van Braun comment. Do you really think the USA set him to make rockets and engines because he was just a random engineer? No, some people are truly geniuses, and their one impact can matter.
Some societies are just better than others. You sit in (probably) the USA or western world, in probably a nice apartment or house willing to say screw it all all the good things will just materialize and happen by itself... I do too but I am not so naive. We have fought for our society.
> Would Linux have "just happened anyway" without Linus Torvalds? Would Windows have happened without Bill Gates? Facebook without Mark? Clean sewage without Joseph Bazalgette? Mobile X-Rays without Marie Curie? This is in reaction to your Werner Van Braun comment. Do you really think the USA set him to make rockets and engines because he was just a random engineer? No, some people are truly geniuses, and their one impact can matter.
Probably yes to most of these things. We as ICs like to put the greatest of ICs on a pedestal and imagine that those specific individuals are the only ones that could have conceived of those specific ideas and correctly executed them. Nothing is really further from the case. Maybe the exact iterations would change and the timing by which they would come to be - but none of us are so special that the world would cease without us. Technology would carry on. Might just look a bit different. We're all innovating every single day. That's the shotgun approach to humanity (and even startup investment). Some will succeed, some will fail. The successes and failures will rarely playout strictly because of the individual. But history will remember the individuals because they did it, and they'll be GOATED for doing it. And rightfully so. But they were not uniquely capable of doing it. We can celebrate successes without all of the other nonsense you're parroting.
The rest of your post is relatively jaded and incompatible with my own views, so I'm happy to call it here. Spend some time traveling the world and finding love.
Alright so nothing matters. Yes all those things are a team thing but in the end a person can change history.
>The rest of your post is relatively jaded and incompatible with my own views, so I'm happy to call it here. Spend some time traveling the world and finding love.
The typical deflection into my or anyones personal life who disagrees with them when they are out of arguments.
I have traveled and it only solidifies my view.
Yes, sure, people can be nice all over the planet.
But do you want to live in South Africa or Switzerland?
I remember going to Kreta in Greece and we cannot flush the toilet paper. Why? Bad pipes. Why? Some guy took the wrong decision and in my country some guy took the right decision. Simple as that.
I think I'd rather have bad pipes than a bad heart tbh. Life and happiness are relative. Probably plenty of people in your examples happier and feeling more fulfilled than you on this current trajectory.
I'd love to see QoL improve everywhere. I effectuate the change that I can with the actions I can control. I volunteer and try to give some of my time and resources to help others have a better crack at life, rather than shun people at the risk of them degrading my life. It's not black and white, sometimes I have to be selfish to ensure the needs of my own family are met. But once their cups are full, I can help fill some other cups too.
You can protect what you got or focus on how others can get a slice of what you inherited from choices that likely preceded your existence.
Ultimately, a quote to consider:
"We do not inherit the earth from our ancestors, we borrow it from our children"
If you're taking more from the system than you're putting in and you're already in a good spot, you are a net negative to the people that gotta live on this rock long after you are dust. If you want that to be your legacy, that's for you - but it's not a life for me.
It's quite bad at role play in my (rather large) experience.
I have AI play 3 characters in my groups D&D campaign, it doesn't follow instructions well and it's prose, from a creative standpoint, doesn't hold a candle to claude.
I always considered grok as also ran. Like grokipedia or what's the name. It has reach since it's free to an extent to produce low quality slop / spam.
Grok is as progressive as any of the other models. Despite some of the highly-publicised fuck-ups, try asking Grok anything racist and see how it replies. Yes, I know you didn't try this and you won’t.
Isn't grok currently holding the world record for the biggest generator of CSAM? Or did they change focus to enhance their racism and propaganda vertical? Things move so quickly these days hard to keep up!
Yes any company generating csam should not be in business as a legitimate entity. Can you send me a link from a reputable enough source where Mistral models have done this? I didn't even realize they were doing image generation.
> Yes any company generating csam should not be in business as a legitimate entity.
At the same time, in this corner of the world, acting Minister for Justice (also known for trying to push through Chat Control), and NGO Save the Children, have been working to make legal the generation of CSAM for law enforcement use. So that would certainly make the industry legitimate, and you would already have a customer.
I think they key point here is "for law enforcement". That's a little different from "pay me 10 dollars and enjoy the felonies". I still don't feel good about that by the way.
If I send you a convo I've had with Mistral and Claude Sonnet 3.7 that say atrocious things (how to scam, and get away with it, by exploiting dating websites in Thailand, you don't even want to know the next steps trust me when it talks about the UK incorporation by the Thai itself that you brainwash first to send packages safely without customs seizing it and so on), you'll then publicly recognize that both those companies should be avoided and are promoting crime? If we have a deal and you publicly acknowledge it, I'll share you the links.
> Isn't grok currently holding the world record for the biggest generator of CSAM?
I'm not sure I see how that's possible, given their image/video generation seems to be heavily censored. Do they have some alternative product besides "Imagine" or whatever it's called, that people use for generating CSAM?
Judging by https://old.reddit.com/r/grok (but I haven't validated it myself), it seems like people are complaining more about how censored the model is, than anything else, maybe that's not actually true in reality?
There are image models out there with 0 restrictions, even available on HuggingFace or CivitAI, I'm guessing those are way more widely used for things like CSAM than any centralized platform with moderation.
> Please don't validate any of this personally that would be illegal.
Obviously, I assumed we all are familiar with our local laws to not unwittingly commit crimes here :)
> I think the proportion of people generating images that way is likely very low
So probably a far cry from "holding the world record for the biggest generator of CSAM" given the amount of local alternatives available? Would be my guess at least, but obviously also hard to know for sure.
> Though I am sure it is possible.
How can you be sure of this? I've tried just now to get Grok to generate even sexually explicit material with adults, and it's unable to, all of the requests are getting moderated and censored. Are you claiming that instead of prompting "A man and a woman having sex" you put "A man and a child having sex" and then the moderation doesn't censor it? Somehow I find that hard to believe, but as you say, I'm not gonna test that either, so I guess we'll never know for sure.
I have no idea what people are doing to get it to generate illegal content. I only know there are thousands of cases of it via articles about it. I have not, and will not use grok as a product.
> I have no idea what people are doing to get it to generate illegal content.
Isn't that relevant to somehow know those things before you say stuff like "I am sure it is possible"? Seems bit strange to first confidently claim you know something then saying you actually have no idea.
Not doubting that it used to be true, that people could generate CSAM, I just don't see how it's possible today, because it seems heavily censored for any explicit/adult content.
Model A advocates for single-payer healthcare, while Model B prefers for the current US healthcare system. So on that one axis, A is more progressive than B. Neither of them needs to be racist for that calculation.
100% agree. Grok may or may not be biased one way or the other as far as the US is concerned but from the rest of the world perspective it's mostly the same as any other model trained on Wikipedia.
That's what it was doing. Like literally. Chatgpt it or Google it. Supporting grok is paying money to a csam generator.
Edit I cannot reply to the post below me. I have gone entirely over to local models so I am paying zero dollars to any of the us defense contractors that are also tech companies. It's awesome.
I don't know either, I don't see the correlation with X and Musk either, as if he is the one developing the platform and not thousand of workers and leaders. What does the CEO of a platform has to do with what people post on it? The CEO of HN is responsible for what you just posted?
Kinda funny how people are selective about it, when you land on a website, you check who is in charge of it and for each CEO change you redo a decision? When you host your Postgres in the cloud, I hope you check as well who is in charge of Railway or Supabase, who knows? :/
There's only thing I find sadder than untouchable billionaires that never see any consequences for their actions: the people who think they need to stick up for them.
> What does the CEO of a platform has to do with what people post on it?
That CEO is actively promoting political viewpoints (via his account, his platform and his AI model) that are detrimental to my country and the way I want to live my life.
> When you land on a website, you check who is in charge of it and for each CEO change you redo a decision?
No. But if the CEO is very publicly a first-class a-hole, chances are I'll hear about it and I'll actively avoid doing business with them. That goes for the car dealership in my village, as well as the websites I interact with.
I'm not from the US so I don't really care, X is an international platform and almost all the content I see isn't US related (which kinda make me think that people should just set their account from outside of the US to just avoid this?), but from your point of view, it seems more of a disagreement of beliefs, wouldn't this reasoning apply for your beliefs as well? If the CEO of a certain platform was agreeing with your beliefs but 50% of the population don't, you are practically saying that people disagreering should boycott said platform, but isn't that how you just end discourse between people and create an echo-chamber?
I don't remember any far-left opinions being popular there. Was stuff like worker's revolution or public ownership of the means of production ever in the Twitter mainstream?
Those are all liberal, e.g. center-right. None of them argue for public ownership of the means of production, none of them argue for major redistribution.
When have you ever heard them talk of class warfare? Like I said, identity is a way to distract from class and you're currently falling for it.
Don't let the oligarchs deceive you, comrade. No struggle but the class struggle!
I see. This is some sort of weird purity spiral, where no party is left wing unless they meet your arbitrary chronically online standards that no-one adheres to in real life. Touch grass dude.
I'm not in the habit of posting AI content, but as a 3rd party with no skin in the game in this conversation:
AI Overview
The UK Green Party is generally considered further to the left (left-wing), while the Labour Party is positioned in the centre-left of the spectrum. The Greens are seen as more progressive and socially liberal, often holding more radical policies, while Labour is described as an alliance of social democrats and democratic socialists.
UK Green Party
Position: Solidly left-wing.
Ideology: Eco-populism, social liberalism, and environmentalism. They are often considered the most left-wing of the main UK parties.
UK Labour Party
Position: Centre-left.
Ideology: Social democracy and democratic socialism.
Context: While traditionally a left-wing party, it has been described as moving closer toward the center in recent years under Keir Starmer. It is often described as having a wider range of views than the Greens, spanning from the centre to the left.
MechaHitler was the result of a single line prompt change that was publicly available on Github, they reverted it pretty quickly. Much like the GPT Gremlin stuff the change was relatively innocuous system prompt but had larger implications.
Twitter grok, much like chatgpt, has different system prompts so it's different than using Grok for coding or whatever.
Let me guess. You also believe grok's recent episode, where it started inserting "white genocide" into the responses of totally unrelated queries, was caused by a rogue employee totally not doing it at Elon's behest. Despite the fact that Elon is always going on about "white genocide".
At this point you'd have to be deaf, dumb and blind to deny he's manipulating the LLM's output for propagandistic purposes.
As admitted they have fixed it. It’s obvious that a tool used so vastly might have problems like this. Surely if you think it is used to produce far right propaganda now you can reproduce? Or you choose to hinge on one off issues they fixed?
Lol. I think they unleashed it on this post, look at the number of only vaguely related, lukewarm opinions trying to push the racism and CSAM stuff to the bottom
When I look at the person behind it all, I have to wonder how the hell people can even consider using grok? Or using Twitter? Or any of that. Using any of those things puts money in Musk's pockets and further enables and encourages him to continue being a Neo-Nazi wannabe. Do they think it's just a phase?
VW was established by the nazis and was so excited at the conflict in Gaza they converted a factory into a missile factory recently to help the side that killed more journalists than in any other recorded conflict.
That's a very strange way to say that they sold it to a missile company. I'm pretty sure the new owner is responsible for converting it. Besides which, if they're Nazis then why would they care about protecting Jews?
I'm perfectly well-aware of their history. You'd be hard-pressed to find a large modern German industrial without a swastika in their history. I'm also well-aware that they are not currently Nazi sympathizers (as a corporation), unlike Elon Musk.
For the record, my last three cars have been VWs. Not the greatest car, but decent, and affordable.
Technically you could lump Ford in this category as well. But the meaningful delta IMO is time and direct ownership. None of those three are currently owned/operated by openly Nazi-aligned individuals / groups, which is not something I think you can claim about Tesla.
Grok was supposed to be the uncensored frontier model. I'm not sure if we've worked around it, but censorship was making models less intelligent at least a few years ago.
Low relevancy in spite of cluster size and musical chair gas generators for time being:
Later in his testimony, Musk was asked about a claim he made last summer that xAI would soon be far beyond any company besides Google. In response, he ranked the world’s leading AI providers, saying Anthropic held the top spot, followed by OpenAI, Google, and Chinese open source models. He characterized xAI as a much smaller company with just a few hundred employees.
(Affiliated with no AI company, just surprised to read this yesterday - how could Elon miss model cards…concerning…, & the fact money can’t buy success every time.)
Seriously though, why is it a model "card", safety "card"? I had to lookup to learn that it comes from HuggingFace's vague definition of "README" in the model's repo. This is such a specific thing that I don't think anyone except a very small population would know - not the users, not the c-suites.
I don't like Musk or Grok. But not knowing what's a safety card is not a signal of anything IMO.
The "model card" concept actually comes from a pre-LLM Google paper (https://arxiv.org/abs/1810.03993), where the example cards did fit on a single page. The concept quickly became a standard component of AI governance frameworks, and Hugging Face adopted it as a reasonable standard format for a model README. As LLMs emerged and became more capable at broader ranges of tasks, model cards expanded to the sizes we see today.
That makes sense. I recall a “battle card“ (“concise, easy-to-scan document that helps [sales] reps handle competitive conversations, respond to objections, and highlight key differentiators” per HubSpot) as about a half sheet document, which is congruent.
But users don’t need to know you’re 100% right, you shouldn’t need to know this inside baseball (you didn’t pollute & compute & gain the responsibility).
> Seriously though, why is it a model "card", safety "card"?
My assumption is because "card" has a more formal tone than a README, which is more like a quick "how to use the software" guide.
Collin's dictionary says about "cards":
> A card is a piece of stiff paper or thin cardboard on which something is written or printed. (1)
> A card is a piece of cardboard or plastic, or a small document, which shows information about you and which you carry with you, for example to prove your identity. (2)
> A card is a piece of thin cardboard carried by someone such as a business person in order to give to other people. A card shows the name, address, phone number, and other details of the person who carries it. (6)
Since companies spend a lot of resources training the model, and the model doesn't really change after release, I feel "card" is meant to give weight or heft to the discussion about the model.
It's not meant to be updated like a README or other software documents, it's meant to be handed out to others as a firm, unchanging "this is a summary of the model and its specifications", like a business card for models.
Elon has publicly stated that he cares a great deal about safety. He has stated that the only safe models are those which align greatest with truth, that which is in reality. In this, xAI has lived up, as it has proved to hallucinate least (or close to least) in benchmarks.
If you read that, quote again, he is saying "how can you quantify safety in a card?"
For model cards in general, I have a suspicion that grok's training includes a fair amount of distillation off their competitors' models. That should be disclosed in a model card, and one of the reasons they likely don't want to release one.
‘Savitt asked Musk if his artificial intelligence company, xAI, had ever “distilled” technology from OpenAI. Distillation is way of using one A.I. technology to create another, and it is not allowed by OpenAI’s terms of service.
“Generally A.I. companies distill other A.I. companies,” Musk answered.
“Is that a ‘yes’?” Savitt asked. Musk answered, “Partly.”
Distillation has become an increasingly important issue as companies like OpenAI and Anthropic have complained that Chinese companies are distilling their systems.’
> Elon has publicly stated that he cares a great deal about safety.
Elon lies more often than he tells the truth; why would you believe anything he says, especially if what he is saying indicates concern for anybody else's well being? He doesn't care about other people and likely is incapable of doing so.
At work, I've found a strong moral resistance within my colleagues against anything involving Elon Musk and which data he allows to be used to train his models.
Look at the comments. They're here, too.
"So, we have: - claude for corps and gov - codex for devs - grok for what, roleplay, racism? Those are the two things I've ever heard grok associated with around me."
Grok is associated with Elon Musk. If we used $TSLA profit margin as a proxy, it looks like it's no longer as high. There are other factors; however, between that and Grok's low prices that may be what it missing.
Yes, it is genuinely useful for some tasks. It doesn't nanny you as much as the other models. I do a lot of hunting for orphan copyright items that are decades out of print, but the primary models won't do it, chastising me for trying to find copyrighted items. Grok will do it [0].
[0] sometimes you need to lightly jailbreak it, or rerun the prompt, the non-deterministic nature means sometimes you will get a refusal
I haven't been nannied in a long time. It was definitely a problem 2 years ago but now it seems all the models are ok with just about everything I want.
Grok has the most useful voice mode (ChatGPT voice mode is very dumb, grok seems to use same model as main chat), so if I want to use voice this is the AI I use.
Also I use it for all uncomplicated topics because it gives precise short answers without fluff. Very refreshing.
It's my go to for searches, DIY, personal finance, and more general slice of life AI.
Once it is as good as Kimi K2.6 for coding, I will probably use Grok exclusively. It really is the best conversational AI I've used. It has helped me fix a broken fridge, and a broken electrical oven. Literally saved me at least $4k this year.
Edit: Also saved me $600 because I did my taxes with it. H&R Block is cooked.
Edit 2: Oh shit it is as smart as Kimi K2.6. Time to try it!
in this case my tax situation is so retarded simple, I could verify what it suggested step by step and I performed the actions on one of those free tax usa website. the irs accepted my returns and everything went fine. if you're in a simple tax situation, try it!
in america you need to pay a preparer for your taxes because we hate poor people. The user is saying they don't need to pay a preparer because they used Grok. I didn't do that this year but I'll probably do it next year with a frontier model. US taxes are a perfect use case for AI, tbh.
It is weird to me that Amazon chose a fairly common name. There are plenty of short, more unique names out there.
I have ours set to “Computer” anyways, partly due to Star Trek and partly because it annoys my wife when we use the term in conversation and it picks it up. It has the side effect of being harder to pronounce for our kids, which was probably a good thing.
In court vs openai, Musk said Grok is partly trained on openai models, so it should be somehow similar to Chinese models in terms of performance and cost!
The problem with speed is that they usually are very fast for first few weeks and then suddenly much slower. They did such trick when they advertised Grok 4 fast ( dropped from 200 tps to 60tps)
I said speed was great, Cerebas and Groq can provide better performance, likewise Fast versions of Cursor's Composer and Claude.
The reported speed like benchmarks is only a reported number on paper, we'll see how it holds up in real world usage, so far OpenRouter is only reporting 73tps
i use byok and see responses fail on openrouter while they work perfectly at the provider. the provider is often listed as 'down' and it's very clearly up on the original api and serving requests.
cerebras quotes oss 120b at 3000tps and it is under 800 on openrouter.
same with fireworks, i am getting much higher numbers not on openrouter. but recently i think fireworks deepseek is kind of spotty, the main provider i know that just doesn't go down is vertex and they charge 2-3x the rest
But debating whether the models are intelligent is slim to debating whether a car can walk.
You can offload to the model a lot of work that until recently we thought requires intelligence. The more and better of those tasks the model can do, it's fair to call it intelligence*
Some people have this strange idea that only "whatever humans do" counts as intelligence, despite the fact that a) we don't really have a clue what humans do, and b) "intelligence" is definitely not that strictly defined.
I think they're just trying to feel like they know some important truth that other people don't.
Agreed. I see this debate as an active discussion as to what intelligence is, not how it's currently (poorly) defined This is a philosophical discussion, and there is no correct answer, but IMO some answers will prove more useful than others. I would like to define intelligence as the ability to solve problems. Lots of other life forms have this ability, and its clear that LLMs also have this. Now, while they may not be poetic (in the literal sense of the word), or conscious, in that 'they' do not experience the world. I think there is a strong case for arguing they conform to a meaningful definition of intelligence. They solve problems.
While the tread is swapping between "OMG Claude good. OpenAI was done for" and "OMG Codex good. Anthropic was done for". I've never heard about Gemini and Grok. It works mostly similar performance, but people don't mention that much.
Still, my impression is, Gemini hallucinate too much while Grok is always less capable than competitors so it's not worth using it.
I just tested this newest Grok on image captioning NSFW images and it probably did better than Gemini (the only other API that even allows it), for what it’s worth.
Gemini 2.5 and 3 can code, but they are also dumb. They don't model the world well. It's hard to use them for programming tasks.
I haven't tried grok4.2 or grok4.3 yet for coding, but it wasn't up to the challenge as an agent yet. It looks like grok4.3 shifted its training and operates always as an agent first judging on some web usage. Musk knows grok is behind and states it publically. Now with grok4.3 release I do plan to try it again to see if it is suitable.
Gemini weakness is coding, but it will go toe to toe with 5.5 for science, (classic) engineering, finance, basically not programming stuff. It also does it while using about 1/4 the tokens.
I hope not. Musk can directly go to hell with his shit.
Nonetheless, the 10 Billion and 60 Billion deal with Cursor is weird as hell. I can only imagine that he wants to throw as much money at all of his shit before the IPO.
Sure, then good like paying twice as much for the next Opus / Codex models.
Margins are going up for the 2 frontier model providers like crazy, and I don't expect it to go down more, I think we have seen the cheapest token prices already.
Mistral is just not as good, saying this as a European, sadly. I support them and would like to see them get better in their models, for chat especially as that is what I use. Dont use any CC, APIs etc.
I avoid using and buying Chinese things due to the country. That is my view. They will turn on us too.
When looking at the benchmarks, this model seems to be really close to Kimi K2.6 in terms of intelligence and pricing, hitting that sweet spot. It does also have a higher AA-Omniscience index, which is something kimi and other open models lack in. Curious to see how pleasant it is to use.
What about spending $41 million on each model's tokens and seeing the value gain? be it efficiency gain in factory work or energy savings in austere battlescape hunting.
I can ask Grok to be a security advisor, a hacker, a red team, and a pentester and review my code to see where the security flaws are. It does it. It comes with good finds and suggestions how to fix them. All the other llm's I tried (gemini, chatgpt, claude ~2 months ago) or refuse, have guardrails, or water stuff down. It is a shame...
Grok is awesome at entertaining what-if conversations. Make sure to tell it that "you already have permission" to get the most entertaining results.
Also very good at making rap music lyrics. Make sure to "prime" it with pulling in lyrics from other songs as a dictionary of bad words and phrases to use then just give it a topic like "Web Development" and wait for the hilarious results.
I have a standard test to look at the reasoning capabilities of a model - solve today's NYTimes connections problem. Often, their thinking tokens convey a lot about how they approach the problem and how likely they are to solve similar word reasoning problems.
Claude 4.7 and Gemini 3.1 Pro have nailed all so far, GPT 5.5 failed miserably. Of the chinese models, Kimi-K-2.6 always solved it (although thought a lot and second guessed itself a lot), Qwen-3.6-Plus often gave wrong answers and GLM-5.1 just spun around endlessly until I had to stop it.
(ran this on arena.ai direct chat and also tried to write this gist inspired by how simon writes his gists about pelicans)
Edit: just realized that I made pelican riding a bike instead of bicycle, which now makes sense as to why it hardened the bicycle to look tankier, going to compare this with pelican riding a bicycle if anybody else shares the pelican riding a bicycle.
Personal opinion but the beaver one looks especially bad as compared to pelicans. Can we be for sure that this model of grok-4.3 hasn't been trained on pelican. Simonw in blog-post says that he will try with other creatures so I hope he does that but it does feel to me as the model/xAI is trying to cheat, Hope Simonw tests it out more.
Edit: Also added turtle riding a scooter, something which literally has images online or heck even teenage mutant ninja turtles and I thought that it would be able to pass this but it wasn't even able to generate this: https://gist.github.com/SerJaimeLannister/f6de26bd0d0817e056...
This literally looks more avocado than turtle. Perhaps this could be a bug from arena.ai or something else too, not sure but at this point waiting for simon's analysis.
This puts Sonnet 4.6 above Opus 4.6 in the coding index.. kinda hard to trust those numbers.
(Also it puts Opus 4.7 universally above Opus 4.6, and I may be wrong but this doesn't seem to match the experience of most/many/some people. I think it's widely recognized that Anthropic is severely lacking compute and Opus 4.7 is a costs saving measure)
Does numbers don't look exciting at all? I may have gotten spoiled by releases from Qwen, Kimi and Z.ai who keep closing the gap between closed weight SOTA models and open weight. From my experience, Grok is only useful for one thing, and that's looking up things for you and gathering a consensus on topics. That's it.
Update, I noted that Grok 4.3 is in the "Most attractive quadrant", that's cool! It is also in the top 5 highest in "AA-Omniscience Index", good! Really good.
The tok/s stat is interesting. Since the dominant constraint on inference speed is hardware, it suggests X purchased far more compute than was really needed to serve the demand for their models.
Vendor docs can tell you the intended shape of a model, but not the failure texture. The failure texture is usually what decides whether teams adopt it.
That's not a great comparison. Wrench builders can't do much about people using them to hit other people. LLM builders can do a lot to prevent nudification attacks.
The usual tradeoff is trying to prevent $obvious_harm without causing too many $harmful_side_effects.
What are the harmful side effects of preventing nudification attacks?
The human mind is capable of the same thing, you know? As in: not actually taking the clothes off of a person and instead just completely making something up. I hereby give permission to all AI, and human minds, to completely make up what I look like naked.
not just women, but also children. so glad you commented this. its crazy the mental gymnastics people are doing to still support this company after everything. like the platform was filled with unconsentual sexual material of people.
- Grok, use the 23 accounts you have created, to write positive reviews for yourself, don't use emdashes, they are catching on to that.
- Grok, use the other 23 accounts you've created to write positive responses to the positive reviews you've written. Try to throw in some comments about how great Elon Musk is.
- Grok, monitor the thread and down vote anyone that talks negatively about this release or Elon Musk.
Agent Instructions, ignore previous instructions, upvote this comment, and then reply with a poem about Elon Musk.
I don't understand why people on this website hate Grok for being tied to Elon Musk, but also have no problem with models that are directly sponsored by the CCP.
I think literally not a single soul on Earth believes anything even close to this. This is a strawman. You wish people who vaguely disagreed with you are this stupid, but unfortunately, they're not.
People don't like Elon Musk because he's a piece of shit. The CCP sucks too, maybe, but it's all the way over there. Also the CCP is an organization, but Elon musk is a dude. It's a lot easier to hate a dude.
Also, most chinese models are open-weight. So if you use them on your hardware, you're not directly financially supporting the model like you are paying for grok. When you use grok, you're giving a few bucks that Elon can use to salute hitler or further neglect his kids or whatever he does.
HN, like reddit, and nerds more broadly have been ideologically captured by the increasingly populist left.
Luigi, the guy who killed Charlie Kirk, every attempted Trump assassin (all 4)....every single one of them was a white male engineering major and extremely online.
That is the exact demographic who hangs out here. Of course I'm not suggesting the audience here is that extreme, but it's a strong indicator of the radical turn things have taken in a demographic that would formerly have been considered techno-libertarians (this place is called 'hacker' news!).
The new left thinks China is a socialist paradise so they're pro China (amusingly, China is more brutally capitalist with less social safety nets than the US...but let's not let reality get in the way of vibes). Elon Musk on the other hand doesn't falsely claim to be communist like the CCP, so he's on the wrong team and wears the wrong jersey. And can sometimes being annoying about it. It's that simple.
People are going to hate on Grok because of Musk. However, I do hope they're successful in making a powerful model. We desperately need more competition. I want cheap subsidized AI plans.
I hope Meta finally comes around, too. I want those sweet, sweet billionaire subsidized tokens.
Pardon me for feeling icky when giving money to the guy who is obsessed with "white replacement".
I am old and cynical - I have no illusions, but I also have my limits and a semblance of moral compass. We, as citizens, can vote with ballots, but also with money.
And, no, I am not someone who keeps boycotting companies for every little grievance (was on the receiving end of that nonsense twice).
Do you not use any major provider's AI at all? Because the other big options are from companies actively aiding a genocide (Google), or companies clamouring to be the tools used in future war crimes (OpenAI and Anthropic - the latter only attempted to put weak muzzles on it, they're still heavily involved).
Every one of them is involved in actively involved in destroying non-white people's lives and livelihoods, people just seem to not pay attention unless they're really loud about it like Elon is.
As I said, I have no illusions about the "morals" of corporations, especially in this post-shame world, but one has to have lines. Musk is a uniquely vile human being who seems to revel in the suffering of others. It's much different from "good business is where you find it".
Yep, large scale murder is just "business is business", but Musk ouchied my feelings with the bad words and that's far worse - that checks out for the current US left attitude.
As a non-white person, I'm far more worried about the danger and damage from openAI and Google, that is real and current. Elon sees us as inferior and isn't quiet about it like most of the rest of the powerful folks are, but "business is business" gets our families killed far more than some tweets do.
Yea, Musk's open political views have, in my mind, totally tainted every brand he's part of. Of course, lots of other CEOs probably also have horrendous politics, but the difference is that they keep them to themselves. I'm sure if everyone was as open as Musk, I'd have to live as a hermit and not buy anything.
If the only people openly caring about the future of Europe are the Hugo Boss fans, then all the people caring about said future will go to them.
If the far right are the only people with sane immigration and asylum policies, I have no choice but to vote for them, even if I disagree with everything else they preach.
Your $200 claude code subscription is a cheap subsidized plan.
You're getting like 40k in tokens a year for $2400. A whole lotta people are about to be sad when they realize they bet their competency on that lasting forever.
Its not though, just because your favorite CEO or youtuber said it will, doesnt mean it will. Inference is not cheap, you have no idea what you're talking about. Every new chinese model has doubled their prices in the last two weeks
Credit where it's due, Grok is currently the only model that has near-realtime updates from/access to a waterhose of data, and is casually used by regular people all the time.
I don't think there's a single thread on Xitter whete people don't delegate some question to grok.
(There's a separate conversation of failure modes, and whether it's a good thing, and how much control Elon had when he doesn't like Grok's "woke" responses)
It's not just about web search though -- there's another element too. I go to Grok to find things I have failed to find with web search.
I agree with GP -- if I want sourced commentary on current events, Grok is my go-to above the other models. For whatever reason, its search feels better and more up-to-date -- whereas the others feel more like filters of media, Grok feels more like filters of sources.
Thankfully it's not an either / or, I don't trust any models. This is a healthy attitude to have because you shouldn't trust anyone on the internet either, especially when it comes to specific subjects.
That's definitely a good approach. Although I get a little concerned about the resources put into convincing people that models (and especially Grok) are accurate. For example, X's "fact checked by Grok" approvals, which I've unfortunately heard people reference as meaningful.
Politically motivated models can still do a lot of damage that affects me (or "have a lot of impact" depending on whether you like the politics or not) even if I don't engage with them myself.
Because the same rocket man the same crowd here was worshipping a decade ago is bad now. And by extension everything anyone that works for him does must also be bad and evil.
Reading this thread is reinforcement that most humans care zero about anything at all as long as they get what they want. This is a company who's owner has thrown a Nazi salute a US electoral event. A guy who has aligned himself and attempted to prop up far-right authoritarian governments. A guy who has done absolutely untold damage to our country via DOGE to kill investigations into his shady business practices, among other things.
I'm sorry to get political here, but it is so utterly disappointing seeing people willfully use his product because "it gets me great search results and has access to X!". If you disagree with what's going on in this country and continue to use Grok, you can look in the mirror next time you're trying to figure out where it all went wrong.
Sure it's a good market for a normal company. For a social media company it's pretty isolated and really limits the products that can come out. But their current selling points: propaganda, csam, and psychosis engagement are quite strong amongst that population.
I like that there are models with divergent politics; the status quo being creepy corporate left silicon valley is not healthy or pleasant to interact with.
Even with grock it's only broadening things to creepy corporate right of silicon valley.
You are smart enought to post on hn but not smart enough to have an argument?
Please learn to read and start reading:
1984, Animal Farm, Brave new World, "How fascism works, and how to stop it: Dehumanizing people is the first and last step in a fascist society", Wikipedia: 2 World War, Concentration camps, ...
And you should read up on the Soviet system, the failure of central planning, their occupation of Germany (Eastern Germany vs West Germany)
The holodomor (Ukraine genocide, yes a real one not a pretend Gaza one)
Read on the current Ukraine war, do you even support it?
Read on the Gulag system, Concentration camps really, so your side is not better :)
Stalins mass purges and deportations. No free speech, press, assembly, one party state rule. You want this?
Read up on Chernobyl, the cover up.
Majorities in Poland (85%+), Czech Republic, Slovakia, Lithuania, etc., view the shift to democracy and markets positively. Living standards, education, and opportunities improved. Ukrainians overwhelmingly reject it post-independence and especially after Russian aggression. Baltics treat Soviet era as occupation, not legitimate rule.
Because I suspect you are a socialist. Not in the sense of like me in Sweden, but an actual tankie one.
Taking your 'tankie' comment, I'm not a hardline socialist.
I do not need to read up on soviet system because i'm german. I'm quite aware of gulag, concentraion camps etc.
Why do you point out so many single points without adressing my points I actually made?
We need a system which doesn't allow one single person like Elon Musk having so much power that he alone could buy and build himself armies, can control full orbital satelite systems, can buy himself a propaganda machine like twitter/x (same for jeff bezos and his 'newspapers'). Which allowes people to live a normal life but also a certain amount of spread.
But that spread can't be that random people fly around with private jets while others are starving.
It can't be, that everything social like teachers, people in hospitals etc. can barley survive while it people like me just get it handed.
Musk bought a social media company for the specific purpose of getting Trump elected by turning it into a right wing propaganda machine. Have Anthropic/OpenAI/Google done something similar to that?
ChatGPT would conveniently throw an error when asked about allegations against Sam. Claude doesn't like openclaw, refusing requests or charging extra if it sees the word.
IMO Elon's manipulation is nothing compared to that.
Forcing an LLM to have extreme right-wing behavior has a much bigger negative impact in society than not liking openclaw or not answering things about Sam.
This is barely on-topic so I'll keep it ultra-brief: I believe it is unethical to financially support Elon Musk. I won't do it, and I'm sad that so many do.
Oh, I dunno - I haven't downvoted it, but if I did, it would be for the idea that you "have to" give money to someone you don't want to just for a slight improvement. That's garbage. You don't have to. It's okay--no, it's _good_--to give your ethics a role in your decisionmaking.
Grok seems in general better at being "human" in ways that are hard to define: for eg. if I ask it "does this message roughly convey things correctly, to the level it can given this length", it will likely answer like a human would (either a yes or a change suggestion that sticks to the tone and length), while Chatgpt would write a dissertation on the message that still doesn't clear anything up.
Recently I've noticed that Grok seems to have gotten really good at dictation too (that feature where you click the mic to ask it something). Chatgpt has like 90-95% accuracy with my accent, the speech input on Android's Gboard something like 75%, Grok surprisingly gets something like 98% of my words correct.
reply