I find it absolutely mindblowing to witness the rate at which Anthropic can ship new features. Only a year ago I couldn't wait to see some sort of Github integration and then it appeared only a week later. Seriously impressive stuff.
Meanwhile, you still can't Sign in with Apple on the website.
But you can Sign in with Google.
If you signed up with your Apple on the iOS Claude app, to access your account on the computer, you have to open the passwords app and copy your random email address and paste it into the Claude website login.
Also if you try to copy-paste a prompt from Notes etc into the Claude chat, it gets added as an attachment, so you can't edit the prompt. If you do the four-finger shortcut to paste it as text, it mangles newlines etc.
Why are they so dumb about such basic UX for so long?
Add to the list backtick handling. If you start a backtick block on the claude web chat, you cannot leave it with the keyboard. You are now stuck between the backticks. It is as if they wanted to reproduce Slack misery.
> you still can't Sign in with Apple on the website.
Apple forces developers to offer Sign in with Apple on iOS devices if any other sign in service is used. Apple can't force them to do it on non-Apple platforms.
> If you signed up with your Apple on the iOS Claude app, to access your account on the computer, you have to open the passwords app and copy your random email address and paste it into the Claude website login.
Isn't this basically Apple's fault? When you signed up, Apple provided a fake email address in leu of your real one. This is great for privacy but means the service has the wrong email.
I'm sure they didn't want to provide an Apple sign in option at all, but it's required by App Store rules.
> I'm sure they didn't want to provide an Apple sign in option at all
But they wanted to provide a Google Sign In? wth?
> This is great for privacy but means the service has the wrong email.
So harm the users to benefit the service? wtf?
I don't want to give my real email or anything to random services, specially not one like Claude where they don't even let you remove your payment info.
> I don't want to give my real email or anything to random services, specially not one like Claude where they don't even let you remove your payment info.
The original complaint was:
>> If you signed up with your Apple on the iOS Claude app, to access your account on the computer, you have to open the passwords app and copy your random email address and paste it into the Claude website login.
Either you use your original email or you use a per-service email. Apple helps you do the latter, but this does come with UX tradeoffs.
Using a per-service email, then complaining that the service does not have your real email, strikes me as misguided.
They could also just implement sign in with apple on their website, they have the ability to sign in with google so not supporting Apple is still a weird choice they are making.
Apple should not have had to require developers to have options other than Google for authentication, but clearly some companies have to be dragged kicking and screaming.
So clearly they support it, and there is no reason it should not work on the web also.
There are a lot of websites that only support third party login, so that is not always an option.
They don't have to bend for another, but they made a choice to put an app on iOS. They added support for apple signin, and then for some reason did not put it on their website.
You can criticize Apple for requiring that all you want, but they clearly have support for it and are choosing to not put it on their website which is causing a worse user experience.
IF apple did not support website loggin than sure, but they do. So the ability to fix this is on Anthropic (and many other websites).
If you are already going to support third party login you should not limit it to only Google accounts and there is no reason to support Apple on iOS and not the web.
Also for the record, Apple only requires sign in with apple if you already support third party authentication. So if you are already going to support that, giving the user more choice (and making it so we are all a bit less dependent on google) is a good thing.
No criticism from me towards apple or Anthropic. Both parties made their choice. Apple was late to the identity business and the other ships had already sailed.
Third party logins are an extension and a massive risk to any website that doesn't include email hosting.
We have see identity providers dissapear, and people may change their mind.
Easiest way is to register you rown domain and use it with an identity provider of your choice and be able to move it anywhere.
Otherwise we are a faceless citizen of a corporation that can handle access to our identity and everything attached to it without recourse or access to anyone.
Are you seriously trying to justify offering Sign in with Google but not ALSO offer Sign in with Apple because of some contorted principle, the method which HELPS users maintain their privacy? What the actual f.
Antrhopic's UX is just trash, the worst of all the major AI products.
They have this "I'm special" syndrome where they think they can get away with doing shit weirdly and not offer basic features that everyone else does, and the reason why I never purchased any of their services again after the first month, and had to replace my payment info with a throwaway card because they wouldn't let me remove it, again unlike everyone else.
I don't think it's hard to understand why a service would want to support Google as an identity provider but not Apple. Google is probably the most commonly used provider out there, at least outside of the enterprise space.
Apple's identity service is not as common, and newer than the ones that were established before.
It's ok that Anthropic wasn't a fit for your prompting preference, it doesn't have to work for everyone, and it doesn't mean it wont' work for others. LLMs in general have proven that trying it once a few months ago can be a great way to miss changes. There's something out there for everyone.
Not really. It's the user's fault. Apple provides an option to hide your email, it's not required. It's an option that shows up when you're prompted to create an account.
My original thinking was that Apple makes it too easy for a general audience to hide their email without considering the implications (the service won't know your email). But of course there's a tension here, since you also want the option to be easy and accessible.
The party I do not consider at fault in this case is Anthropic.
They clearly vibe code a lot (most? all?) of their stuff, and it shows. Elementary features are broken regularly and while I appreciate them trying new features, I'd appreciate it more if existing ones were reliable and promptly fixed if broken.
Trouble is, vibe coding refinements and bug fixes works well but probably isn't a good track to promotions at Anthropic (or virtually any other company.)
They also pay... insane salaries, like double industry average. That coupled with an IPO on the horizon means they probably have their pick of engineers.
I think to the extent they are making a speed v quality tradeoff, I think they are making the right call. 10x speed over quality any day for me. Reminds me of:
"If brute force doesn't work, you aren't using enough of it." - Isaac Arthur
Everyone is
making this tradeoff now. Surely nothing bad could come from it.
In the meantime I can’t even continue a Claude Code session I started on desktop on my phone. What’s the point of shipping a billion features of they are all half baked?
It’s a phase, for sure things will turn around in the future once the hype of “oh we can now ship fast” is gone.
fwiw I’ve had this open source browser ui that sits on top of your claude code, gemini and codex and picks up/starts your sessions from any device https://github.com/siteboon/claudecodeui
Everyone is NOT making that tradeoff. Maybe we will be forced into it someday, but my team is leveraging AI to increase the quality of code far beyond what we would have done without it. Some of us are using it to engineer better solutions.
Example: we are putting a lot of energy into removing technical debt, reorganizing the code to remove unneeded abstraction and complexity, and creating missing tests and automation. We're not just burping out new untested and poorly reviewed functionality.
The Claude Code TUI app is pretty solid. I use it heavily and I get great results from it. But with the mobile app, Claude Code remote is basically unusable (weird disconnect bugs) and Claude Code cloud has issues as well (UI hides approval confirmations; must reconnect to see them). So yeah, I imagine what you're saying is true. There are at least some major gaps in their QA process. It's ironically a pretty convincing case to keep humans in the loop. It's honestly shocking to me that those features were actually shipped in their current state. You run into the problems immediately.
I have a very different experience. Claude code tui is the worst tui I have ever used. How is it possible that an inactive tui regularly eats 8gb of ram, has freezing issues and rendering issues?
If I wasn’t forced to use it I wouldn’t as there are better options available.
I agree with you about the Claude Code TUI. I switched to it weeks after it was released. The browser interface is great for quick chats and talking through ideas/concepts, but not for coding. What I love about the TUI is that it can see all your repos as once so it has the full picture all the time. You can't get that with the web version.
I don't find criticism like this particularly compelling. Most products (written by humans) have the same failings. The few that aren't are exceptions to the rule or develop very very slowly and carefully.
It was inevitable until the point all apps will disappear and AI will be the entry point for all work. You can see how anything required appear based on a single request. After which world models and other forms of interaction that are more dynamic will make sense and we'll need something that's not a screen.
Its a large leap from “we made a config driven diagram tool and trained an llm on that config” to “all apps will disappear”. If you’re predicting such grand claims please be more precise than “AI” which is a term we cant define.
Yeah an app doesn't "disappear" because you put an AI interface in front of it and then use a bunch of old school programming to parse LLM output and feed that into your old app. 99% of the work is still building the old app.
I feel like this is a feature which improves the perceived confidence of the LLM but doesn't do much for correctness of other outputs, i.e. an exacerbation of the "confidently incorrect" criticism.
This was my first thought as well, all this does is further remove the user from seeing the chat output and instead makes it appear as if the information is concretely reliable.
I mean is it really that shocking that you can have an LLM generate structured data and shove that into a visualizer? The concern is if is reliable, which we know it isnt.
Its' a reasonable concern. Often it can be mitigated by prompting in a manner that invokes research and verification instead of defaulting to a corpus.
It's a usability / quality of life feature to me. Nothing to do with increasing perceived confidence. I guess it depends on how much you already (dis)trust LLMs.
I'm finding more and more often the limiting factor isn't the LLM, it's my intuition. This goes a way towards helping with that.
A similar thing happened when Google started really pushing generating flowcharts as a use-case with Nano Banana. A slick presentation can distract people from the only thing that really matters - the accuracy of the underlying data.
As a slightly different tack, I’ve been using Copilot to generate flowcharts from some of the fiendishly complex (and badly written) standard operating procedures we have at work.
People find them quite easy to check - easier than the raw document. My angle with teams is use these to check your processes. If the flow is wrong it’s either because the LLM has screwed up, or because the policy is wrong/badly written. It’s usually the latter. It’s a good way to fix SOPs
It’s interesting you mentioned that. One of the things I’ve started doing recently is throwing a large LLM such as codex-5.3 (highest level of reasoning) at some of the more complex systems we have to produce nicely formatted ASCII diagrams.
I still review each diagram afterward, but the great thing is that, unlike image-based diagrams, they remain fully text-readable and searchable. And you can even expose them as part of the knowledge base for the LLM to reference when needed going forward.
> A recent LinkedIn post that I came across as an example of people trusting (or learning to trust) AI too much while not realizing that it can make up numbers too
Honestly, people make them up just as much or generate equally incorrect graphs.
It's about time our trust into random visualizations is destroyed, without the actual formulas and data behind being exposed.
It's a mismatch with our intuition about how much effort things take.
If there's humans involved, "I took this data and made a really fancy interactive chart" means that you put a lot more work into it, and you can probably somewhat assume that this means some more effort was also put into the accuracy of the data.
But with the LLM it's not really very much more work to get the fancy chart. So the thing that was a signifier of effort is now misleading us into trusting data that got no extra effort.
(Humans have been exploiting this tendency to trust fancy graphics forever, of course.)
It is not limited to graphics, better packaged products, better dressed / good looking well spoken person and so on. Celebrity endorsements depend on this thesis.
There has always been a bias towards form over function.
This is the tension I keep hitting when building data tools on top of LLMs. A nice-looking chart makes the output feel more trustworthy, but the data can still be wrong. The chart just makes it harder to notice. LLMs still need to come with receipts of where the data came from and the math they did. It's as bad as "I read the headline so I know everything in the article."
Wow, I asked it to build me a simple diagram explaining agile development and it did an amazing job. Wow it felt magical to watch that diagram slowly animating to life.
Like a much prettier version of Mermaid.
Kudos, Anthropic. Geez, this is so nice.
Now I'm going to ask it to draw a diagram of a pelican riding a bicycle, why not?
This is pretty neat and I am experimenting with it now, but hasn't ChatGPT had capability to create graphs and interact with data for a while? "ChatGPT advanced data analysis" for example. I'm asking in good faith as maybe some of you have been using that and can compare the two and give an informed opinion.
I usually use a lot of other tools for data analysis or write code with Claude code or another LLM to do data analysis and visualization.
> but hasn't ChatGPT had capability to create graphs and interact with data for a while?
It's pretty bad (for me). I have to use extremely prescriptive language to tell ChatGPT what to create. Even down to the colours in the chart, because otherwise it puts black font on black background (for example). Then I have to specifically tell it to put it in a canvas, and make it interactive, and make it executable in the canvas. Then if I'm lucky I have to hit a "preview" button in the top right and hope it works (it doesn't). I could write several paragraphs telling it to do something like what Claude just demo'd and it wouldn't come close. I'm trying Claude now for financial insights and it's effortless with beautiful UX.
For posterity, Gemini is pretty good with these interactive canvases. Not nearly as good, but FAR better than ChatGPT.
Interesting. So if I'm reading this correctly, this is distinctly different than the artifacts that Claude creates? If that's the case, why create it inline as opposed to an artifact? Any time I get a visual, I tend to find them so useful that I _want_ them to be an artifact that I can export and share.
I tried the periodic table in their examples using sonnet 4.6 on the $20/mo plan. After a few minutes Claude told me it reached the max message length and bailed. I pressed continue and eventually it generated the table, but it wasn't inline, it was a jsx artifact, and I've now hit my daily usage limit.
I’m intermittently getting artifacts vs the new visuals api, depending on which version of the Claude app I use. iOS/iPadOS apps are not yet supporting the visualization API, and I don’t see an app-store update yet.
Unable to reproduce the recipe image in the iOS app. It first gave a normal text answer. Then when referencing this blog post it produced a wonky HTML artifact.
I asked it to do some portfolio analysis for me and it created BEAUTIFUL, tabbed, interactive charts UNPROMPTED. This is kind of magical. The charts were not just beautiful, but actually super useful in understanding the data faster. I honestly could not have produced those in a week if you asked me to.
Likewise, it created a couple visualizations for me that weren't requested but were very useful. That's a nice surprise. My takeaway was essentially the same in that I'd take much longer to make something comparable. I'll take advantage of this quite a bit, I think. I spend a lot of time visualizing data
Reliability has been the real bottleneck for multi-agent setups in production. The hard part isnt getting one agent to do something clever once - its making repeated runs observable and bounded when tools fail halfway through. Idempotency checks, explicit handoff state, and human review gates have mattered more for us than adding another model or another agent role.
The artifact output model is more useful than it looks at first. We use Claude in a multi-agent pipeline and discovered that structured artifact outputs reduce parse errors significantly compared to freeform text responses -- the model seems to reason differently when it knows the output will be rendered. Curious whether Anthropic sees similar quality improvements in tool-use tasks when the output has a concrete format constraint.
When I ask chatgpt to create a mermaid diagram for me it regularly will add new lines to certain labels that will break the parse. If you then feed the parse error back to it the second version is always correct And it seems to exactly know the problem. There are some other examples where it will almost always get it wrong the first time but right if nudged to correct itself. I wonder what the underlying cause is
Today I asked Claude to create me a squidward looking out the window meme and it started generating HTML & CSS to draw squidward in a style best described as "4 year old preschooler". Not quite it yet.
The issue for Claude is that Anthropic don't have an imagen that I know of, so the only tool available for the LLM to draw something is to start doing vector stuff in CSS, which is very hard for it (see the pelicans).
Gemini, ChatGPT or Grok would find this a lot easier as they could gen an image inline, although IP restrictions might bite you. Even Grok wants to lecture on IP these days, but at least it's fairly trivial to jailbreak.
It responds with the statistically most probable text based on its training data, which happens to be different with the errors vs without. I suspect high-fidelity diagramming requires a different attention architecture from the common ones used in sentence-optimized models.
Mermaid is really bad about cutting off text after spaces, so you have to insert <br>s everywhere. I’m guessing this is getting rendered instead of escaped by your interface. Or just lost in translation at the tokenizer.
Observed from 5.2, on chatgpt.com. earlier versions did worse.. as in, they might take a few prompts to generate a parseable syntax. Newer versions just usually deliver one unparseable version then get it right second try. Likely I could prompt engineer to one shot but I think I would always need the specific warning about newlines.
I don't think it's about repeating the instructions, but rather providing feedback as to why it's not working.
I've noticed the same thing when creating an agentic loop, if the model outputs a syntax error, just automatically feed it back to the LLM and give it a second chance. It dramatically increases the success rate.
This is one of the issues I’ve attempted to tackle with the Mermaid Studio plugin for IntelliJ.
It provides both syntax guides and syntax/semantic analysis as MCP Tools, so you can have an agent iteratively refine diagrams with good context for patterns like multi-line text and comments (LLMs love end-of-line comments, but Mermaid.js often doesn’t).
I think the problem should be defined as "why does it not loop back the errors from the first attempt so it can fix it on the second attempt" rather than why it fails to produce a fully correct implementation on the first pass.
you’ve got to give it a way (eg rendering with playwright and friends) and tell it to use that way to verify correctness. it’s not going to create the guard rail for you but if you provide it with one the output is much better.
Chat --> Notebook: Jupyter is so much more functional than slack for communicating real work product!
Next up: exporting or sharing selections from the chat as a document or interactive page. If they allow share with non-subscribers, subscriptions could hockey stick -- particularly if the document/page included prompts necessary to replicate (or modify and adapt).
Claude can do mermaid diagrams and I started with those, but I have been asking it to generate draw.io diagrams as of late. I haven't actually tried the AI integration recommendations for draw.io, yet. I will have to pull the skill and references to see if it makes the process faster.
Honestly, I initially thought that everyone already does it, amazing it seems they don't yet - neither teachers, nor class. The artefact was created with care and love through a very long conversation, so this is not a 1-shot slop, rather a cared-for-slop :D. Besides I don't think it is easy to get this right from the first time, and the model usually expounds on the irrelevant details if not properly guided by a human hand.
People are really amazed at something excel has been capable of for like 30 years now. PowerBI/Tableau are free and do much more. Making charts and diagrams is impressive or a selling point now? Am I missing something here?
fixxation92 | a day ago
Razengan | a day ago
But you can Sign in with Google.
If you signed up with your Apple on the iOS Claude app, to access your account on the computer, you have to open the passwords app and copy your random email address and paste it into the Claude website login.
Also if you try to copy-paste a prompt from Notes etc into the Claude chat, it gets added as an attachment, so you can't edit the prompt. If you do the four-finger shortcut to paste it as text, it mangles newlines etc.
Why are they so dumb about such basic UX for so long?
bombela | a day ago
catgirlinspace | a day ago
radley | a day ago
Apple forces developers to offer Sign in with Apple on iOS devices if any other sign in service is used. Apple can't force them to do it on non-Apple platforms.
Wowfunhappy | a day ago
Isn't this basically Apple's fault? When you signed up, Apple provided a fake email address in leu of your real one. This is great for privacy but means the service has the wrong email.
I'm sure they didn't want to provide an Apple sign in option at all, but it's required by App Store rules.
Razengan | a day ago
But they wanted to provide a Google Sign In? wth?
> This is great for privacy but means the service has the wrong email.
So harm the users to benefit the service? wtf?
I don't want to give my real email or anything to random services, specially not one like Claude where they don't even let you remove your payment info.
Wowfunhappy | a day ago
The original complaint was:
>> If you signed up with your Apple on the iOS Claude app, to access your account on the computer, you have to open the passwords app and copy your random email address and paste it into the Claude website login.
Either you use your original email or you use a per-service email. Apple helps you do the latter, but this does come with UX tradeoffs.
Using a per-service email, then complaining that the service does not have your real email, strikes me as misguided.
Razengan | 22 hours ago
Only when a dumb service refuses to support Sign In with one pro-privacy provider but does for another anti-privacy one.
Anyway I've voted by having a ChatGPT/Codex subscription for 1 year and only tried Claude for 1 month. Not missing anything.
nerdjon | a day ago
Apple should not have had to require developers to have options other than Google for authentication, but clearly some companies have to be dragged kicking and screaming.
So clearly they support it, and there is no reason it should not work on the web also.
j45 | a day ago
Always best to sign in with your own email address.
nerdjon | a day ago
They don't have to bend for another, but they made a choice to put an app on iOS. They added support for apple signin, and then for some reason did not put it on their website.
You can criticize Apple for requiring that all you want, but they clearly have support for it and are choosing to not put it on their website which is causing a worse user experience.
IF apple did not support website loggin than sure, but they do. So the ability to fix this is on Anthropic (and many other websites).
If you are already going to support third party login you should not limit it to only Google accounts and there is no reason to support Apple on iOS and not the web.
Also for the record, Apple only requires sign in with apple if you already support third party authentication. So if you are already going to support that, giving the user more choice (and making it so we are all a bit less dependent on google) is a good thing.
j45 | a day ago
Third party logins are an extension and a massive risk to any website that doesn't include email hosting.
We have see identity providers dissapear, and people may change their mind.
Easiest way is to register you rown domain and use it with an identity provider of your choice and be able to move it anywhere.
Otherwise we are a faceless citizen of a corporation that can handle access to our identity and everything attached to it without recourse or access to anyone.
Razengan | 22 hours ago
Are you seriously trying to justify offering Sign in with Google but not ALSO offer Sign in with Apple because of some contorted principle, the method which HELPS users maintain their privacy? What the actual f.
Antrhopic's UX is just trash, the worst of all the major AI products.
They have this "I'm special" syndrome where they think they can get away with doing shit weirdly and not offer basic features that everyone else does, and the reason why I never purchased any of their services again after the first month, and had to replace my payment info with a throwaway card because they wouldn't let me remove it, again unlike everyone else.
Wowfunhappy | 22 hours ago
j45 | 8 hours ago
It's ok that Anthropic wasn't a fit for your prompting preference, it doesn't have to work for everyone, and it doesn't mean it wont' work for others. LLMs in general have proven that trying it once a few months ago can be a great way to miss changes. There's something out there for everyone.
squeaky-clean | a day ago
Using a randomly generated email per service is a huge improvement over always using the same email.
Razengan | 22 hours ago
Oh boy
Saying this in 2026 is just.. oh man. just wow
tstrimple | a day ago
Wowfunhappy | a day ago
My original thinking was that Apple makes it too easy for a general audience to hide their email without considering the implications (the service won't know your email). But of course there's a tension here, since you also want the option to be easy and accessible.
The party I do not consider at fault in this case is Anthropic.
IcyWindows | 20 hours ago
short_sells_poo | a day ago
mikkupikku | a day ago
ericmcer | a day ago
dheera | a day ago
brcmthrowaway | a day ago
throwatdem12311 | a day ago
They write 100% of their code with Claude. Some of their engineers apparently burn over 100k worth of tokens per month.
It’s not surprising they ship fast at all when the product is actually falling apart at the seams and they just vibe code everything.
mchusma | a day ago
"If brute force doesn't work, you aren't using enough of it." - Isaac Arthur
throwatdem12311 | a day ago
In the meantime I can’t even continue a Claude Code session I started on desktop on my phone. What’s the point of shipping a billion features of they are all half baked?
WhrRTheBaboons | 23 hours ago
simosmik | 23 hours ago
fwiw I’ve had this open source browser ui that sits on top of your claude code, gemini and codex and picks up/starts your sessions from any device https://github.com/siteboon/claudecodeui
bmurphy1976 | 23 hours ago
Example: we are putting a lot of energy into removing technical debt, reorganizing the code to remove unneeded abstraction and complexity, and creating missing tests and automation. We're not just burping out new untested and poorly reviewed functionality.
davesque | 23 hours ago
not_ai | 23 hours ago
If I wasn’t forced to use it I wouldn’t as there are better options available.
fixxation92 | 23 hours ago
bmurphy1976 | 23 hours ago
asim | a day ago
joshribakoff | a day ago
bogzz | a day ago
abnercoimbre | a day ago
elliotbnvl | a day ago
ericmcer | a day ago
shiftyck | a day ago
gkfasdfasdf | a day ago
[0]: https://github.com/vercel-labs/json-render
[1]: https://github.com/thesysdev/openui
gavinray | a day ago
atonse | a day ago
alansaber | a day ago
captainbland | a day ago
nerdjon | a day ago
I mean is it really that shocking that you can have an LLM generate structured data and shove that into a visualizer? The concern is if is reliable, which we know it isnt.
j45 | a day ago
Passive questions generate passive responses.
ericmcer | a day ago
elliotbnvl | a day ago
I'm finding more and more often the limiting factor isn't the LLM, it's my intuition. This goes a way towards helping with that.
vunderba | a day ago
Angostura | a day ago
People find them quite easy to check - easier than the raw document. My angle with teams is use these to check your processes. If the flow is wrong it’s either because the LLM has screwed up, or because the policy is wrong/badly written. It’s usually the latter. It’s a good way to fix SOPs
vunderba | 23 hours ago
I still review each diagram afterward, but the great thing is that, unlike image-based diagrams, they remain fully text-readable and searchable. And you can even expose them as part of the knowledge base for the LLM to reference when needed going forward.
programmertote | a day ago
P.S. Credit to the poster, she posted a correction note when someone caught the issue: https://www.linkedin.com/posts/mariamartin1728_correction-on...
Avamander | 22 hours ago
Honestly, people make them up just as much or generate equally incorrect graphs.
It's about time our trust into random visualizations is destroyed, without the actual formulas and data behind being exposed.
mikkupikku | a day ago
kemayo | a day ago
If there's humans involved, "I took this data and made a really fancy interactive chart" means that you put a lot more work into it, and you can probably somewhat assume that this means some more effort was also put into the accuracy of the data.
But with the LLM it's not really very much more work to get the fancy chart. So the thing that was a signifier of effort is now misleading us into trusting data that got no extra effort.
(Humans have been exploiting this tendency to trust fancy graphics forever, of course.)
manquer | 23 hours ago
There has always been a bias towards form over function.
ipython | a day ago
https://www.reddit.com/r/dataisugly/comments/1mk5wdb/this_ch...
outlore | 23 hours ago
jameschaearley | 21 hours ago
amadeuspagel | 8 hours ago
atonse | a day ago
Like a much prettier version of Mermaid.
Kudos, Anthropic. Geez, this is so nice.
Now I'm going to ask it to draw a diagram of a pelican riding a bicycle, why not?
drewda | a day ago
Great for summarizing a multi-step process and quick to render with simple tools.
JoshGG | a day ago
I usually use a lot of other tools for data analysis or write code with Claude code or another LLM to do data analysis and visualization.
article about the ChatGPT charts and graphs https://www.zdnet.com/article/how-to-use-chatgpt-to-make-cha...
Gareth321 | a day ago
It's pretty bad (for me). I have to use extremely prescriptive language to tell ChatGPT what to create. Even down to the colours in the chart, because otherwise it puts black font on black background (for example). Then I have to specifically tell it to put it in a canvas, and make it interactive, and make it executable in the canvas. Then if I'm lucky I have to hit a "preview" button in the top right and hope it works (it doesn't). I could write several paragraphs telling it to do something like what Claude just demo'd and it wouldn't come close. I'm trying Claude now for financial insights and it's effortless with beautiful UX.
For posterity, Gemini is pretty good with these interactive canvases. Not nearly as good, but FAR better than ChatGPT.
HotGarbage | a day ago
razerbeans | a day ago
I_am_tiberius | a day ago
I_am_tiberius | a day ago
czk | a day ago
data-ottawa | a day ago
karussell | 23 hours ago
edit: claude just confirmed the initial version has a bug and 104-117 are not visible
googlehater | 18 hours ago
Rendered it in the right pane, instead of inline. Dark theme. 2% of Daily limit.
jzig | a day ago
mehdibl | a day ago
Gareth321 | a day ago
steve_adams_86 | 12 hours ago
wuweiaxin | a day ago
alansaber | a day ago
wuweiaxin | a day ago
darepublic | a day ago
stefan_ | a day ago
qingcharles | 23 hours ago
Gemini, ChatGPT or Grok would find this a lot easier as they could gen an image inline, although IP restrictions might bite you. Even Grok wants to lecture on IP these days, but at least it's fairly trivial to jailbreak.
overfeed | a day ago
It responds with the statistically most probable text based on its training data, which happens to be different with the errors vs without. I suspect high-fidelity diagramming requires a different attention architecture from the common ones used in sentence-optimized models.
deckar01 | a day ago
ar0b | a day ago
What instance of ChatGPT are you doing that with? (Reasoning?)
darepublic | a day ago
alex_duf | 22 hours ago
I've noticed the same thing when creating an agentic loop, if the model outputs a syntax error, just automatically feed it back to the LLM and give it a second chance. It dramatically increases the success rate.
quintu5 | 23 hours ago
It provides both syntax guides and syntax/semantic analysis as MCP Tools, so you can have an agent iteratively refine diagrams with good context for patterns like multi-line text and comments (LLMs love end-of-line comments, but Mermaid.js often doesn’t).
dworks | 23 hours ago
karmajunkie | 16 hours ago
groby_b | a day ago
(Literally nobody needs an image of a cake when asking for a cake recipe)
alansaber | a day ago
groby_b | a day ago
johsole | a day ago
smusamashah | 23 hours ago
https://petergpt.github.io/bullshit-benchmark/viewer/index.v...
w10-1 | 22 hours ago
Next up: exporting or sharing selections from the chat as a document or interactive page. If they allow share with non-subscribers, subscriptions could hockey stick -- particularly if the document/page included prompts necessary to replicate (or modify and adapt).
tamimio | 22 hours ago
mleo | 21 hours ago
skybrian | 19 hours ago
pks016 | 15 hours ago
DeepYogurt | 14 hours ago
https://www.wolfram.com/cdf/
larodi | 11 hours ago
https://claude.ai/public/artifacts/1bded4db-c4c2-4089-aa36-5...
Honestly, I initially thought that everyone already does it, amazing it seems they don't yet - neither teachers, nor class. The artefact was created with care and love through a very long conversation, so this is not a 1-shot slop, rather a cared-for-slop :D. Besides I don't think it is easy to get this right from the first time, and the model usually expounds on the irrelevant details if not properly guided by a human hand.
kate23_human | 5 hours ago
sieep | 4 hours ago