![]() |
|
![]() |
| Or it could be the start of the enshittification of Anthropic, like OpenAI ruined GPT-4 with GPT-4o by overly simplifying it.
I hope not, because Claude is much better, especially at programming. |
![]() |
| Claude 3.5 Sonnet is the first model that made me realize that the era of AI-aided programming is here. Its ability to generate and modify large amounts of correct code - across multiple files/modules - in one response beats anything I've tried before. Integrating that with specialized editors (like https://www.cursor.com) is an early vision of the future of software development.
|
![]() |
| This sentiment is so far from the truth that I find it hilarious. How can a technically adept person be so out of touch with what these systems are already capable of? |
![]() |
| With all of that, ChatGPT is actually one of the top authors in Amazon e-books.
But I agree that for some creative tasks, like writing or explaining a joke, or some novel algorithms, it's very bad. |
![]() |
| I'm a firm believer that the best benchmark is playing around with the model for like an hour. On the type of tasks that are relevant to you and your work, of course. |
![]() |
| Who cares about the interface? Not everyone is interested in conversational tasks. Corporations in particular need LLMs to process their data. A restful API is more than enough. |
![]() |
| > Except a "good product" is safe.
Depends on how you define "safe". The kind of "safe" we get from OpenAI today seems to be mostly censorship, I don't think we need more of that. |
![]() |
| Even when we get a gen AI that exceeds all human metrics, there will 100% still be people who with a straight face will say "Meh, I tried it and found it be pretty useless for my work." |
![]() |
| I have, yeah.
Still useless for my day to day coding work. Most useful for whipping up a quick bash or Python script that does some simple looping and file io. |
![]() |
| Summarizing, doc qa, and unstructured text ingestion are the killer features I’ve seen.
The 3rd one still being quite involved, but leaps and bounds easier than 5 years ago. |
![]() |
| Re: NVIDIA. I wholeheartedly agree. Google/TPU is an existence proof that it is entirely possible and rational to do so. My surprise was that everyone except Google missed. |
![]() |
| I think you are in one of the extreme bubbles. The general tech industry is not subscribed to the drama and has less personal feelings on individuals they do not directly know. |
![]() |
| Well as mentioned I don't even see Google's ads unless I deliberately turn the blocker off. I much prefer that to the content being subtly biased which you see in blogs, newspapers and the like. |
![]() |
| Kinda easy if you look where the stuff is being trained. A single joke post on Reddit was enough to convince Google's A"I" to put glue on pizza after all [1].
Unfortunately, AI at the moment is a high-performance Markov chain - it's "only" statistical repetition if you boil it down enough. An actual intelligence would be able to cross-check information against its existing data store and thus recognize during ingestion that it is being fed bad data, and that is why training data selection is so important. Unfortunately, the tech status quo is nowhere near that capability, hence all the AI companies slurping up as much data as they can, in the hope that "outlier opinions" are simply smothered statistically. [1] https://www.businessinsider.com/google-ai-glue-pizza-i-tried... |
![]() |
| > Quite a lot of humans are bad at that too. It's not so much that AIs are markov chains but that you really want better than average human fact checking.
Let's take a particularly ridiculous piece of news: Beatrix von Storch, a MP of the far-right German AfD party, claimed a few years ago that the sun's activity (changes) were responsible for climate change [1]. Due to the sheer ridiculousness of that claim, it was widely reported on credible news sites, so basically prime material for any AI training dataset. A human can easily see from context and their general knowledge: this is an AfD politician, her claims are completely and utterly ridiculous, it's not the first time she has spread outright bullshit and it's widely accepted scientific fact that climate change is caused by humans, not by sun activity changes. An AI at ingestion time "knows" neither of these four facts, so how can it take that claim of knowledge and store it in its database as "untrustworthy, do not use in answers about climate change" and as "if someone asks about counterfactual claims relating to climate change, show this"? [1] https://www.tagesschau.de/faktenfinder/weidel-klimawandel-10... |
![]() |
| "I noticed your desire to be ad-free, but puppies die all the time. If you want to learn more about dog mortality rates, you can subscribe to National Geographic by clicking this [link]". |
![]() |
| And then the new "adblockers" will be AI based too, and will take the AI's answer as input and remove all product placement.
It's just a cat and mouse game, really |
![]() |
| If you can solve the technical problem of ensuring an AI acts on behalf of its user's interests, please post the solution on the AI Alignment Forum: https://www.alignmentforum.org/
So far, that is not a feature of existing or hypothesized AI systems, and it's a pretty important feature to add before AI exceeds human capabilities in full generality. |
![]() |
| AI is a commodity right now, or at least - text. I just realized when paying the bills this month I got 1kg of cucumbers and a few KBs of text from opanai. They literally sell text by the kilo. |
![]() |
| AI (of the type that OpenAI is doing) already is a commodity. right now.
So the question would be "what makes you think AI will stop being a commodity?". |
![]() |
| Search needs to constantly update its catalog. I‘d say there are lots of AI use-cases that will (eventually?) be good for a long while after training. Like audio input/output, translations, … |
![]() |
| I'm betting on fully integrated agents.
And for good agents you need a lot of crucial integrations like email, banking etc. that can only provide companies like Google, Microsoft, Apple etc. |
![]() |
| > The "AI answer" industry might become 10 times bigger than the search industry.
Whenever I see people saying things like this it just makes me think we are at, or very near, the top. |
![]() |
| Good for him, seems like OpenAI is moving towards a business model of profitability, and Anthropic seems to be more aligned with the original goals of OpenAI.
What is open about Anthropic ? |
![]() |
| It was actually a serious and open question, but I can see, given the hypocrisy found in a lot of these self-proclaimed "open AI" companies, how it would come across like I was refuting something ;) |
![]() |
| Is it just me or is Brockman leaving absolutely huge ? I can’t believe this isn’t front page. Basically everyone who is anyone has left or is leaving. It’s ridiculous. |
![]() |
| It does say that, seems kind of strange though, rapidly growing company, apparently an absolutely key member of facilitating that growth and, poof, gone for 6 months at least. |
![]() |
| Claude 3.5 Sonnet by Anthropic is the best model out there, if you are trying to have an extremely talented programmer paired to you.
Somehow, OpenAI is playing catch with them rather than vice versa. |
![]() |
| I'd replace "extremely talented programmer" with "knowledgeable junior", in my experience. It's much better than GPT-4o, but still not great. |
![]() |
| Advanced ML products are forbidden[0] to export to many places, so those who skimp on KYC are playing with fire. Paid products do not have this issue since you provide a billing address, but there is no good, free, and legal LLM that does not use a reliable way of verifying at least user’s location.
Whether they are serious about it or use it as an excuse to collect more PII (or both/neither), collecting verified phone numbers presumably allows them to demonstrate compliance. [0] https://cset.georgetown.edu/article/dont-forget-the-catch-al... |
![]() |
| I'm not affiliated with Claude, but assuming you're serious:
> Umm... why? https://support.anthropic.com/en/articles/8287232-why-do-i-n... My guess is, these models are incredibly expensive to run, Claude has a fairly generous free tier, and phone numbers are one of the easiest ways to significantly reduce the number of duplicate accounts. > Nobody else in the AI space wants to track my number. Given they're likely hoovering up all of the data you're sending to them, and they have your email address to identify you, this seems like an odd hill to die on. |
OpenAI has a burn rate of about 5 billion a year and they need to raise ASAP. If the fundraising isn't going well or if OpenAI is forced to accept money from questionable investors that would also be a good reason to jump ship.
In situations like these it's good to remember that people are much more likely to take the ethical and principled road when they also stand to gain from that choice. People who put their ideals above pragmatic self-interest self-select out of positions of power and influence. That is likely to be the case here as well.