![]() |
|
![]() |
|
I appreciate it too, and they're of course going to call it "open weights", but I reckon we (the technically informed public) should call it "weights-available".
|
![]() |
|
AI is taking marketshare from search slowly. More and more people will go to the AI to find things and not a search bar. It will be a crisis for Google in 5-10 years.
|
![]() |
|
1. Free rlhf
2. They cookie the hell out of you to breadcrumb your journey around the web. They don't need you to login to get what they need, much like Google |
![]() |
|
This is always the case. But the fact that open models are beating state of the art from 6 months ago is really telling just how little moat there is around AI. |
![]() |
|
>We’re rolling out Meta AI in English in more than a dozen countries outside of the US. Now, people will have access to Meta AI in Australia, Canada, Ghana, Jamaica, Malawi, New Zealand, Nigeria, Pakistan, Singapore, South Africa, Uganda, Zambia and Zimbabwe — and we’re just getting started. https://about.fb.com/news/2024/04/meta-ai-assistant-built-wi... |
![]() |
|
It's trivial to comply with EU privacy regulation if you're not depending on selling customer data. But if you say "It's because of regulations!" I hope you have a source to back that up. |
![]() |
|
"Censored" is the word that you're looking for, and is generally what you see when these models are discussed on Reddit etc. Not to worry - uncensored finetunes will be coming shortly. |
![]() |
|
Public benchmarks are broadly indicative, but devs really should run custom benchmarks on their own use cases. Replicate created a Llama 3 API [0] very quickly. This can be used to run simple benchmarks with promptfoo [1] comparing Llama 3 vs Mixtral, GPT, Claude, and others:
Still testing things but Llama 3 8b is looking pretty good for my set of random programming qs at least.Edit: ollama now supports Llama 3 8b, making it easy to run this eval locally.
[0] https://replicate.com/blog/run-llama-3-with-an-api
|
![]() |
|
Maybe at 1 or 2 bits of quantization! Even the Macs with the most unified RAM are maxxed out with much smaller models than 405b (especially since it's a dense model and not a MOE).
|
![]() |
|
Depends on your size threshhold. For anything beyond 100 bn in market cap certainly. There is some relatively large companies with a similar flair though, like Cohere and obviously Mistral.
|
![]() |
|
Counter counter point: apples hardware division has been doing great work in the last 5 years, it’s their software that seems to have gone off the rails (in my opinion).
|
![]() |
|
Good thing that he's only 39 years old and seems more energetic than ever to run his company. Having a passionate founder is, imo, a big advantage for Meta compared to other big tech companies.
|
![]() |
|
Love how everyone is romanticizing his engineering mindset. But have we already forgotten that he was even more passionate about the metaverse which, as far as I can tell, was a 50B failure?
|
![]() |
|
Think of it as a 50B spending spree where he gave that to VR tech out of enthusiasm. Even I, with the cold dark heart that I have, has to admit he's a geek hero with his open source attitude.
|
![]() |
|
Do you have a source? Here's the license when you request access from Meta for Llama, unless there's something I'm missing? https://ai.meta.com/blog/large-language-model-llama-meta-ai/ EDIT: Looks like they did open up commercial use with version 2 with the explicit restriction to prevent any major competitor to Meta from using Llama, and that any improvements related to Llama can only apply to Llama. So an attempt to expand the scope of usage and adoption of their proprietary model without their main competitors being able to use it, which still fits my original point. |
![]() |
|
> It's the same as "lowering prices to the benefit of consumer" vs "price dumping to become a monopoly". Where has that ever worked? Predatory pricing is highly unlikely. See eg https://www.econlib.org/library/Columns/y2017/Hendersonpreda... and https://www.econlib.org/archives/2014/03/public_schoolin.htm... > Facebook never did it at scale though. Google did. Please provide some examples. > There's a difference between "paying higher salaries in fair competition for talents" and "buying people to let them rot to make sure they don't work for competition". It's up to the workers themselves to decide whether that's a good deal. And I'm not sure why as a worker you would decide to rot? If someone pays me a lot to put in a token effort, just so I don't work for the competition, I might happily take that over and practice my trumpet playing while 'working from home'. I can also take that offer and shop it around. Perhaps someone else has actual interesting work, and comparable pay. |
![]() |
|
https://en.wikipedia.org/wiki/Predatory_pricing says > For a period of time, the prices are set unrealistically low to ensure competitors are unable to effectively compete with the dominant firm without making substantial loss. The aim is to force existing or potential competitors within the industry to abandon the market so that the dominant firm may establish a stronger market position and create further barriers to entry.[2] Once competition has been driven from the market, consumers are forced into a monopolistic market where the dominant firm can safely increase prices to recoup its losses.[3] What you are describing is not predatory pricing, that's a big part of why I was confused. > Funnily enough the second author got a good example but still failed to see it under his nose: public schools do have 90% of the market, and in many countries almost 100%. Obviously it works. Please consider reading the article more carefully. Your interpretation requires the author to be an idiot. --- What you are describing about browsers is interesting. But it's more like bundling and cross subsidies. Neither Microsoft nor Google were ever considering making money from raising the price of their browser after competition had been driven out. That's required for predatory pricing. |
![]() |
|
I am fine with large pool of greedy people trying their hand at programming. Some of them will stick and find meaning in work. Rest will wade out in downturn. Net positive.
|
![]() |
|
> but driving software engineer salaries out of reach of otherwise profitable, sustainable businesses is not a good thing. I'm not convinced he's actually done that. Pretty much any 'profitable, sustainable business' can afford software developers. Software developers are paid pretty decently, but (grabbing a couple of lists off of Google) it looks like there's 18 careers more lucrative than it (from a wage perspective), and computers-in-general are only 3 of the top 25 highest paying careers - https://money.usnews.com/careers/best-jobs/rankings/best-pay... Medical, Legal, Finance, and Sales as careers (roughly in that order) all seem to pay more on average. |
![]() |
|
They're commoditizing their complement [0][1], inasmuch as LLMs are a complement of social media and advertising (which I think they are). They've made it harder for competitors like Google or TikTok to compete with Meta on the basis of "we have a super secret proprietary AI that no one else has that's leagues better than anything else". If everyone has access to a high quality AI (perhaps not the world's best, but competitive), then no one -- including their competitors -- has a competitive advantage from having exclusive access to high quality AI. [0]: https://www.joelonsoftware.com/2002/06/12/strategy-letter-v/ |
![]() |
|
He went into the details of how he thinks about open sourcing weights for Llama responding to a question from an analyst in one of the earnings call last year after Llama release. I had made a post on Reddit with some details. https://www.reddit.com/r/MachineLearning/s/GK57eB2qiz Some noteworthy quotes that signal the thought process at Meta FAIR and more broadly * We’re just playing a different game on the infrastructure than companies like Google or Microsoft or Amazon * We would aspire to and hope to make even more open than that. So, we’ll need to figure out a way to do that. * ...lead us to do more work in terms of open sourcing, some of the lower level models and tools * Open sourcing low level tools make the way we run all this infrastructure more efficient over time. * On PyTorch: It’s generally been very valuable for us to provide that because now all of the best developers across the industry are using tools that we’re also using internally. * I would expect us to be pushing and helping to build out an open ecosystem. |
![]() |
|
It’s a shame it can’t just be giving back to the community and not questioned. Why is selfishness from companies who’ve benefited from social resources not a surprising event vs the norm. |
![]() |
|
It's because Elon, Mark, and Jensen are true founders. They aren't MBAs who got voted in because shareholders thought they would make them the most money in the shortest amount of time.
|
Since llama is open source, we're going to see fine tunes and LoRAs though, unlike opus.