![]() |
|
![]() |
| GPUs are easily and irreparably broken by overheating, so GPU compute is something that's high maintenance. It won't stick around like a building or a bridge. |
![]() |
| I don’t think you’ve used LLMs enough. They are revolutionary, every day. As a coder I’m several times more productive than I was before, especially when trying to learn some new library or language. |
![]() |
| Developer productivity doesn't map very directly to compensation. If one engineer is 10x as productive as another, they're lucky if they get 2x the compensation. |
![]() |
| No but I can work 2-3 hours a day (WFH) while delivering results that my boss is very happy with. I would prefer to be paid 3 times as much and working 8 hours a day, but I'm ok with this too. |
![]() |
| My experience is similar. Most of the results are not really useful so I have to put work in to fix them. But at that point I can do the small extra step of doing it completely myself. |
![]() |
| > I do think those juicy salaries and compensation packages will.
I think that's inevitable with or without LLMs in the mix. I also think the industry as a whole will be better for it. |
![]() |
| "no no, it's not that the technology has reached it's current limits and these companies are bleeding money, they're just withholding their newest releases not to spook the normies!" |
![]() |
| Here's an experiment you can try, go to https://www.udio.com/home and grab a free account which comes with more than enough credits to do this. Use a free chat LLM like Claude 3.5 Sonnet or ChatGPT 4o to workshop some lyrics that you like, just try a few generations and ask it to rewrite parts you don't like until you have something that you don't find too cringe. Then go back over to Udio, go to the create tab turn on the Manual Mode toggle and type in only 3 or 4 comma separated tags that describe the genre you like keep them very basic like Progressive Rock, Hip Hop, 1995, Male Vocalist or whatever you don't need to combine genres these are just examples of tags. Then under the Lyrics section choose Custom and paste in just the chorus or a single verse from the lyrics you generated and then click Create. It'll create two samples for you to listen to, if you don't like either of them then just click Create again to get another two but normally it doesn't take too many tries to get something that sounds pretty good. After you have one you like then click on the ... menu next to the song title and click Extend, you can add sections before or after and you just have to add the corresponding verse from the lyrics you generated or choose Instrumental if you want a guitar solo or something. You'll wind up with something pretty good if you really listen to each sample and choose the best one.
Music generation is one of the easiest ways to "spook the normies" since most people are completely unaware of the current SOTA. Anyone with a good ear and access to these tools can create a listenable song that sounds like it's been professionally produced. Anyone with a good ear and competence with a DAW and these tools can produce a high quality song. Someone who is already a professional can create incredible results in a fraction of the time it would normally take with zero budget. One of the main limitations of generative AI at the moment is the interface, Udio's could certainly be improved but I think they have something good here with the extend feature allowing you to steer the creation. Developing the key UI features that allow you to control the inputs to generative models is an area where huge advancements can be made that can dramatically improve the quality of the generated output. We've only just scratched the surface here and even if the technology has reached its current limits, which I strongly believe it hasn't since there are a lot of things that have been shown to work but haven't been productized yet, we could still see steady month over month improvements based on better tooling built around them alone. Text generation has gone from markov chain babblers to indistinguishable from human written. Image generation has gone from acid trip uncanny valley to photorealistic. Audio generation has gone from 1930's AM radio quality to crystal clear. Video generation is currently in fugue dream state but is rapidly improving. 3D is early stages. ???? is next but I'm guessing it'll be things like CAD STL models, electronic circuits, and other physics based modelling outputs. The ride's not over yet. |
![]() |
| Have you tried the latest model? It's night and day.
Edit: There's obviously still skill involved in creating a good song, it's not like you can just click one button and get a perfect hit. I outlined the simplest process in my first comment and specifically said you could create a "listenable" song, it's not going to be great but it probably rivals some of the slop you often hear on the radio. If you're a skilled music producer you can absolutely create something good especially now with access to the stemmed components of the songs. It's going to be a half manual process where you first generate enough to capture the feeling of the song and then download and make edits or add samples, upload and extend or remix and repeat. If you're looking for links and don't care to peruse the trending section they have several samples on the announcement page https://www.udio.com/blog/introducing-v1-5 |
![]() |
| Your observation is spot on. LLMs represent a transformative capability, offering new ways to handle tasks that were previously more challenging or resource-intensive. |
![]() |
| Why do you need an LLM if you know what you want it to do? Just write the code rather than wrangling with the LLM, it isn't like writing code take much time when you know what it should do. |
![]() |
| Not op but my response: Because I am lazy and would like to save the 1-5 minutes it would take me to actually write it. When there are dozens of these small things a day the saved time really adds. |
![]() |
| I know it’s not a completely fair comparison, but to me this question is kind of missing the point. It’s like asking “Why take a cab if you know where you want to go?” |
![]() |
| It's such a poor comparison it's ridiculous. A better analogy is "why take a cab if you know where you want to go and provide the car and instructions on how to drive" |
![]() |
| I assume it knows the big stuff like the PyTorch API/major JS and React libs then just paste the docs or even impl code for any libs it needs to know beyond that. |
![]() |
| Well, time to run it locally then :) Check out ollama.com. llama 3.1 is pretty crazy, especially if you can run the 405B one. Otherwise, use Mistral/Mixtral or something similar. |
![]() |
| The worst part about LLMs is this attitude it's giving to people who get a few helpful answers in a row
You're like a gambling addict who thinks he's smarter than everyone else |
![]() |
| I was skeptical like you, but recently decided to try it out. I wasn't expecting much, and as such I was slightly surprised.
For example, just now my NAS stopped working because the boot device went offline. So I got to thinking about writing a simple syslog server. I've never looked at the syslog protocol before, and I've never done any low-level TCP/UDP work in C# yet. So I asked ChatGPT to generate some code[1], and while the result is not perfect it's certainly better than nothing, and would save me time to get going. As another example, a friend who's not very technical wanted to make an Arduino circuit to perform some automated experiment. He's dabbled with programing and can modify code, but struggles to get going. Again just for kicks, I asked ChatGPT and it provided a very nice starting point[2]. For exploratory stuff like this, it seems to provide a nice alternative to searching and piecing together the bits. Revolutionary is a quite loaded word, but it's certainly not just a slight improvement on what we had before LLMs and instead feels like a quantum leap. [1]: https://chatgpt.com/share/f4343939-74f1-404d-bfac-b903525f61... (modified, see reply) [2]: https://chatgpt.com/share/fc764e73-f01f-4a7c-ab58-f43da3e077... |
![]() |
| Very interesting. A couple of notes here on the C# version.
Its using the old format where the Program.cs file has an actual class, whereas as of .NET 6 thats not required. You said barebones, but for any real server you would want to use the generic host https://learn.microsoft.com/en-us/dotnet/core/extensions/gen... which gets you a lot of the boilerplate and enables you program to be wrapped in a windows or systemd service. Finally, parsing can be simplified since ASCII is a proper subset of UTF-8, you can just parse the entire string as UTF-8. IMHO I am disappointed that the AI didn't point that out. |
![]() |
| I wonder if quality would improve if one would upload the latest LOVR documentation and uploaded LOVR codebases using the newest version and instruct it properly? |
![]() |
| Yes, you have figured it out. LLMs are terrible for graphics programming. Web development - much better. Sonnet 3.5 is the only good model around for now. GPT 4o is very poor. |
![]() |
| what did it do when you told it all of those things? was it able to fix the problems when you pointed them out? did you give it one prompt and expect perfect code out on the first try? is that how you code? all your code complies and runs flawlessly first try? I'm jealous. it usually takes me a bunch of passes before I get things right.
here's a chat for a uc and LCD chip that I picked at random (and got the name wrong for) (and didn't want raspberry pi code for so it stopped it short on that response) https://chatgpt.com/share/2004ac32-b08b-43d7-b762-91543d656a... |
![]() |
| one trick is to let it come up with what it wants (lots of functions, no texture), then run the code, give it the errors until that's fixed. then ask it to inline them, then add the texture, etc. |
![]() |
| Not-so-subtly mocking the top-level for not replying "yet", when they replied almost immediately after with a video of the relevant workflow, was not a move that made you look smart or nice. |
![]() |
| By god, andrewflnr. Very "nice". /s. See point 4 below.
You just showed that you are inaccurate, pompous and fake, all of that, in one single comment of yours, above. How? I'll tell you: 1. inaccurate: That commenter's username (the one who started this subthread) is _acco, not acco_ as you wrote above. Check that in their comment, or here: https://news.ycombinator.com/user?id=_acco I was careful to check the spelling before mentioning their name, unlike you, even when I referred to them earlier. The fact that you cannot even get the position of an underscore in a name, correct, seems to indicate that you are sloppy. which leads me to my next point. 2. pompous: You said: >Really, I'm being generous in assuming they didn't reply before you. This is the pompous bit. Generous? Laughable. I neither need not want your generosity. If anything, I prefer objectivity, and that people give others the benefit of the doubt, instead of assuming bad intentions by them: I had actually checked for a second comment by _acco (in reply to layer8) just before I wrote my comment to layer8, the one that got all of you in a tizzy. But you not only got the times wrong (see your edit, and point 3 below), but also assumed bad faith on my part. 3. fake. You first said above that both those replies to layer8 showed as 13 hours ago, then edited your comment to say 14 and 13 hours. It shows that you don't use your brains. The feature of software showing time deltas in the form of "hours ago" or "days ago", versus an exact time stamp, is pretty old by now. It dates back to Web 2.0 or earlier, maybe it was started by some Web 2 startups or by Google. If you think you are so clever as to criticize me without proof, or say that you are generous in your assumptions about me, you should have been equally clever or generous about the time delta point above, and so realized that I could have replied to layer8 before _acco, which was indeed the case. Obviously I cannot prove it, but the fact that I got _acco's name correct, while you did not, lends credence to my statement. It shows that I took care while writing my comment. 4. So you are fake because you don't bother to think before bad-mouthing others, and even more fake because you did not apply (to yourself) your own made-up "rule" in this other comment of yours, where you criticized my comment as being neither smart nor nice, so not of value: https://news.ycombinator.com/item?id=41310460 I should not have had to write such a long comment to refute your silly and false allegations, and I will not always do that, but decided to do it this time, to make the point. And, wow: you managed to pack 3 blunders (being inaccurate, pompous and fake) into a comment of just a few lines. That's neither smart not nice. Instead, it's toxic. |
![]() |
| 2 hours in a discussion forum, where the discussion spans days or sometimes weeks is certainly an ”almost immediate” response.
Perception of time is subjective. |
![]() |
| If a comment is not at least one of smart or nice, it's a waste of space and attention. That may not be your purpose, but don't act shocked when people respond with negativity. |
![]() |
| In my experience ChatGPT seems particularly bad at Elixir, presumably because there is a comparative lack of published code and discussion about it. |
![]() |
| Have you tried using the shortcut for turning copilot on/off? I know what you mean and in those cases I just turn it off for a second and type freely, then turn it back on. |
![]() |
| It's just a requirement that everyone has a license and did 1-hour introductory training. Whether you actually use it or not is up to you, but it's encouraged. |
![]() |
| My GH copilot kept recommending incorrect things when I use it along side common libraries and frameworks, I don't know, I just don't really find it very useful. |
![]() |
| I think the non obvious benefit is using LLMs nudge you into putting your thoughts in narrative form and training that ability, something that someone with more experience does subconsciously. |
![]() |
| > BigCos won't let their employees use these tools to write code because you can't let code leak off prem.
Ridiculous blanket statement. A bunch of places use external LLMs. |
![]() |
| As I mentioned in a sibling comment, I now "pair program" all day. Instead of being the driver and navigator all day, I can mostly sit "one layer up" in the navigator seat. |
![]() |
| How is this a change in job description? It may be a failure of imagination on my part, but it sounds like you're still doing what you have always done - you've changed the how. |
![]() |
| I don't. But that's probably because I'm very opinionated about the implementation details, so I'm scrutinizing and tweaking its output a lot. |
But as to the hype, we are in a brief pause before the election where no company wants to release anything that would hit the news cycle in a bad way and cause knee-jerk legislation. Are there new architectures and capabilities waiting? Likely some. Sora showed state of the art video generation, OpenAI has demoed an impressive voice mode, and Anthropic has teased that Opus 3.5 will be even more capable. OpenAI also clearly has some gas in the tank as they have focused on releasing small models such as GPT-4o and 4o mini. And many have been musing about agents and methods to improve system 2 like reasoning.
So while there’s a soft moratorium on showing scary new capability there is still evidence of progress being made behind-the-scenes. But what will a state of the art model look like when all of these techniques have been scaled up on brand new exascale data centers?
It might not be AGI, but I think it will at least be enough for the next hype Investment bubble.