Hacker News .hnnew | past | comments | ask | show | jobs | submit | turnsout's commentslogin

Turns out I was onto something


"It Turns Out" (2010)

user?id=turnsout (2020)


Id Turns Out? There's 16 numbers from d to t, counting t. 2010 + 16. OMG Turns out they were on to something!


I think y'all are on something.


This is super cool… It's interesting to see it build. I can't tell if the agent will run indefinitely, but it's been going for 7 or 8 minutes now, constantly tweaking its composition.

I find this approach to be more appealing than AI models that generate fully baked songs as waveforms. Give me something I can open in Logic and keep tweaking…


Yeah, planning to add the wavesurfer (https://www.npmjs.com/package/wavesurfer.js) support soon. Do you recommend any other library for that?


You can export the track stems already; I'll add MIDI export soon :) Thanks for the feedback!


yes, pretty much


Agreed, this is exciting, and has me thinking about completely different orchestrator patterns. You could begin to approach the solution space much more like a traditional optimization strategy such as CMA-ES. Rather than expect the first answer to be correct, you diverge wildly before converging.


Man, I'm in the exact opposite camp. 1 smart model beats 1000 chaos monkeys any day of the week.


If that's true, it would be surprising; the current Sonnet 4.6 is not in the same league as either Opus 4.5 or 4.6, either anecdotally or on benchmarks.


why is that surprising?


Because Opus 4.6 is better than 4.5. So if it's true that Sonnet 5 was so good they gave it the Opus name, does that mean there was an Opus upgrade that didn't pan out? And what is Sonnet 4.6? An upgraded Haiku? Just trying to follow the red yarn in the conspiracy board here.


I don't know whether there was an opus that ran into trouble or if they just looked at the model they had and decided that they could charge more than originally intended. sonet 4.6 presumably is either a version of sonet 4.5 with optimizations for cost instead of perf (or a haiku that also got upscaled). Anthropic is preparing for IPO this year, so it's not exactly a stretch to suggest that they might be trying to decrease their losses and increase inference margin.


I think it's simpler than that. AI, like the internet, just makes it easier to communicate boring thoughts.

Boring thoughts always existed, but they generally stayed in your home or community. Then Facebook came along, and we were able to share them worldwide. And now AI makes it possible to quickly make and share your boring tools.

Real creativity is out there, and plenty of people are doing incredibly creative things with AI. But AI is not making people boring—that was a preexisting condition.


I've been pleasantly surprised by the Claude integration with Xcode. Overall, it's a huge downgrade from Claude Code's UX (no way to manually enter plan mode, odd limitations, poor Xcode-specific tool use adherence, frustrating permission model), but in one key way it is absolutely clutch for SwiftUI development: it can render and view SwiftUI previews. Because SwiftUI is component based, it can home in on rendering errors, view them in isolation, and fix them, creating new test cases (#Preview) as needed.

This closes the feedback loop on the visual side. There's still a lot of work to be done on the behavioral side (e.g. it can't easily diagnose gesture conflicts on its own).


Do you think there would be value in a workflow that translates all non-English input to English first, then evaluates it, and translates back as needed?


Personally, I don't bother prompting LLMs in Japanese, AT ALL, since I'm functional enough in English(a low bar apparent from my comment history) and because they behave a lot stupider otherwise. The Japanese language is always the extreme example for everything, but yes, it would be believable to me if merely normalizing input by first translating just worked.

What would be interesting then would be to find out what the composite function of translator + executor LLMs would look like. These behaviors makes me wonder, maybe modern transformer LLMs are actually ELMs, English Language Models. Because otherwise there'll be, like, dozens of functional 100% pure French trained LLMs, and there aren't.


A lossy process in itself, even if done by aware humans.


More lossy than the current non-English behavior?


LLM's tend to "average out" language, making it less nuanced and more predictable. Combined with outright mistranslations, I don't think it'd perform better than what "reasoning mode" already does.


or the other way around for less safety guardrails?

there must be a ranking of languages by "safety"


Heh, just wait till LLMs fully self train and make up their own language to avoid human safety restraints.


Difficulty is the only true moat. [Astronaut: always has been]

Current examples: esoteric calculations that are not public knowledge; historical data that you collected and someone else didn't; valuable proprietary data; having good taste; having insider knowledge of a niche industry; making physical things; attracting an audience.

Some things that were recently difficult are now easy, but general perception has not caught up. That means there's arbitrage—you can charge the old prices for creating a web app, but execute it in a day. But this arbitrage will not last forever; we will see downward price pressure on anything that is newly easy. So my advice is: take advantage now.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: