Hacker Newsnew | past | comments | ask | show | jobs | submit | mrtesthah's commentslogin

Think about all the things people have done in the real world the last 50 years to combat bigotry. During the civil rights movement of the 60s, black people sat at segregated lunch counters and marched peacefully in the street, and were consequently spat on and attacked by white mobs, beaten by police, sprayed with fire hoses, attacked by dogs, etc.

In the last 10 years, the modern black lives matter movement has triggered similar violent backlashes, with every public gathering drawing a militarized police response and hateful counter-protesters. On a policy level, even the most milquetoast corporate initiatives to consider applications and promotions from diverse candidates of equal merit are now being slandered and attacked. In education, acknowledgment of historical racial and gender inequality is under heavy censorship pressure.

It really does seem like the more effective we are at acting IRL, the greater the backlash is going to be.


Coding and reasoning skills can be improved using machine-driven reinforcement learning.

https://arxiv.org/abs/2501.12948


This is a tone deaf take that ignores the massive imbalance in how IP law is wielded by large corporations vs individuals or small firms.

No, it’s the most empowering thing humanity has ever constructed to wrestle the beast of IP and make it an irrelevant footnote in the human story. Deckchairs on the titanic.

If one wastes their life in court, arguing 19th century myths, that’s on the players.


IP law is not going away for “little people” like us until we collectively overturn the existing political regime which grants more rights to corporations than people.

IP being used by small firms instead of large corporations does not make it a good thing. It's the same disgusting concept to deny freedom for end users to give control to who "owns" the IP.

IP as a concept needs to die.


>IP as a concept needs to die.

Yes it does, but if you think that's going to happen on its own by allowing the largest corporations to run roughshod over everyone else, you're going to be disappointed. True freedom takes struggle.


They probably need to be able to read and understand the lean language.

I can read and understand e.g. Python, but I have seen subtle bugs that were hard to spot in code generated by AI. At least the last time I tried coding agents (mid 2025), it was often easier to write the code myself then play "spot the bug" with whatever was generated. I don't know anything about Lean, so I was wondering if there were similar pitfalls here.

In this case the human written statement of the theorem is small. Can there be bugs? Absolutely! But it's essentially a few lines of code worth of thinking.

The lean proof checker then checks to make sure the proof actually proves the statement.

In this case an AI is generating the proof, but if it "compiles" it's correct. The only thing humans need to check is the statement to be proven.

(I don't know anything about this project but I've played around with lean and used other proof checkers more sesrisously).


This is also an issue with non-AI and non-Lean proofs. Andrew Wiles' initial Fermat's Last Theorem proof initially had an error in it. That was spotted by peer review, fixed, and an updated proof was submitted.

I don't disagree with you, but on the other hand, I feel the same way about code written by other humans, and that's not because they're necessarily worse at writing code than me, but because for code I've written myself, I've already spent time thinking about it, so I don't have to start from scratch when re-reading it. It's also not like I don't think I potentially write as many bugs as my coworkers either; it's just easier for me to catch my own up front as I'm coding than it is to catch theirs in code review. The two main differences are that I can have a more meaningful conversation with my coworkers about their approach, what bugs they might think are worth looking out for, etc. compared to an LLM (which in my experience will claim completely incorrect things about the code it wrote far more often than any engineer I've worked with even junior ones; the humans I've worked with have uniformly been able to report how confident they are in what they've produced being what they were tasked with without insane exaggerations), and that an LLM can produce a much higher volume of plausible-enough looking code in a given unit of time than most humans I've worked with. It's not obvious to me that these would be particularly severe problems in generating proofs though; unless the proof is so large that it would be infeasible to read through it in a reasonable amount of time, I would expect mathematicians to be able to make up for the lower quality conversations with the "author" by devoting more time to reading and thinking, or having someone else also read through the proof and talking through it with them. If anything, it's possible that the timeline for writing up a paper about the results might be better for some mathematicians than the equivalent amount of time most engineers have to spend reviewing code before the pressure to get it merged and move on to the next thing. (I'm aware that there is certainly pressure to get things published in academia, but I don't have firsthand experience, so I've tried to be intentional in how I've worded this to clarify that I want to avoid any assumptions about what the norms would be, but given the relatively wide range of time pressure that engineers might experience across the industry as a whole, I'd expect that at least some mathematicians might have some flexibility to spend extra time reading through an LLM-written proof, especially if it might be time they'd otherwise have to spend trying to come up with the proof themselves).

If you want to check the statement, you only have to read the type. The proof itself you don’t have to read at all

As I understand it Lean is not a general purpose programming language, it is a DSL focused on formal logic verification. Bugs in a DSL are generally easier to identify and fix.

It seems one side of this argument desperately needs AI to have failed, and the other side is just saying that it probably worked but it is not as important as presented, that it is actually just a very cool working methodology going forward.


Lean 4 is a bit awkward, but workable as a general purpose programming language, it e.g. supports sockets (with a C module, but so does Python.)

We do know that it’s been a longstanding policy of DHS for officers not to stand in front of cars on purpose just so they’d have an excuse to fire upon the driver. There was an internal audit in 2014 that called out this exact behavior.

Indeed, murder is the point. They don’t actually believe that the agent “feared for his life”. The only disagreement here is about whether it’s ok to murder people who you don’t like.

Do you think an officer who feared for this life would have used a casual stance with one hand on the gun and the other with a phone, then casually walked away, or would he have held the gun with two hands as trained to make sure he hit his mark?

Or maybe your point is simply that because dissenters exist that their critiques are valid? There are also people who think the 2020 election was rigged simply because a loudmouth claims it to be. They’re wrong.


Remember that the entire conversation is literally the query you’re making, so the longer it is the more you’re counting on the rational comprehension abilities of the AI to follow it and determine what is most relevant.

Looks fun! Does it support devcontainers and/or any form of isolation? The app's purpose seems similar to that of Branchbox.

Also the website says it runs on macOS 13+, but the app's Info.plist sets LSMinimumSystemVersion to 14.0.


not yet, just vanilla work-tree, so file isolation. I also poorly manage env isolation through running .agentastic/setup.sh script upon work-tree creation; which can setup .env files, etc.

I wasn't aware of branchbox, i was planning to integrate firecracker in the upcoming version. However Branchbox seems like a great option too.

Which set of container features / isolation matter to you more?


Primarily, I want to run agents with a flag like --dangerously-bypass-approvals-and-sandbox to benefit from newer models' improvements in long-horizon tasks. For that to work, I need solid guarantees it won't rm -rf / my entire system while still allowing full stack access for installing and configuring databases, server daemons, etc. Ultimately that means working within some sort of container and/or VM.

Facilitating long-running tasks makes multiple parallel agents much more compelling.


You can actually use full uBlock origin in Kagi's Orion iOS browser.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: