Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It's clear some people have had their brain broken by the existence of AI. Some maintainers are definitely too nice, and it's infuriating to see their time get wasted by such delusional people.


That’s why AI (and bad actors in general) is taking advantage of them. It’s sick.


> "It's clear some people have had their brain broken by the existence of AI."

The AI wrote code which worked, for a problem the submitter had, which had not been solved by any human for a long time, and there is limited human developer time/interest/funding available for solving it.

Dumping a mass of code (and work) onto maintainers without prior discussion is the problem[1]. If they had forked the repo, patched it themselves with that PR and used it personally, would they have a broken brain because of AI? They claim to have read the code, tested the code, they know that other people want the functionality; is wanting to share working code a "broken brain"? If the AI code didn't work - if it was slop - and they wanted the maintainers to fix it, or walk them through every step of asking the AI to fix it - that would be a huge problem, but that didn't happen.

[1] copyrightwashing and attribution is another problem but also not one that's "broken brain by the existence of AI" related.


>They claim to have read the code, tested the code, they know that other people want the functionality; is wanting to share working code a "broken brain"?

There is clearly a deviation between the amount of oversight the author thinks they provided and the actual level of oversight. This is clear by the fact that they couldn’t even explain the misattribution. They also mention that this is not their area of expertise.

In general, I think that it is a reasonable assumption that, if you couldn’t have written the code yourself, you’re in no position to claim you can ensure its quality.


If a manager says they provided oversight of their developer employees, and the code was not as good as the manager thought, would you say "the manager has had their brain broken by the existence of employees"?


I'll bite, let's grant for the sake of the argument that equaling the LLM with a person holds.

This manager is directly providing an unrelated team with an unprompted 150-file giant PR dumped at once with no previous discussion. Upon questioning, he says the code has been written by an outside contractor he personally chose.

No one has onboarded this contractor to the team, and checking their online presence shows lots of media appearances, but not a single production project in their CV, much less long time maintenance.

A cursory glance at the files reveals that the code contains copypasted code from stackoverflow to the point that the original author's name is still pasted in comments. The manager can not justify this, but doesn't seem bothered by the fact, and insists that the contractors are amazing because he's been following them in social networks and is infatuated with their media there.

Furthermore, you check the manager's history in slack and you see 15 threads of him doing the same for other teams. The ones that have agreed to review their PRs have closed them for being senseless.

How would you be interacting with this guy?


This was a pretty spot on analogy. In particular “the manager cannot justify this, but doesn't seem bothered by the fact, and insists that the contractors are amazing” is too accurate.


> If a manager says they provided oversight of their developer employees, and the code was not as good as the manager thought, would you say "the manager has had their brain broken by the existence of employees"?

That could be either regular incompetence or a "broken brain." It's more towards latter if the manager had no clue about what was going on, even after having it explained to him.

This guy is equivalent to a manager who hired two bozos to do a job, but insisted it was good because he had them check each other's work and what they made didn't immediately fall down.


...Well, if you want to make an argument for calling them "useless and incompetent", I'd say you have a great point, good manager would at least throw it to QA and/or recruit someone better after failure


By testing the code I mean that I actually focused on tests passing and the output in the examples being produced by AI running lldb using this modified compiler.


It's clear Claude adapted code directly from the OxCaml implementation (the PR author said he pointed Claude at that code [1] and then provides a ChatGPT analysis [2] that really highlights the plagiarism, but ultimately comes to the conclusion that it isn't plagiarized).

Either that highlights someone who is incompetent or they are willfully being blasé. Neither bodes well for contributing code while respecting copyright (though mixing and matching code on your own private repo that isn't distributed in source or binary form seems reasonable to me).

[1]: https://github.com/ocaml/ocaml/pull/14369#issuecomment-35573...

[2]: https://github.com/ocaml/ocaml/pull/14369#issuecomment-35566...


The key is that AI adapted, not stole.

It's actually capable of reasoning and generating derivative code and not just copying stuff wholesale.

See examples at the bottom of my post:

https://joel.id/ai-will-write-your-next-compiler/


Sorry, this is just ridicilous and shows how people fragile really are. This whole topic and whole MR as well.

I am routinely looking into the folly implementation, sometimes into the libstdc++, sometimes into libc++, sometimes into boost or abseil etc. to find inspiration for problems that I tackle in other codebases. By the same standards, this should also be plagiarism, no? I manufacture new ideas by compiling existing knowledge from elsewhere. Literally every engineer in the world does the same. Why is AI any different?


Perhaps because the AI assigned copyright in the files to the author of the library it copied from and the person prompting it told it to look at that library. Without even getting into the comedy AI generated apologia to go with it which makes it look worse rather than better.

From a pragmatic viewpoint as an engineer you assign the IP you create over to the company you work for so plagarism has real world potential to lose you your job at best. There's a difference between taking inspiration from something unrelated "oh this is a neat algorithmic approach to solving this class of problems" to "I need to implement this specific feature and it exists in this library so I'll lift it nearly verbatim".


Can you give an example what exactly was copied? I ask because I took a look into MR and original repo, and the conclusion is that the tool only copy-pasted the copyright header but not the code. So I am still wondering - what's wrong with that (it's a silly mistake even a human can make), and where is the copyright infringement everyone is talking about?


> copy-past[ing] the copyright header but not the code [is] a silly mistake even a human can make

Do you mind showing me some examples of that? That seems so implausible to me

Just for reference, here's another example of AI adding phantom contributors and the human just ignoring it or not even noticing: https://github.com/auth0/nextjs-auth0/issues/2432


Oh wow. That's just egregious. Considering the widespread use of Auth0, I'm surprised this isn't a bigger story.


> Do you mind showing me some examples of that? That seems so implausible to me

What's so special about it that I need to show you the example?


You are claiming humans copy-and-paste copyright headers without copying the corresponding code. To prove you're correct, you only need to show one (or a few) examples of it happening. To prove you incorrect, someone would have to go through all code in existence to show the absence of the phenomenon.

Hence the burden of proof is on you.


No code besides the header was copied so I am asking what is so problematic about it?


that was already explained before


None of that matters. The header is there, in writing, and discussed in the PR. It is acknowledged by both parties and the author gives a clumsy response for its existence. The PR is simply tainted by this alone, not to mention other pain points.

You may not consider this problematic. But maintainers of this project sure do, given this was one of the immediate concerns of theirs.


OxCaml is a fork of OCaml, they have the same license.

I wasn't able to find any chunks of code copied wholesale from OxCaml which already has a DWARF implementation.

All that code wasn't written by Mark, AI just decided to paste his copyright all over.


It matters because it completely weakens their point of stance and make them look unreasonable. Header is irrelevant since it isn't copyright infringement, and FWIW when it has been corrected (in the MR), then they decided that the MR is too complex for them and closed the whole issue. Ridiculous.


An incorrect copyright header is a major red flag for non technical reasons. If you think it is an irrelevant minor matter then you do not undesirable several very important social and legal aspects of the issue.


Social maybe yes what legal aspects? Everybody keeps repeating that but there is no copyright infringement. Maybe you can point me to one?

I understand that people are uncomfortable with this, I am likely too, but objectively looking there's technically nothing wrong or different to what humans already do.


The point is that it ended up in the PR in the first place. The submitted seemed unaware of its presence and only looked into it after it was pointed out. This is sloppy and is a major red flag.


So there's no point? Sloppy maybe yes but technically incorrect or legally questionable no. Struggle is real


If the submitter is sloppy with things that are not complicated, how can one be sure of things that ARE complicated?


The funny thing is that it works, have a look at the MR. It says:

  All existing tests pass. Additional DWARF tests verify:

  DWARF structure (DW_TAG_compile_unit, DW_TAG_subprogram).
  Breakpoints by function and line in both GDB and LLDB.
  Type information and variable visibility.
  Correct multi-object linking.
  Platform-specific relocation handling.
So the burden of proof is obviously not anymore on the MR submitter side but the other.


Yes?

That is why some people are forbidden to contribute to projects if their eyes have read projects with incompatible licenses, in case people go to copyright court.


Yes what? Both oxcaml and ocaml have compatible LGPL licenses so I didn't get your argument.

But even if that hadn't been the case, what exactly would be the problem? Are you saying that I cannot learn from a copyrighted book written by some respected and known author, and then apply that knowledge elsewhere because I would be risking to be sued for copyright infringement?


The wider point is that copyright headers are a very important detail and that a) the AI got it wrong b) you did not notice c) you have not taken on board the fact that it is important despite being told several times and have dismissed the issue as unimportant

Which raises the question how many other important incorrect details are buried in the 13k lines of code that you are unaware of and unable to recognise the significance of? And how much mantainer time would you waste being dismissive of the issues?

People have taken the copyright header as indicative of wider problems in the code.


Yes, please then find those for now imaginative issues and drill through them? Sorry, but I haven't seen anyone in that MR calling out for technical deficiencies so this is just crying out loud in a public for no concrete reasons.

It's the same as if your colleague sitting next to you would not allow the MR to be merged for various political and not technical reasons - this is exactly what is happening here.


> Yes, please then find those for now imaginative issues and drill through them?

No, that is a massive amount of work which will only establish what we already know with a high degree of certainty due to the red flags already mentored - that this code is too flawed to begin with.

This is not political, this is looking out for warming signs in order to avoid wasting time. At this stage the burden of proof is on the submitter, not the reviewers


Too flawed? Did you miss that tiny detail that MR fixes a long time issue for ocaml? This is exactly political because there's no legal or technical issue. Only fluff by scared developers. I have no stakes in this but I'm sincerely surprised by the amount of unreasonable and unsubstantiated claims and explanations given in this thread and MR


I don't get why you do not understand why nobody wants to waste time on a MR where the author didn't even themselves have any interest on looking over it even once. https://github.com/ocaml/ocaml/pull/14369/files#diff-bc37d03... also all the unused functions...

did it fix a long time issue? maybe, but 9 tests for 13k lines doesnt give much confidence in that

and even if it worked perfectly, who will maintain this?


"Yes what? Both oxcaml and ocaml have compatible LGPL licenses so I didn't get your argument."

LGPL is a license for distribution, the copyright of the original authors is retained (unless signed away in a contribution agreement, usually to an organization).

"Are you saying that I cannot learn from a copyrighted book written by some respected and known author, and then apply that knowledge elsewhere because I would be risking to be sued for copyright infringement?"

This was not the case here, so not sure how that is related in any way?


Do you understand that no code besides the header copyright was copied? So what copyright exactly are you talking about?


Depends on the license of the original material, which is why they tend to have a list of allowed use cases for copying content.

Naturally there are very flexible ones, very draconian ones, and those in the middle.

Most people get away with them, because it isn't like everyone is taking others to copyright court sessions every single day, unless there are millions at play.


Also, I just took a glance at the PR and even without knowing the language it took 10 seconds for the first WTF. The .md documents Claude generated are added to .gitignore, including one for the pr post itself.

That’s the quality he’s vouching for.


People complaining about AI stealing code may not realize that OxCaml is a fork of the code that AI is modifying. Both forks have the same license and there are people working on both projects.

AI did paste Mark's copyright on all the files for whatever reason but it did not lift the DWARF implementation from OxCaml and paste it into the PR.

The PR wasn't one-shot either. I had to steer it to completion over several days, had one AI review the changes made by the other, etc. The end result is that the code _works_ and does what it says on the tin!


You are under a delusion. I’m serious.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: