Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I found this principle particularly interesting:

    Human oversight: The use of AI must always remain under human control. Its functioning and outputs must be consistently and critically assessed and validated by a human.


Interesting in what sense? Isn't it just stating something plainly obvious?


It is, but unfortunately the fact that to you - and me - it is obvious does not mean it is obvious to everybody.


Quite. One would hope, though, that it would be clear to prestigious scientific research organizations in particular, just like everything else related to source criticism and proper academic conduct.


Did you forget the entire DOGE episode where every government worker in the US had to send an weekly email to an LLM to justify their existence?


I'd hold CERN to a slightly higher standard than DOGE when it comes to what's considered plainly obvious.


Sure, but the way you maintain this standard is by codifying rules that are distinct from the "lower" practices you find elsewhere.

In other words, because of the huge DOGE clusterfuck demonstrated how horrible practices people will actually enact, you need to put this into the principles.


Oddly enough nowadays CERN is very much like a big corpo, yes they do science, but there is a huge overhead of corpo-like people who running CERN as an enterprise that should bring "income".


Can you elaborate on this, hopefully with details and sources including the revenue stream that CERN is getting as a cooperation?


I want to see how obvious this becomes when you start to add agents left and right that make decisions automagically...


Right. It should be obvious that at an organization like CERN you're not supposed to start adding autonomous agents left and right.


Where is “human oversight” in an automated workflow? I noticed the quote didn’t say “inputs”.

And with testing and other services, I guess human oversight can be reduced to _looking at the dials_ for the green and red lights?


Someone's inputs is someone else's outputs, I don't think you have spotted an interesting gap. Certainly just looking at the dials will do for monitoring functioning, but falls well short of validating the system performance.


The real interesting thing is how does that principle interplay with their pillars and goals i.e. if the goal is to "optimize workflow and resource usage" then having a human in the loop at all points might limit or fully erode this ambition. Obviously it not that black and white, certain tasks could be fully autonomous where others require human validation and you could be net positive - but - this challenge is not exclusive to CERN that's for sure.


Do they hold the CERN Roomba to the same standard? If it cleans the same section of carpet twice is someone going to have to do a review?


It's still just a platitude. Being somewhat critical is still giving some implicit trust. If you didn't give it any trust at all, you wouldn't use it at all! So they endorse trusting it is my read, exactly the opposite of what they appear to say!

It's funny how many official policies leave me thinking that it's a corporate cover-your-ass policy and if they really meant it they would have found a much stronger and plainer way to say it


"You can use AI but you are responsible for and must validate its output" is a completely reasonable and coherent policy. I'm sure they stated exactly what they intended to.


If you have a program that looks at CCTV footage and IDs animals that go by.. is a human supposed to validate every single output? How about if it's thousands of hours of footage?

I think parent comment is right. It's just a platitude for administrators to cover their backs and it doesn't hold to actual usecases


I don't see it so bleakly. Using your analogy, it would simply mean that if the program underperforms compared to humans and starts making a large amount of errors, the human who set up the pipeline will be held accountable. If the program is responsible for a critical task (ie the animal will be shot depending on the classification) then yes, a human should validate every output or be held accountable in case of a mistake.


I take an interest in plane crashes and human factors in digital systems. We understand that there's a very human aspect of complacency that is often read about in reports of true disasters, well after that complacency has crept deep into an organization.

When you put something on autopilot, you also massively accelerate your process of becoming complacent about it -- which is normal, it is the process of building trust.

When that trust is earned but not deserved, problems develop. Often the system affected by complacency drifts. Nobody is looking closely enough to notice the problems until they become proto-disasters. When the human finally is put back in control, it may be to discover that the equilibrium of the system is approaching catastrophe too rapidly for humans to catch up on the situation and intercede appropriately. It is for this reason that many aircraft accidents occur in the seconds and minutes following an autopilot cutoff. Similarly, every Tesla that ever slammed into the back of an ambulance on the back of the road was a) driven by an AI, b) that the driver had learned to trust, and c) the driver - though theoretically responsible - had become complacent.


Sure, but not every application has dramatic consequences such as plane or car crashes. I mean, we are talking about theoretical physics here.


Theoretical? I don't see any reason that complacency is fine in science. If it's a high school science project and you don't actually care at all about the results, sure.


Half-Life showed a plausible story of how high energy physics could have unforeseen consequences.


The problem is that the original statement is too black and white. We make tradeoffs based on costs and feasibility

"if the program underperforms compared to humans and starts making a large amount of errors, the human who set up the pipeline will be held accountable"

Like.. compared to one human? Or an army of a thousand humans tracking animals? There is no nuance at all. It's just unreasonable to make a blanket statement that humans always have to be accountable.

"If the program is responsible for a critical task .."

See how your statement has some nuance? and recognizes that some situations require more accountability and validation that others?


Exactly.

If some dogs chew up an important component, the CERN dog-catcher won't avoid responsibility just by saying "Well, the computer said there weren't any dogs inside the fence, so I believed it."

Instead, they should be taking proactive steps: testing and evaluating the AI, adding manual patrols, etc.


That doesn't follow. Say you write a proof for a something I request, I can then check that proof. That doesn't mean I don't derive any value from being given the proof. A lack of trust does not imply no use.


> So they endorse trusting it is my read, exactly the opposite of what they appear to say!

They endorse limited trust, not exactly a foreign concept to anyone who's taken a closer look at an older loaf of bread before cutting a slice to eat.


I think you're more reading what you want to read out of that - but that's the problem, it's too ambiguous to be useful




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: