site banner

Small-Scale Question Sunday for May 7, 2023

Do you have a dumb question that you're kind of embarrassed to ask in the main thread? Is there something you're just not sure about?

This is your opportunity to ask questions. No question too simple or too silly.

Culture war topics are accepted, and proposals for a better intro post are appreciated.

2
Jump in the discussion.

No email address required.

Someone recently pointed out that with the advent of tools like GitHub Copilot, we will reach a point where most of the code Copilot is trained on will have been generated by Copilot or other, similar tools.

What weird/negative effects do you predict this will have on tools like Copilot and ChatGPT? Will successive generations of AIs mindlessly amplify small quirks in the original human-generated data set? Or will AIs become good at detecting AI generated content and assign it a lower weight? Or will something else happen?

I'm not sure it matters because the Copilot code that has been committed has been filtered by a developer, so it's a bit like RLHF. The human is still in the loop, so the only qualities that get amplified are the ones the humans want.