Comment by myrmidon

5 days ago

What is your worst-case scenario here?

Something like a pop-sci article along the lines of "Mad scientists create racist, imperialistic AI"?

I honestly don't see publication of the weights as a relevant risk factor, because sensationalist misrepresentation is trivially possible with the given example responses alone.

I don't think such pseudo-malicious misrepresentation of scientific research can be reliably prevented anyway, and the disclaimers make your stance very clear.

On the other hand, publishing weights might lead to interesting insights from others tinkering with the models. A good example for this would be the published word prevalence data (M. Brysbaert et al @Ghent University) that led to interesting follow-ups like this: https://observablehq.com/@yurivish/words

I hope you can get the models out in some form, would be a waste not to, but congratulations on a fascinating project regardless!

It seems like if there is an obvious misuse of a tool, one has a moral imperative to restrict use of the tool.

  • Every tool can be misused. Hammers are as good for bashing heads as building houses. Restricting hammers would be silly and counterproductive.

    • Yes but if you are building an voice activated autonomous flying hammer then you either want it to be very good at differentiating heads from hammers OR you should restrict its use.

      1 reply →