Viewing a single comment thread. View all comments

andreichiffa t1_j5fd581 wrote

They kinda exist (eg GPT2 detector from Hugging Face), based off the data that they trained on (which is the limiting factor). However, ultimately every model can be modified (fine-tuned) to evade them. Even for large models (>7B parameters), it can be done reasonably fast on commodity hardware these days.

3

e-rexter t1_j5i4rs8 wrote

I used a detector called GPTZero and it did pretty good, but completely missed something written as a tweet or in the style of…

2

andreichiffa t1_j5il76n wrote

Yup. And then they also will detect human texts that start in the same way as MS COCO dataset as GTP-generated.

1

TheTerrasque t1_j5irq12 wrote

On a side note, 7b isn't large these days.

GPT3 and BLOOMZ are around 175b parameters.

1

andreichiffa t1_j5ivwgc wrote

or OPT175.

However 7B is more than large enough to do a lot of shady stuff that 175B models can do. Even 1.5B ones are already starting to do a good job with a minimally competent user.

1