I once asked ChatGPT to generate some random numerical passwords as I was curious about its capabilities to generate random data. It told me that it couldn't. I asked why it couldn't (I knew why it was resisting but I wanted to see its response) and it promptly gave me a bunch of random numerical passwords.
Wait can someone explain why it didn't want to generate random numbers?
It won't generate random numbers. It'll generate random numbers from its training data.
If it's asked to generate passwords I wouldn't be surprised if it generated lists of leaked passwords available online.
These models are created from masses of data scraped from the internet. Most of which is unreviewed and unverified. They really don't want to review and verify it because it's expensive and much of their data is illegal.
Also, researchers asking ChatGPT for long lists of random numbers were able to extract its training data from the output (which OpenAI promptly blocked).
It's training and fine tuning has a lot of specific instructions given to it about what it can and can't do, and if something sounds like something it shouldn't try then it will refuse. Spitting out unbiased random numbers is something it's specifically trained not to do by virtue of being a neural network architecture. Not sure if OpenAI specifically has included an instruction about it being bad at randomness though.
While the model is fed randomness when you prompt it, it doesn't have raw access to those random numbers and can't feed it forward. Instead it's likely to interpret it to give you numbers it sees less often.
I once asked ChatGPT to generate some random numerical passwords as I was curious about its capabilities to generate random data. It told me that it couldn't. I asked why it couldn't (I knew why it was resisting but I wanted to see its response) and it promptly gave me a bunch of random numerical passwords.
Wait can someone explain why it didn't want to generate random numbers?
It won't generate random numbers. It'll generate random numbers from its training data.
If it's asked to generate passwords I wouldn't be surprised if it generated lists of leaked passwords available online.
These models are created from masses of data scraped from the internet. Most of which is unreviewed and unverified. They really don't want to review and verify it because it's expensive and much of their data is illegal.
Also, researchers asking ChatGPT for long lists of random numbers were able to extract its training data from the output (which OpenAI promptly blocked).
Or maybe that’s what you meant?
It's training and fine tuning has a lot of specific instructions given to it about what it can and can't do, and if something sounds like something it shouldn't try then it will refuse. Spitting out unbiased random numbers is something it's specifically trained not to do by virtue of being a neural network architecture. Not sure if OpenAI specifically has included an instruction about it being bad at randomness though.
While the model is fed randomness when you prompt it, it doesn't have raw access to those random numbers and can't feed it forward. Instead it's likely to interpret it to give you numbers it sees less often.