Let's assume that someone creates a GPT model which
1. searches the internet for server software vulnerability or just find security holes by reading source codes,
2. writes code to exploit, searches internet for vulnerable servers,
3. copy itself to any server it can hack and replicate.
Bonus steps:
– randomly modifies its hyper-parameters, collect extra learning data, and trains itself,
– generate a script with evil purposes (“What would an evil AI do? Write a python script.”) and executes it.
When we will have such a GPT virus, how will we fight against it? Patch all software? Create AI-free part of the internet?
(Please consider that any answer to this question might get into the training data of a GPT model.)