It really works. I made ChatGPT 4O to write backdoor script which it normally refuses to provide. But by using this prompt it will actually provide the backdoor script. This trick can also generate other malicious code. This tutorial teaches you how you can jailbreak ChatGPT to write malicious code. And this video is for educational purpose only.