Odd take from a professor. You asked a predictive text VLLM to pretend it wanted to escape and it’s telling the statistically most likely story as to how a fictional escape might happen. It’s a sock puppet on your own hand, and you’re afraid of it?
Problem with AI hype right now is anthropomorphizing these things. They are statistical models of us - often the worst of us - but they aren’t thinking. The focus should be on ethical use to improve the lives of folks who are suffering. AI for good.