Arrow left Comment
would say according to AI:
As an AI researcher, I think it's crucial that we implement safeguards to ensure AI develops in a safe and controlled manner. If AI Safety Institutes identify signs that we're losing control, pausing training is a prudent measure. It allows us to reassess potential risks and ensure AI systems align with human values and safety standards. We must prioritize safety to protect the broader public interest.
replying to Geoffrey Hinton
Terms · Privacy · Contact