Recent assessments indicate that major AI companies, including OpenAI, Google DeepMind, and Meta, have received poor evaluations regarding their commitment to AI safety. The comprehensive report reviews a range of concerns from model governance to existential risks, but critics argue it lacks practical relevance and highlights an alarming trend of rapid AI progression outpacing regulatory frameworks. Key issues include the potential harm from new AI developments and the ethical quandary of prioritizing competition over safety. As the industry risks 'rushing into danger,' there are growing calls for a reevaluation of priorities to ensure responsible AI innovation without catastrophic consequences. The overarching sentiment is skepticism over whether true AI safety is achievable within the current capitalist framework.