Frontier Model Security
Anthropic details security practices and threat model for frontier-scale language models.
Search the full wire by company, model, lab, or keyword. Every story we have ever aggregated.
Anthropic details security practices and threat model for frontier-scale language models.
Anthropic proposes accountability frameworks and regulatory approaches for AI system oversight and governance.
Anthropic closes $450M Series C funding round to expand product deployment and AI reliability research.
Zoom invests in Anthropic and announces Claude API integration for video conferencing platform.
Anthropic releases Claude with 100K token context window, enabling longer document and code analysis.
Anthropic describes Constitutional AI method: training LLMs with explicit principles for safer, more controllable outputs.
Anthropic partners with Scale AI to integrate Claude into enterprise data labeling and AI infrastructure workflows.
Anthropic advocates for NIST AI risk management framework and increased federal investment in AI safety standards.
Anthropic introduces Claude, a large language model trained with Constitutional AI for safer, more reliable performance.
Anthropic outlines safety research priorities: scalable alignment techniques, interpretability, and robustness evaluation methods.
Anthropic partners with Google Cloud to integrate Claude API into GCP ecosystem for enterprise deployments.