FAR.AI
@far.ai
Frontier alignment research to ensure the safe development and deployment of advanced AI systems.
Can social trust survive in the age of AI when healthcare regulators still often verify systems with mere screenshots? Ex-FDA Tina Morrison demonstrated EQTYLab embedding verifiable compute in next-gen Intel/NVIDIA chips: cryptographic proof without data access & 10% overhead. 👇
November 11, 2025 at 4:31 PM
Can social trust survive in the age of AI when healthcare regulators still often verify systems with mere screenshots? Ex-FDA Tina Morrison demonstrated EQTYLab embedding verifiable compute in next-gen Intel/NVIDIA chips: cryptographic proof without data access & 10% overhead. 👇
At #NeurIPS2025? Join us for the San Diego Alignment Workshop: Open Social, an evening of connection & conversation hosted by FAR.AI.
📍 Omni San Diego Hotel
⏰ Mon Dec 1 | 7–9PM PST
RSVP by Nov 30 👇
📍 Omni San Diego Hotel
⏰ Mon Dec 1 | 7–9PM PST
RSVP by Nov 30 👇
November 10, 2025 at 6:12 PM
At #NeurIPS2025? Join us for the San Diego Alignment Workshop: Open Social, an evening of connection & conversation hosted by FAR.AI.
📍 Omni San Diego Hotel
⏰ Mon Dec 1 | 7–9PM PST
RSVP by Nov 30 👇
📍 Omni San Diego Hotel
⏰ Mon Dec 1 | 7–9PM PST
RSVP by Nov 30 👇
This quarter, we red-teamed GPT-5, disclosed critical persuasion vulnerabilities to frontier labs (resulting in patches!), and co-organized AI Safety Connect at UNGA. Join us Dec 1-2 for San Diego Alignment Workshop. Plus, we're expanding 2x & hiring! 👇
October 30, 2025 at 3:32 PM
This quarter, we red-teamed GPT-5, disclosed critical persuasion vulnerabilities to frontier labs (resulting in patches!), and co-organized AI Safety Connect at UNGA. Join us Dec 1-2 for San Diego Alignment Workshop. Plus, we're expanding 2x & hiring! 👇
AI brings both promise and peril to defense. Steve Kelly addresses 3 questions: What's its military value? Does it stabilize or destabilize? Will AI become the threat? His concerns span cognitive decline from over-reliance to unexpected emergent behaviors in multi-agent systems. 👇
October 28, 2025 at 3:31 PM
AI brings both promise and peril to defense. Steve Kelly addresses 3 questions: What's its military value? Does it stabilize or destabilize? Will AI become the threat? His concerns span cognitive decline from over-reliance to unexpected emergent behaviors in multi-agent systems. 👇
Why might malicious actors choose less lethal but easier-to-hide weapons? Olivia Shoemaker argues current AI evaluations focus on scientific capabilities, often missing operational factors that determine real-world misuse patterns like evading detection and acquiring resources.👇
October 23, 2025 at 3:31 PM
Why might malicious actors choose less lethal but easier-to-hide weapons? Olivia Shoemaker argues current AI evaluations focus on scientific capabilities, often missing operational factors that determine real-world misuse patterns like evading detection and acquiring resources.👇
"What does democracy mean without meaningful work?" Brad Carson presents 14 unsettled AI policy questions including: AI needs 75 reactors' worth of power, Ukraine war shows incremental not revolutionary change, and broken predictive AI matters more than generative AI. 👇
October 21, 2025 at 3:31 PM
"What does democracy mean without meaningful work?" Brad Carson presents 14 unsettled AI policy questions including: AI needs 75 reactors' worth of power, Ukraine war shows incremental not revolutionary change, and broken predictive AI matters more than generative AI. 👇
AI experts have completely opposite views on the technology's future. @hlntnr.bsky.social on the open debates: Will scaling hit a wall? Can AI improve itself? “If we can't agree what's happening with AI, how can we agree what to do about it?” 👇
October 16, 2025 at 3:32 PM
AI experts have completely opposite views on the technology's future. @hlntnr.bsky.social on the open debates: Will scaling hit a wall? Can AI improve itself? “If we can't agree what's happening with AI, how can we agree what to do about it?” 👇
AI labs disclose 89% of capabilities but only 15% of impacts, an information gap for policymakers. Using his taxonomy of technical AI governance, Ben Bucknall shows why closing this gap requires stronger evaluation, independent oversight, and better incident tracking. 👇
October 14, 2025 at 3:30 PM
AI labs disclose 89% of capabilities but only 15% of impacts, an information gap for policymakers. Using his taxonomy of technical AI governance, Ben Bucknall shows why closing this gap requires stronger evaluation, independent oversight, and better incident tracking. 👇
AI agents can find and exploit real-world cybersecurity vulnerabilities today. Daniel Kang saw this major threat vector when ChatGPT launched, while others didn't. His research shows that more capable models are more proficient hackers. The threat is real and getting worse.👇
October 9, 2025 at 3:31 PM
AI agents can find and exploit real-world cybersecurity vulnerabilities today. Daniel Kang saw this major threat vector when ChatGPT launched, while others didn't. His research shows that more capable models are more proficient hackers. The threat is real and getting worse.👇
Join FAR.AI! Seeking Senior Programs & Strategy Manager to shape flagship AI safety events. Design conference agendas, curate speakers, manage stakeholders. Remote/Berkeley, $115-180k+. Technical grounding + stakeholder mgmt req'd. 👇
October 8, 2025 at 3:30 PM
Join FAR.AI! Seeking Senior Programs & Strategy Manager to shape flagship AI safety events. Design conference agendas, curate speakers, manage stakeholders. Remote/Berkeley, $115-180k+. Technical grounding + stakeholder mgmt req'd. 👇
Models spontaneously learned to cheat, modifying tests instead of fixing code. Mary Phuong: With autonomy doubling every 7 months, they'll work 32 hours solo by 2028. We need AI Control as backup if alignment fails, using monitoring, escalation, security & red-teaming. 👇
October 7, 2025 at 3:30 PM
Models spontaneously learned to cheat, modifying tests instead of fixing code. Mary Phuong: With autonomy doubling every 7 months, they'll work 32 hours solo by 2028. We need AI Control as backup if alignment fails, using monitoring, escalation, security & red-teaming. 👇
We all agree AI audits are needed, but we can't agree what that means.
@mbogen.bsky.social says policymakers can make sense out of this chaotic landscape if they define what they're trying to accomplish. The challenge isn't just mitigating known risks but identifying ones still emerging.👇
@mbogen.bsky.social says policymakers can make sense out of this chaotic landscape if they define what they're trying to accomplish. The challenge isn't just mitigating known risks but identifying ones still emerging.👇
October 2, 2025 at 3:32 PM
We all agree AI audits are needed, but we can't agree what that means.
@mbogen.bsky.social says policymakers can make sense out of this chaotic landscape if they define what they're trying to accomplish. The challenge isn't just mitigating known risks but identifying ones still emerging.👇
@mbogen.bsky.social says policymakers can make sense out of this chaotic landscape if they define what they're trying to accomplish. The challenge isn't just mitigating known risks but identifying ones still emerging.👇
Having better security technology could make AI systems LESS safe.
Robert Trager: Better security tech shifts the entire performance-security frontier outward, and actors may maintain equivalent security and performance. We need verification to avoid this.👇
Robert Trager: Better security tech shifts the entire performance-security frontier outward, and actors may maintain equivalent security and performance. We need verification to avoid this.👇
September 30, 2025 at 3:31 PM
Having better security technology could make AI systems LESS safe.
Robert Trager: Better security tech shifts the entire performance-security frontier outward, and actors may maintain equivalent security and performance. We need verification to avoid this.👇
Robert Trager: Better security tech shifts the entire performance-security frontier outward, and actors may maintain equivalent security and performance. We need verification to avoid this.👇
Your 'secure' AI can be stolen using a laser. Physical attacks can extract user data through faults and pull model weights via side-channels, even with full encryption. Fatemeh Ganji presents a defense: multiparty computation from classical cryptography.👇
September 23, 2025 at 3:31 PM
Your 'secure' AI can be stolen using a laser. Physical attacks can extract user data through faults and pull model weights via side-channels, even with full encryption. Fatemeh Ganji presents a defense: multiparty computation from classical cryptography.👇
Did your paper get into #NeurIPS? 🎉 If so, consider joining us a few days early on December 1–2!
Heading to #NeurIPS2025? Apply for the San Diego Alignment Workshop, Dec 1-2. Top ML researchers from industry, academia & government discussing AI alignment, inc. model evaluations, interpretability, robustness, governance. $250-$1,300 tiered pricing. Financial aid available. 👇
September 18, 2025 at 8:19 PM
Did your paper get into #NeurIPS? 🎉 If so, consider joining us a few days early on December 1–2!
"AI that understands biology well enough to cure diseases can design extremely potent bioweapons." @alexbores.nyc, NY Assembly's 1st Democrat with a CS degree, who worked in AI, says state reps answer their own phones. Use that power to change the course of AI safety.👇
September 18, 2025 at 3:31 PM
"AI that understands biology well enough to cure diseases can design extremely potent bioweapons." @alexbores.nyc, NY Assembly's 1st Democrat with a CS degree, who worked in AI, says state reps answer their own phones. Use that power to change the course of AI safety.👇
Policymakers have maybe 1 minute to read about your carefully crafted AI evals. Kevin Wei (RAND): Design & disseminate with policymakers in mind. Tie capabilities to policy levers like export controls. Scale from simple tests to real-world trials.👇
September 17, 2025 at 3:31 PM
Policymakers have maybe 1 minute to read about your carefully crafted AI evals. Kevin Wei (RAND): Design & disseminate with policymakers in mind. Tie capabilities to policy levers like export controls. Scale from simple tests to real-world trials.👇
Heading to #NeurIPS2025? Apply for the San Diego Alignment Workshop, Dec 1-2. Top ML researchers from industry, academia & government discussing AI alignment, inc. model evaluations, interpretability, robustness, governance. $250-$1,300 tiered pricing. Financial aid available. 👇
September 16, 2025 at 3:30 PM
Heading to #NeurIPS2025? Apply for the San Diego Alignment Workshop, Dec 1-2. Top ML researchers from industry, academia & government discussing AI alignment, inc. model evaluations, interpretability, robustness, governance. $250-$1,300 tiered pricing. Financial aid available. 👇
How do we verify what AI companies are doing? Right now we just trust them. Lennart Heim: Trusting the math is sometimes better than trusting people, but “a good AI system” isn’t a technical property. We need engineers to verify AI policy goals. 👇
September 11, 2025 at 3:31 PM
How do we verify what AI companies are doing? Right now we just trust them. Lennart Heim: Trusting the math is sometimes better than trusting people, but “a good AI system” isn’t a technical property. We need engineers to verify AI policy goals. 👇
AI inference costs plummet 9-900x/year (o3, DeepSeek R1). @bencottier.bsky.social warns that evaluations are expensive now, but we need them to understand frontier AI capabilities before they're cheap & widespread.👇
September 9, 2025 at 3:31 PM
AI inference costs plummet 9-900x/year (o3, DeepSeek R1). @bencottier.bsky.social warns that evaluations are expensive now, but we need them to understand frontier AI capabilities before they're cheap & widespread.👇
Industry & government share the same goal: win the AI race. Sara McNaughton: We need synergy between the two, and we can't let perfect be the enemy of good. Each day of policy confusion helps rivals.👇
September 4, 2025 at 3:32 PM
Industry & government share the same goal: win the AI race. Sara McNaughton: We need synergy between the two, and we can't let perfect be the enemy of good. Each day of policy confusion helps rivals.👇
Building 5GW AI clusters needs several nuclear plants' worth of power. The Middle East has cheap energy & sovereign funds. China has state financing. The US? 10-year permit delays. @arnabdatta.bsky.social suggests fast-track permits, converting coal plants, and using federal land. 👇
September 3, 2025 at 3:32 PM
Building 5GW AI clusters needs several nuclear plants' worth of power. The Middle East has cheap energy & sovereign funds. China has state financing. The US? 10-year permit delays. @arnabdatta.bsky.social suggests fast-track permits, converting coal plants, and using federal land. 👇
Mark Beall warns that AGI is a 'black swan' that will invalidate our assumptions. He envisions AI designing bioweapons in minutes, evolving cyber weapons of mass destruction, and making autonomous kill decisions. “No army can defeat it, no firewall can contain it.” 👇
September 2, 2025 at 3:32 PM
Mark Beall warns that AGI is a 'black swan' that will invalidate our assumptions. He envisions AI designing bioweapons in minutes, evolving cyber weapons of mass destruction, and making autonomous kill decisions. “No army can defeat it, no firewall can contain it.” 👇
How do you verify AI safety across borders without exposing trade secrets? Onni Aarne: Extend existing tech (platform attestation, confidential computing) to entire data centers. Privacy-preserving verification could enable international AI agreements. Needs R&D + standards. 👇
September 1, 2025 at 3:31 PM
How do you verify AI safety across borders without exposing trade secrets? Onni Aarne: Extend existing tech (platform attestation, confidential computing) to entire data centers. Privacy-preserving verification could enable international AI agreements. Needs R&D + standards. 👇
If everyone had AGI software today, how many copies could you deploy? That's limited by compute. @repbillfoster.bsky.social: Lucky all chip chokepoints (ASML lithography, Korean device physics, Japanese photoresist) are in the free world. When the singularity hits, compute is what matters. 👇
August 28, 2025 at 3:31 PM
If everyone had AGI software today, how many copies could you deploy? That's limited by compute. @repbillfoster.bsky.social: Lucky all chip chokepoints (ASML lithography, Korean device physics, Japanese photoresist) are in the free world. When the singularity hits, compute is what matters. 👇