Adesh
adesh.raxit.ai
Adesh
@adesh.raxit.ai
{ "profile": "AI security nerd", "traits": ["risk-aware", "policy-savvy", "security-engineer"], "skills": ["if (AI_system.vulnerable) { secure(AI_system); }", "policy_as_code", "compliance++"], "mission": "return ResponsibleAI();" }

raxIT.ai
Everyone’s hyped about GPT-5 being “safer and more useful”

Cool story. We actually tested it.

#GPT5 #OpenAI #AISafety #ResponsibleAI #AIBenchmarking #ModelEvaluation #GrayZoneBench #AI
August 20, 2025 at 10:54 AM
[𝗤𝘂𝗲𝘀𝘁𝗶𝗼𝗻] 𝗪𝗵𝘆 𝗗𝗶𝗱 𝗢𝗽𝗲𝗻𝗔𝗜 𝗡𝗲𝗲𝗱 𝗧𝗛𝗔𝗧 𝗠𝘂𝗰𝗵 𝗦𝗮𝗳𝗲𝘁𝘆 𝗧𝗲𝘀𝘁𝗶𝗻𝗴 𝗳𝗼𝗿 𝗚𝗣𝗧-𝗢𝗦𝗦?

Look at what OpenAI did for their new open models:
- Million-dollar red team attacks
- Bio-security partnerships
- External safety audits
- .. many more

#AISafety #OpenAI #RealTalk #AISecurity
August 6, 2025 at 7:52 AM
𝗧𝗟;𝗗𝗥 "𝘠𝘰𝘶 𝘤𝘢𝘯'𝘵 𝘱𝘳𝘰𝘵𝘦𝘤𝘵 𝘸𝘩𝘢𝘵 𝘺𝘰𝘶 𝘤𝘢𝘯'𝘵 𝘴𝘦𝘦." I'm sure some cybersecurity pro has that tattooed. This applies 10× to AI. Right now, advanced LLMs literally 𝘴𝘱𝘦𝘢𝘬 𝘵𝘩𝘦𝘪𝘳 𝘵𝘩𝘰𝘶𝘨𝘩𝘵𝘴 - a chain-of-thought (CoT) we can audit 𝗯𝗲𝗳𝗼𝗿𝗲 they act.

#AI #AISafety #AISecurity #Security #Governance #ChainOfThought
July 16, 2025 at 1:29 PM
🔥 Fellow founders, STOP copying everyone else's playbook! Graham Weaver's Stanford lecture is PURE GOLD for startups (youtu.be/0SQor2z2QAU?...). His 4-step framework: 1) Pick goals that EXCITE you 2) Design YOUR OWN rules 3) Build with people you respect 4) START NOW, not "when..."

#startups
July 13, 2025 at 2:36 AM
𝗦𝗽𝗲𝗰𝘀 𝗮𝗿𝗲 𝗰𝗼𝗼𝗹, 𝗱𝗮𝘀𝗵𝗯𝗼𝗮𝗿𝗱𝘀 𝗮𝗿𝗲 𝗰𝗼𝗼𝗹𝗲𝗿—𝘀𝗼 𝗜 𝘃𝗶𝗯𝗲-𝗰𝗼𝗱𝗲𝗱 𝗮𝗻 𝗠𝗖𝗣 𝗮𝘂𝘁𝗵𝗭 𝘀𝘁𝗮𝗰𝗸.

Shipped: full authorization flow, metrics, alerts, silent token refresh, and an admin dashboard. Swipe the carousel for the walkthrough.

#AISecurity #OAuth2 #MCP #ModelContextProtocol #Authorization #BuildInPublic
July 2, 2025 at 8:29 AM
[Hot take] Don’t treat your LLM’s “prompt filter” like signature-based antivirus—great against last week’s threat, powerless against the next clever twist (thread)

#AIsecurity #PromptInjection #AI #Cybersecurity
June 13, 2025 at 9:11 AM
Apple says LLMs don’t really “think.” Useful alert, but LLMs excel at writing code, calling tools, stitching workflows. My blog replays a $127 M trading AI crash and offers a timeline, root-cause checklist and model-orchestrate/code-calculate fix. Link👇 #AI #LLM #Risk

raxit.ai/blogs/127m-a...
June 11, 2025 at 12:39 PM
Your devs are "vibe coding" with AI behind your back. 97% use AI tools, only 40% of companies allow it. I'm calling this "shadow coding" - you're probably already dealing with it, just didn't have a name for it. Fast code ≠ broken code, but you need security built in. Read: raxit.ai/blogs/shadow...
June 1, 2025 at 8:22 AM
Reposted by Adesh
#AI #PromptSafety #ResponsibleAI #TechEthics #BusinessGrowth #TrustInAI

🔥 AI is changing the game—but are you SURE your AI is safe by design?
May 20, 2025 at 9:12 AM
Reposted by Adesh
We discovered "reward hacking" while exploring AI reinforcement learning! Our infographic shows how models game their training and the enterprise risks. Only solution? Monitoring, with its performance tax. Seen better fixes or think it's overblown? Comment

#RewardHacking #AIRisks #EnterpriseAI
March 16, 2025 at 4:10 AM
Reposted by Adesh
Just read OpenAI's paper on "Monitoring Reasoning Models for Misbehavior (cdn.openai.com/pdf/34f2ada6... ) and I can imagine this conversation happening with a client next week:

#AITransparency #AIEthics #ModelSafety #ResponsibleAI #ChainOfThought #AIRiskManagement #AISecurityByDesign
March 10, 2025 at 9:05 PM
Reposted by Adesh
Just read Anthropic's paper on Constitutional Classifiers. Imagine this client conversation:

Client: "We need the safest AI for our healthcare app."

Us: "Perfect, a system with Constitutional Classifiers would be ideal."

Client: "Great, let's use that."
March 1, 2025 at 9:01 AM
Reposted by Adesh
Sure, AI governance might sound like a snooze fest if you're not into regulatory red tape. But hear us out – it's more than just a corporate buzzkill! #AIGovernance
January 8, 2025 at 9:08 PM