Security Superstream: Secure Code in the Age of AI - O'Reilly Media
AI tools are transforming the ways that we write and deploy code, making development faster and more efficient, but they also introduce new risks and vulnerabilities. To protect organizations, securit...
I’m excited to be hosting the O’Reilly Security Superstream: Secure Code in the Age of AI on October 7 at 11:00 AM ET.
We’ll be diving into practical insights, real-world experiences, and emerging trends to address the full spectrum of AI security.
✨ Save your free spot here: bit.ly/4nEWzgj
30.09.2025 17:52
👍 1
🔁 0
💬 0
📌 0
Persistent prompt injections can manipulate LLM behavior across sessions, making attacks harder to detect and defend against. This is a new frontier in AI threat vectors.
Read more: dl.acm.org/doi/10.1145/...
#PromptInjection #Cybersecurity #AIsecurity
10.07.2025 18:14
👍 2
🔁 0
💬 0
📌 0
New research reveals timing side channels can leak ChatGPT prompts, exposing confidential info through subtle delays. AI security needs to consider more than just inputs.
Read more: dl.acm.org/doi/10.1145/...
#AIsecurity #SideChannel #LLM
09.07.2025 23:22
👍 1
🔁 0
💬 0
📌 0
Magistral
We introduce Magistral, Mistral's first reasoning model and our own scalable reinforcement learning (RL) pipeline. Instead of relying on existing implementations and RL traces distilled from prior…
Magistral is Mistral’s first reinforcement‑learning‑only reasoning model.
Shows gains in math, code, and multimodal reasoning—all built from the ground up. Worth a look if RL‑based LLMs are on your radar.
🔗 arxiv.org/abs/2506.10910
08.07.2025 18:07
👍 0
🔁 0
💬 0
📌 0
OpenAI can rehabilitate AI models that develop a “bad boy persona”
Researchers at the company looked into how malicious fine-tuning makes a model go rogue, and how to turn it back.
OpenAI shows that fine-tuning on biased data can induce misaligned 'personas' in language models, but such behavioral shifts can often be detected and reversed.
Read more: www.technologyreview.com/2025/06/18/1...
#Bias #OpenAI
25.06.2025 23:22
👍 0
🔁 0
💬 0
📌 0
To ensure AI is truly open source, we need full access to:
1. The datasets for training and testing
2. The source code
3. The model's architecture
4. The parameters of the model.
Without these, transparency and replicating outcomes are lacking.
#OpenSourceAI #Transparency
19.06.2025 17:18
👍 0
🔁 0
💬 0
📌 0
Antagonistic AI
The vast majority of discourse around AI development assumes that subservient, "moral" models aligned with "human values" are universally beneficial -- in short, that good AI is sycophantic AI. We…
While most AI aims to be compliant and "moral," this study explores the potential benefits of antagonistic AI—systems that challenge and confront users—to promote critical thinking and resilience, emphasizing ethical design grounded in consent, context, and framing.
arxiv.org/abs/2402.07350
13.06.2025 13:42
👍 2
🔁 0
💬 1
📌 0
CTRAP: Embedding Collapse Trap to Safeguard Large Language Models from Harmful Fine-Tuning Attacks
CTRAP is a promising pre-deployment alignment method that makes AI models resistant to harmful fine-tuning by causing them to "break" if malicious tuning occurs, while remaining stable under benign changes.
anonymous.4open.science/r/CTRAP/READ...
12.06.2025 13:47
👍 1
🔁 0
💬 0
📌 0
LLMs Get Lost In Multi-Turn Conversation
Large Language Models (LLMs) are conversational interfaces. As such, LLMs have the potential to assist their users not only when they can fully specify the task at hand, but also to help them define,…
Large language models (LLMs) see a 39% drop in effectiveness in multi-turn dialogues versus single-turn tasks due to their tendency for hasty assumptions and premature response finalization, leading to inconsistency and error correction challenges.
arxiv.org/abs/2505.06120
#AI #MachineLearning
10.06.2025 13:39
👍 1
🔁 0
💬 0
📌 0
Tina: Tiny Reasoning Models via LoRA
How cost-effectively can strong reasoning abilities be achieved in language models? Driven by this fundamental question, we present Tina, a family of tiny reasoning models achieved with high…
Tina models leverage low-rank adaptation and reinforcement learning to offer robust, economical reasoning capabilities, making advanced AI more accessible and budget-friendly for innovators.
For more details, visit: arxiv.org/abs/2504.15777
#AI #Innovation #MachineLearning
09.06.2025 13:51
👍 1
🔁 0
💬 0
📌 0
Power Hungry
An unprecedented look at the state of AI’s energy and resource usage, where it is now, where it is headed in the years to come, and why we have to get it right.
MIT Tech Review's AI Energy Package highlights the enormous energy and water usage involved in AI model training and operation. This is crucial for grasping AI's environmental impact and its implications for sustainable technology. #AI #Sustainability
www.technologyreview.com/supertopic/a...
04.06.2025 13:42
👍 1
🔁 0
💬 0
📌 0
Most AI chatbots easily tricked into giving dangerous responses, study finds
Researchers say threat from ‘jailbroken’ chatbots trained to churn out illegal information is ‘tangible and concerning’
A new study reveals that AI chatbots can be easily tricked into offering guidance on hacking, creating explosives, cybercrime methods, and other illicit or dangerous activities.
#AI #CyberSecurity
www.theguardian.com/technology/2...
03.06.2025 13:39
👍 1
🔁 0
💬 0
📌 0