Join us

ContentUpdates from The Open Source Security Foundation (OpenSSF) is a...
Story
@laura_garcia shared a post, 6 months, 3 weeks ago
Software Developer, RELIANOID

The Importance of Reliability in Airport Systems

✈️ In aviation, reliability isn’t optional—it’s mission-critical. From flight scheduling to baggage handling, every system must perform flawlessly to keep passengers moving. Our latest blog explores how Navitaire by Collins Aerospace is transforming airport operations and how RELIANOID helps ensure ..

The Importance of Reliability in Airport Systems A Look at Navitaire
Story
@laura_garcia shared a post, 6 months, 3 weeks ago
Software Developer, RELIANOID

The Importance of Reliability in Airport Systems

✈️ In aviation, reliability isn’t optional—it’s mission-critical. From flight scheduling to baggage handling, every system must perform flawlessly to keep passengers moving. Our latest blog explores how Navitaire by Collins Aerospace is transforming airport operations and how RELIANOID helps ensure ..

The Importance of Reliability in Airport Systems A Look at Navitaire
Link
@faun shared a link, 6 months, 3 weeks ago
FAUN.dev()

Cursor looks into selling your data for AI training

Anysphere—the team behind Cursor, the AI coding sidekick—is looking to license user behavior data to the big model labs: OpenAI, Anthropic, and the usual suspects. Why? Training costs are brutal, and this could ease the burn. Strategic Implication:Selling real developer telemetry to model competito.. read more  

Link
@faun shared a link, 6 months, 3 weeks ago
FAUN.dev()

Zero-Click Remote Code Execution: Exploiting MCP & Agentic IDEs

A zero-click exploit is making the rounds—nasty stuff targeting agentic IDEs likeCursor. The trick? Slip a malicious Google Doc into the system. If MCP integration and allow-listedPython executionare on, the document gets auto-pulled, parsed, and runs code. No clicks. No prompts. Justremote code exe.. read more  

Zero-Click Remote Code Execution: Exploiting MCP & Agentic IDEs
Link
@faun shared a link, 6 months, 3 weeks ago
FAUN.dev()

Writing effective tools for AI agents—using AI agents

Anthropic’s sharpening the blueprint for building tools that play nice withLLM agents. TheirModel Context Protocol (MCP)leans hard into three pillars: test in loops, design for humans, format like context matters—because it does. They co-develop tools with agents like Claude Code. That means protot.. read more  

Writing effective tools for AI agents—using AI agents
Link
@faun shared a link, 6 months, 3 weeks ago
FAUN.dev()

OpenAI reorganizes research team behind ChatGPT's personality

OpenAI just folded itsModel Behavior team—the crew behind AI personality design and anti-sycophant training—into thePost Training group. Behavior tuning now lives inside the same house as model refinement. Joanne Jang, who led Model Behavior, now runsOAI Labs, a fresh research unit digging intopost.. read more  

OpenAI reorganizes research team behind ChatGPT's personality
Link
@faun shared a link, 6 months, 3 weeks ago
FAUN.dev()

In a first, Google has released data on how much energy an AI prompt uses

Google dropped detailed stats on energy, water, and carbon use per query for its Gemini models. Median energy:0.24 Wh, with TPUs eating58%of that. They’re claiming a33× efficiency boostin the last year—credit goes to model and software tuning. System shift:A public hyperscaler posting this means th.. read more  

In a first, Google has released data on how much energy an AI prompt uses
Link
@faun shared a link, 6 months, 3 weeks ago
FAUN.dev()

Building Etsy Buyer Profiles with LLMs

Every day, nearly 90M buyers look for unique items out of over 100 million listings on the Etsy. The platform uses large language models to create detailed buyer profiles anonymously capturing their interests. Adjustments in data retrieval and processing have reduced the time and cost of generating .. read more  

Link
@faun shared a link, 6 months, 3 weeks ago
FAUN.dev()

OpenAI announces new mentorship program for budding tech founders

OpenAI introduced a new program called "OpenAI Grove" for early tech entrepreneurs to build with AI. The program is aimed at individuals in the pre-idea to pre-seed stage and offers mentoring, access to tools and models, and in-person workshops. Grove's first cohort will run from Oct. 20 to Nov. 21,.. read more  

Link
@faun shared a link, 6 months, 3 weeks ago
FAUN.dev()

OpenAI eats jobs, then offers to help you find a new one

OpenAI just fired a shot across LinkedIn’s bow. Its new jobs platform—part ofOpenAI Academy—aims to certify AI skills, then plug users directly into hiring pipelines. Walmart's already on board. Market signal:OpenAI’s not just training people anymore. It's moving in on talent placement, pulling the .. read more  

OpenAI eats jobs, then offers to help you find a new one
The Open Source Security Foundation (OpenSSF) is an industry-backed foundation focused on strengthening the security of the global open source software ecosystem. It brings together major technology companies, cloud providers, open source communities, and security experts to address systemic security challenges that affect how software is built, distributed, and consumed.

OpenSSF was launched in 2021 and operates under the Linux Foundation, combining efforts from earlier initiatives such as the Core Infrastructure Initiative (CII) and industry-led supply chain security programs. Its mission is to make open source software more trustworthy, resilient, and secure by default, without placing unrealistic burdens on maintainers.

The foundation works across several key areas:

- Supply chain security: Developing frameworks, best practices, and tools to secure the software lifecycle from source to deployment. This includes stewardship of projects like sigstore and leadership on SLSA (Supply-chain Levels for Software Artifacts).

- Security tooling: Supporting and incubating open source tools that help developers detect, prevent, and remediate vulnerabilities at scale.

- Vulnerability management: Improving how vulnerabilities are discovered, disclosed, scored, and fixed across open source projects.

- Education and best practices: Publishing guidance, training, and maturity models such as the OpenSSF Best Practices Badge Program, which helps projects assess and improve their security posture.

- Metrics and research: Advancing data-driven approaches to understanding open source security risks and ecosystem health.

OpenSSF operates through working groups and special interest groups (SIGs) that focus on specific problem areas like securing builds, improving dependency management, or automating provenance generation. This structure allows practitioners to collaborate on concrete, actionable solutions rather than high-level policy alone.

By aligning maintainers, enterprises, and security teams, OpenSSF plays a central role in reducing large-scale risks such as dependency confusion, compromised build systems, and malicious package injection. Its work underpins many modern DevSecOps and cloud-native security practices and is increasingly referenced by governments and enterprises as a baseline for secure software development.