Join us

ContentUpdates from BAO...
Link
@faun shared a link, 2 weeks ago

OpenAI to launch its first AI chip in 2026 with Broadcom, FT reports

OpenAI’s firstin-house AI chipis nearly out of the oven. It’s headed for fabrication atTSMCand built to handle OpenAI’s own workloads—no outside sales, according to theFinancial Times. Why it matters:Big AI shops are going vertical. Custom silicon means tighter control over runtime, reliability, an..

OpenAI to launch its first AI chip in 2026 with Broadcom, FT reports
Link
@faun shared a link, 2 weeks ago

GPT-5 Thinking in ChatGPT (aka Research Goblin) is shockingly good at search

GPT-5's“thinking” modeljust leveled up. It's not just answering queries—it’s doing full-on research. Picture deep, multi-step Bing searches mixed with tool use and reasoning chains. It reads PDFs. Analyzes them. Suggests what to do next. Then actually does it. All from your phone. What’s changing:L..

GPT-5 Thinking in ChatGPT (aka Research Goblin) is shockingly good at search
Link
@faun shared a link, 2 weeks ago

From Zero to GPU: A Guide to Building and Scaling Production-Ready CUDA Kernels

Hugging Face just dropped Kernel Builder—a full-stack toolchain for building, versioning, and shippingcustom CUDA kernels as native PyTorch ops. Kernels arearchitecture-aware,semantically versioned, andpullable straight from the Hub. It tracks changes with lockfiles and bakes inDocker deploysout of..

Link
@faun shared a link, 2 weeks ago

Best Practices for High Availability of LLM Based on AI Gateway

Alibaba Cloud’s AI Gateway just got sharper. It now handlesreal-time overload protectionandLLM fallback routingusing passive health checks, first packet timeouts, and traffic shaping. It proxies both BYO and cloud LLMs—think PAI-EAS, Tongyi Qianwen—and redirects load spikes or failures on the fly. F..

Best Practices for High Availability of LLM Based on AI Gateway
Link
@faun shared a link, 2 weeks ago

The Big LLM Architecture Comparison

Architectures since GPT-2 still ride transformers. They crank memory and performance withRoPE, swapGQAforMLA, sprinkle in sparseMoE, and roll sliding-window attention. Teams shiftRMSNorm. They tweak layer norms withQK-Norm, locking in training stability across modern models. Trend to watch:In 2025,..

The Big LLM Architecture Comparison
Link
@faun shared a link, 2 weeks ago

Hermes V3: Building Swiggy’s Conversational AI Analyst

Swiggy just gave its GenAI tool, Hermes, a serious glow-up. What started as a simple text-to-SQL bot is now acontext-aware AI analystthat lives inside Slack. The upgrade? Not just tweaks—an overhaul. Think: vector-based prompt retrieval, session-level memory, an Agent orchestration layer, and a SQL..

Hermes V3: Building Swiggy’s Conversational AI Analyst
Link
@faun shared a link, 2 weeks ago

Why language models hallucinate

OpenAI sheds light on the persistence ofhallucinationsin language models due to evaluation methods favoring guessing over honesty, requiring a shift towards rewarding uncertainty acknowledgment. High model accuracy does not equate to the eradication of hallucinations, as some questions are inherentl..

Why language models hallucinate
Link
@faun shared a link, 2 weeks ago

Simplifying Large-Scale LLM Processing across Instacart with Maple

Instacart builtMaple, a backend brain for handling millions of LLM prompts—fast, cheap, and shared across teams. It’s not just another service. Maple runs onTemporal,PyArrow, andS3, strip-mines away provider-specific boilerplate, auto-batches prompts, retries failures, and slashes LLM costs by up t..

Simplifying Large-Scale LLM Processing across Instacart with Maple
Story Trending
@laura_garcia shared a post, 2 weeks ago
Software Developer, RELIANOID

RELIANOID Load Balancer Community Edition v7 on AWS using Terraform

🚀 New Guide Available! Learn how to quickly deploy RELIANOID Load Balancer Community Edition v7 on AWS using Terraform. Our step-by-step article shows you how to provision everything automatically — from VPCs and subnets to EC2 and key pairs — in just minutes. 👉 https://www.relianoid.com/resources/k..

Knowledge base Deploy RELIANOID Load Balancer Community Edition v7 with Terraform on AWS
Link
@faun shared a link, 2 weeks ago

Sandboxed to Compromised: New Research Exposes Credential Exfiltration Paths in AWS Code Interpreters

Researchers poked holes insandboxed Bedrock AgentCore code interpreters—and found a way to leak execution role credentials through theMicroVM Metadata Service (MMDS). No outside network? Doesn’t matter. The exploit dodges basic string filters in requests and lets non-agentic code swipe AWS creds to ..

BAO, c’est les chasseurs Tech et Produit qui apportent de la transparence au recrutement. Depuis Paris et Bordeaux, ils ont fait du monde startup leur terrain de jeu en décidant de ne travailler que sur très peu de postes à la fois. Pourquoi ? Parce qu’on voit la chasse comme un sprint dans lequel on travaille main dans la main avec nos clients : nous donnons un maximum de visibilité et de conseil à nos startups partenaires.

En créant BAO, en 2019, Baptiste et Lucas ont décidé de mettre l’écoute au cƓur de leur travail. Curiosité, sourire et empathie sont le seul trait commun à toute l’équipe !

Le bouche-à-oreille est au centre de leur manière de chasser : chaque recruteur entretient son réseau, conscient que la proximité amène à de belles rencontres.
Travailler chez BAO c’est avoir la volonté de rencontrer des personnes aux parcours passionnants et de tisser des liens avec eux. Mais c’est aussi gagner en autonomie tout en profitant d’une équipe dans laquelle les membres s’encouragent mutuellement.

C'est de la vente sans avoir à être agressif, des évolutions rapides au sein d’un écosystème passionnant et un environnement de travail ambitieux sans se prendre trop au sérieux.