Join us

ContentUpdates and recent posts about Pelagia..
Link
@faun shared a link, 1 month, 4 weeks ago

The Big LLM Architecture Comparison

Architectures since GPT-2 still ride transformers. They crank memory and performance withRoPE, swapGQAforMLA, sprinkle in sparseMoE, and roll sliding-window attention. Teams shiftRMSNorm. They tweak layer norms withQK-Norm, locking in training stability across modern models. Trend to watch:In 2025,..

The Big LLM Architecture Comparison
Link
@faun shared a link, 1 month, 4 weeks ago

Hermes V3: Building Swiggy’s Conversational AI Analyst

Swiggy just gave its GenAI tool, Hermes, a serious glow-up. What started as a simple text-to-SQL bot is now acontext-aware AI analystthat lives inside Slack. The upgrade? Not just tweaks—an overhaul. Think: vector-based prompt retrieval, session-level memory, an Agent orchestration layer, and a SQL..

Hermes V3: Building Swiggy’s Conversational AI Analyst
Link
@faun shared a link, 1 month, 4 weeks ago

Simplifying Large-Scale LLM Processing across Instacart with Maple

Instacart builtMaple, a backend brain for handling millions of LLM prompts—fast, cheap, and shared across teams. It’s not just another service. Maple runs onTemporal,PyArrow, andS3, strip-mines away provider-specific boilerplate, auto-batches prompts, retries failures, and slashes LLM costs by up t..

Simplifying Large-Scale LLM Processing across Instacart with Maple
Link
@faun shared a link, 1 month, 4 weeks ago

Why language models hallucinate

OpenAI sheds light on the persistence ofhallucinationsin language models due to evaluation methods favoring guessing over honesty, requiring a shift towards rewarding uncertainty acknowledgment. High model accuracy does not equate to the eradication of hallucinations, as some questions are inherentl..

Why language models hallucinate
Link
@faun shared a link, 1 month, 4 weeks ago

Best Practices for High Availability of LLM Based on AI Gateway

Alibaba Cloud’s AI Gateway just got sharper. It now handlesreal-time overload protectionandLLM fallback routingusing passive health checks, first packet timeouts, and traffic shaping. It proxies both BYO and cloud LLMs—think PAI-EAS, Tongyi Qianwen—and redirects load spikes or failures on the fly. F..

Best Practices for High Availability of LLM Based on AI Gateway
Story
@laura_garcia shared a post, 1 month, 4 weeks ago
Software Developer, RELIANOID

RELIANOID Load Balancer Community Edition v7 on AWS using Terraform

🚀 New Guide Available! Learn how to quickly deploy RELIANOID Load Balancer Community Edition v7 on AWS using Terraform. Our step-by-step article shows you how to provision everything automatically — from VPCs and subnets to EC2 and key pairs — in just minutes. 👉 https://www.relianoid.com/resources/k..

Knowledge base Deploy RELIANOID Load Balancer Community Edition v7 with Terraform on AWS
Link
@faun shared a link, 1 month, 4 weeks ago

Deploy a containerized application with Kamal and Terraform

A Docker-first workflow combinesTerraformandKamalinto a lean, Elastic Beanstalk-ish alternative—without the bloat. Terraform spins up a three-tier VPC and wires it toECR. Kamal takes it from there, booting containers on a raw EC2 box: app, proxy, monitor. One script. Done...

Deploy a containerized application with Kamal and Terraform
Link
@faun shared a link, 1 month, 4 weeks ago

Sandboxed to Compromised: New Research Exposes Credential Exfiltration Paths in AWS Code Interpreters

Researchers poked holes insandboxed Bedrock AgentCore code interpreters—and found a way to leak execution role credentials through theMicroVM Metadata Service (MMDS). No outside network? Doesn’t matter. The exploit dodges basic string filters in requests and lets non-agentic code swipe AWS creds to ..

Link
@faun shared a link, 1 month, 4 weeks ago

AWS, Microsoft and Google unite behind Linux Foundation DocumentDB database to cut enterprise costs and limit vendor lock-in

Document databases are crucial for AI apps in the gen AI era. Microsoft's open-source DocumentDB project, based on PostgreSQL, is moving to the Linux Foundation, offering a vendor-neutral, open-source alternative to MongoDB. DocumentDB's compatibility with MongoDB drivers and open source governance ..

Link
@faun shared a link, 1 month, 4 weeks ago

Measuring Developer Productivity with Amazon Q Developer and Jellyfish

Amazon Q Developer now plugs into Jellyfish. Teams get a clearer view of how AI fits into the real flow of work—prompt usage, code adoption, PR throughput. Not just surface stats. The setup pipes data from AWS S3 straight into Jellyfish’s analytics engine. It tags AI users, tracks velocity gains, an..

Measuring Developer Productivity with Amazon Q Developer and Jellyfish
Pelagia is a Kubernetes controller that provides all-in-one management for Ceph clusters installed by Rook. It delivers two main features:

Aggregates all Rook Custom Resources (CRs) into a single CephDeployment resource, simplifying the management of Ceph clusters.
Provides automated lifecycle management (LCM) of Rook Ceph OSD nodes for bare-metal clusters. Automated LCM is managed by the special CephOsdRemoveTask resource.

It is designed to simplify the management of Ceph clusters in Kubernetes installed by Rook.

Being solid Rook users, we had dozens of Rook CRs to manage. Thus, one day we decided to create a single resource that would aggregate all Rook CRs and deliver a smoother LCM experience. This is how Pelagia was born.

It supports almost all Rook CRs API, including CephCluster, CephBlockPool, CephFilesystem, CephObjectStore, and others, aggregating them into a single specification. We continuously work on improving Pelagia's API, adding new features, and enhancing existing ones.

Pelagia collects Ceph cluster state and all Rook CRs statuses into single CephDeploymentHealth CR. This resource highlights of Ceph cluster and Rook APIs issues, if any.

Another important thing we implemented in Pelagia is the automated lifecycle management of Rook Ceph OSD nodes for bare-metal clusters. This feature is delivered by the CephOsdRemoveTask resource, which automates the process of removing OSD disks and nodes from the cluster. We are using this feature in our everyday day-2 operations routine.