ContentPosts from @abhilash_malli..
Link
@faun shared a link, 3 months, 4 weeks ago
FAUN.dev()

How In-Memory Caching Works in Redis

Redis isn’t just a cache anymore. Sure, it still owns the in-memory speed game—with **key expiration**, **data persistence**, and **horizontal scaling** via **replication** and **clustering**. But if you're only using it to stash a few keys, you're missing the point. This thing handles **streams**,.. read more  

How In-Memory Caching Works in Redis
Link
@faun shared a link, 3 months, 4 weeks ago
FAUN.dev()

Experimenting with local LLMs on macOS

Running **open-weight LLMs locally on macOS**? This post breaks it down clean. It compares **llama.cpp**—great for tweaking things—to **LM Studio**, which trades control for simplicity. Covers what fits in memory, which quantized models to grab (hint: 4-bit GGUF), and what’s coming down the pipe: *.. read more  

Experimenting with local LLMs on macOS
Link
@faun shared a link, 3 months, 4 weeks ago
FAUN.dev()

TIOBE Programming Index News September 2025: Perl Regains the Spotlight

Perl 5 has risen to **10th place in the TIOBE Index**, increasing in popularity even though the exact reason is unknown. Perl 6, or Raku, lags behind Perl 5 in rankings and has not seen the same rise in attention. Other top languages like C and Java have experienced slight falls in rankings... read more  

Link
@faun shared a link, 3 months, 4 weeks ago
FAUN.dev()

Building an AI Server on a Budget ($1.3K)

A developer rolled their own AI server for $1.3K—Ubuntu 24.04.2 LTS, an Nvidia RTX GPU, and a sharp eye on Tensor cores, VRAM, and resale value. The rig handles small models locally and punts big jobs to the cloud when needed. Local-first, cloud-when-it-counts... read more  

Building an AI Server on a Budget ($1.3K)
Link
@faun shared a link, 3 months, 4 weeks ago
FAUN.dev()

Using Claude Code to modernize a 25-year-old kernel driver

A long-dead Linux kernel driver for QIC-80 tape drives just got dragged into the present—with help from **Claude Code** and a lot of tinkering. It now builds cleanly and runs as a **standalone module** on **Linux 6.8**, playing nice with modern setups like **Xubuntu 24.04**. **The bigger picture:**.. read more  

Using Claude Code to modernize a 25-year-old kernel driver
Link
@faun shared a link, 3 months, 4 weeks ago
FAUN.dev()

You Vibe It, You Run It?

Vibe Coding lets developers create software by chatting with AI, skipping traditional coding. But the non-determinism of AI prompts poses significant risks for reliability and maintainability, potentially leading to addiction-like dependence on this new tool. Think twice before fully embracing this .. read more  

Link
@faun shared a link, 3 months, 4 weeks ago
FAUN.dev()

Guardians of the Agents 

A new static verification framework wants to make runtime safeguards look lazy. It slaps **mathematical safety proofs** onto LLM-generated workflows *before* they run—no more crossing fingers at execution time. The setup decouples **code from data**, then runs checks with tools like **CodeQL** and .. read more  

Link
@faun shared a link, 3 months, 4 weeks ago
FAUN.dev()

GitHub Copilot on autopilot as community complaints persist

GitHub's biggest debates right now? Whether to shut down AI-generated "noise" fromCopilot—stuff like auto-written issues and code reviews. No clear answers from GitHub yet. Frustration is piling up. Some devs are ditching the platform altogether, shifting their projects toCodebergor spinning upself-.. read more  

GitHub Copilot on autopilot as community complaints persist
Link
@faun shared a link, 3 months, 4 weeks ago
FAUN.dev()

LLM Evaluation: Practical Tips at Booking.com

Booking.com built Judge-LLM, a framework where strong LLMs evaluate other models against a carefully curated golden dataset. Clear metric definitions, rigorous annotation, and iterative prompt engineering make evaluations more scalable and consistent than relying solely on humans. **The takeaway**:.. read more  

Link
@faun shared a link, 3 months, 4 weeks ago
FAUN.dev()

Building Agents for Small Language Models: A Deep Dive into Lightweight AI

Agent engineering with **small language models (SLMs)**—anywhere from 270M to 32B parameters—calls for a different playbook. Think tight prompts, offloaded logic, clean I/O, and systems that don’t fall apart when things go sideways. The newer stack—**GGUF** + **llama.cpp**—lets these agents run loc.. read more Â