Reddit DevOps
266 subscribers
30.9K links
Reddit DevOps. #devops
Thanks @reddit2telegram and @r_channels
Download Telegram
Do DevOps teams at newer companies still choose Terraform for IaC, or native IaC services (like CloudFormation/Bicep)?

Terraform has been the go to for companies with cloud resources across multiple platforms or migrating from onprem, because of its great cross platform support. But for newer startups or organisations starting out in the cloud, I’d say using platform specific IaC services is usually easier than picking up Terraform, and the platform integration is probably better too. Native tools also don’t require installing extra CLIs or managing state files.

If you're at a newer company or helping clients spin up infra, what are you using for IaC? Are platform native tools good enough now, or is Terraform still the default?

https://redd.it/1mc7p46
@r_devops
Free DevOps Tool Developer Experience Audit

I'm offering free developer experience audits specifically focused on DevOps tools.

My background: Helped dyrectorio (deployment orchestration and container management) and Gimlet (GitOps deployment) gain significant GitHub adoption through improved developer onboarding and documentation. Not affiliated with them anymore.

I specialize in identifying friction points in CI/CD pipelines, infrastructure tooling adoption, and developer-facing automation workflows.

What I'll analyze:

Developer onboarding for your DevOps tools
CI/CD pipeline user experience and documentation
Infrastructure-as-code developer workflows
Tool integration friction points

DM me if you'd like an audit of your developer-facing DevOps processes.

https://redd.it/1mc8qna
@r_devops
Problem when fetching image via api gateway

I'm trying to use KrakenD as an api gateway. I have this endpoint on a flask microservice (both the gateway the microservice are conteinerized)

/images/<date>/<hour>/<filename>

When I fetch the image with a direct connection there are no errors. When I use the endpoint on the gateway it gives back a 404 error. This is the endpoint. I have other endpoints but those work.

        {
            "endpoint": "/api/images/{date}/{hour}/{filename}",
            "method": "GET",
            "inputparams": [
                "date",
                "hour",
                "filename"
            ],
            "backend": [
                {
                    "url
pattern": "/images/{date}/{hour}/{filename}",
                    "host":
                        "https://data_processor:8080"
                   
                }
            ]
        }


This is the configuration of the endpoint.

https://redd.it/1mc9nq4
@r_devops
Rollouts

Hello folks,

I want to understand how you guys handles the rollouts.

We are hosting services on Azure.

While rollout, we have few manual changes in app config, kv, DB, etc. and then push services one by one to AKS, how do you handles it, so that everybody will understand different approaches and can implement.

https://redd.it/1mc7v24
@r_devops
If I hear "treat your platform as a product" one more time...

Let's just admit it that we've all been there:
You start with a clean slate. You build a platform tailored perfectly to your org.

Custom pipelines. Custom tooling. A CI/CD “stack” that makes sense to you.

And it works… until it doesn’t.

Suddenly, your internal platform is this black box only you and your team understand.

It’s brittle, hard to onboard new people to, impossible to scale cleanly, and when something breaks, you’re reinventing the wheel again.

We all say things like “our business is unique”, “our scale is different”, “our use case is too complex”. But in reality, the foundations are the same across the board.

https://redd.it/1mcc78f
@r_devops
Tried Jenkins again, was not that bad as I had in mind!

Hi everyone,
as the title says, I gave Jenkins another shot. The last time I used it was at my former company, with a pretty archaic setup: several VMs running Docker Engine, the Docker plugin to spin up workers, and some static servers for on-site deployments in a local datacenter. All of it glued together with some cool Ansible playbooks (still proud of those, ngl). The goal back then was to avoid the classic pet server scenario. If you know me personally, you probably know the company I worked for!

Now I gave it a fresh spin and I approached it with a Kubernetes-first mindset. Deployed everything via Helm charts and used the Kubernetes plugin. And since I like working with Pulumi (and work since then for them), I used that too. You could likely do the same with Terraform and the Kubernetes/Helm provider.

I wrote it all down here: https://www.pulumi.com/blog/jenkins-pulumi-2025-experience/

Any "old" DevOps tech you gave also a new lock/try?

https://redd.it/1mcg2kk
@r_devops
Farewell to my dad

https://blog.mattsbit.co.uk/2025/07/23/dad/

I originally wrote the speach in my blog repo, just for writing purposes.

My dad's funeral was a couple of days ago and wondered, maybe, someone might appreciate it - either because they've lost their dad or it makes them appreciate their dad a little more.
Particularly in this community, as I assume you probably grew up with messing with computers and/or servers and probably had a similar influence from your dads.

https://redd.it/1mcheuv
@r_devops
Is there a proper way to get depot sizes on perforce ?

I wrote a script for our perforce server , but sooner after it crashed our server.
The server was a 4 CPU and 8GB RAM system that was stable. But after running my script it crashed the server (linux) . After our crash I doubled the CPU to 8 and RAM to 16GB .

Still wary of using my script below and asking how perforce admins query depot sizes safely.


depot_sizes.sh
—————————————————

 \#!/bin/bashfor

depot in $(p4 depots | awk '{print $2}'); do   
echo "Depot: $depot"   
p4 sizes //$depot/... | awk '{total += $4} END {print "  Total Size: " total " bytes\\n"}'
done

—————————————————


https://redd.it/1mcherq
@r_devops
DevOps Confessions

Hey guys. just ran into something funny on YouTube, thought you might enjoy it.

Plus, AI videos are terrifying.

https://www.youtube.com/watch?v=Y1xIRAjzTjM

https://redd.it/1mcgq4j
@r_devops
Test your database backups before they fail you in production

Hey devs! 👋

Just shipped BackupGuardian - tired of backup validation tools that only check syntax but don't actually test restoration.

This one spins up Docker containers and actually restores your entire backup to see what breaks. Supports PostgreSQL/MySQL + has a CLI for CI/CD.

Built it after a 3 AM incident where a "validated" backup was missing half the constraints 😅

Demo: https://www.backupguardian.org

Anyone else been burned by "good" backups before?

https://redd.it/1mctf97
@r_devops
Good tip

I came across this tip and couldn't help but share it it's so good and useful

I think we should follow him and help him reach 10 followers

https://x.com/username_husan/status/1950368472258793538?s=46

https://redd.it/1mcwusn
@r_devops
Are these real-time DevOps projects?

I came across this website which has DevOps projects. The author mentioned that those are real-time DevOps projects. There are 7 pages on this website and 37 projects(as of now). Could some experienced DevOps engineers please visit the link below and confirm if these are real-time projects if possible?


https://projects.prodevopsguytech.com/blog

Thanks for your valuable time.

https://redd.it/1mcyeig
@r_devops
Using Vector search for Log monitoring / incident report management?

Hi I wanted to know if anyone in the DevOps community has used vector search / Agentic RAG for performing the following:

🔹 Log monitoring + triage
Some setups use agents to scan logs in real time, highlight anomalies, and even suggest likely root causes based on past patterns. Haven’t tried this myself yet, but sounds promising for reducing alert fatigue.

This agent could help reduce Mean Time to Recovery (MTTR) by analyzing logs, traces, and metrics to suggest root causes and remediation steps. It continuously learns from past incidents to improve future diagnostics.Stores structured incident metadata and unstructured logs as JSON documents. Embeds and indexes logs using Vector Search for similarity-based retrieval. High-throughput data ingestion + sub-millisecond querying for real-time analysis.

One might argue - why do you need a vector database for it? Storing logs as vector doesn't make sense. But I just wanted to see if anyone has a different opinion or even has an open source repository.

Also would love to know if we could use vector search for some other use-case apart from log monitoring - like incident management reporting



https://redd.it/1mczb0a
@r_devops
Project N1netails

# 🧠 Story time:

I started building N1netails after a moment at work that really stuck with me. One of my production support teammates started flipping tables (literally) after getting a Splunk alert 15 minutes too late. By the time we were notified, the issue had already escalated. That experience got me thinking:

I actually like Splunk, but I also think there are some real problems with it:

1. High learning curve – You basically need to take a course just to be productive with Splunk. Because of this, most of our production support folks weren’t using it properly — or even at all.
2. Poor context – I’d get notified by a Splunk alert, but then I had to spend valuable time digging to figure out what actually went wrong. The alert itself wasn’t enough.
3. Query throttling – In big organizations, querying Splunk often means getting throttled. You’re hunting down a bug, and suddenly your queries stop loading. It’s frustrating and slows everything down.
4. Centralization – Again, great for security teams. But as a developer, I just want to be alerted on issues related to my services. Competing for Splunk resources across a large org is overkill if all I want is simple service-level alerting.

So that’s why I built N1netails.

The name comes from two ideas:

* N1 = Think “Big O” notation — O(1), O(n), etc. — but the goal is to get fast, direct insights. N=1.
* ne = Any
* Tails = Like tail -f, watching logs in real-time.

Put it all together and you get N1netails.

The goal? Get notified ASAP when something breaks in the systems that matter to me and my team.

As a developer, I don’t need a full-blown SIEM to monitor the entire company. I just want to know when my stuff is broken — and ideally have some help understanding what happened.

That’s why N1netails includes:

* A prebuilt dashboard (no setup required)
* Stack trace capture
* LLM assistance for debugging (through a helper named Inari)

I also made it easy to self-host. You can check it out here:

* SaaS: [https://app.n1netails.com](https://app.n1netails.com/)
* Docs: [https://n1netails.com](https://n1netails.com/)
* GitHub: [https://github.com/n1netails](https://github.com/n1netails)

Right now, it’s optimized for Java and Spring Boot, but I’m working on expanding support to other languages and platforms.

I know people will probably say, “Why make this? There are tools for this already.” And that’s fair. But I’m building this because I’ve used those tools, and I still believe there’s room for something better — or at least something simpler.

I’m not trying to replace Splunk. N1netails can supplement the tools you already use and help with the day-to-day debugging, triage, and monitoring that’s often overlooked.


N1netails is an open-source project that provides practical alerting and monitoring for applications. If you’re tired of relying on overly complex SIEM tools to identify issues — or if your app lacks alerting altogether — N1netails gives you a straightforward way to get notified when things break.

Thanks for reading. If you want to try it, give feedback, or contribute, check out the repo.

And feel free to leave your hate comments or tell me why you love Splunk. I don’t care. I’m building this because I believe there’s a better way to handle alerts — and I want to help others who feel the same.

https://redd.it/1mczyji
@r_devops
Building a Game

Im looking for a devs to design a game from A to Z. Html based with crypto wallet connection and remote playing. Contact me for more details

https://redd.it/1mcynfo
@r_devops
Why Git Branching Strategy Matters in Database DevOps?

Hey folks,

I've been working a lot with CI/CD and GitOps lately, especially around databases and wanted to share some thoughts on Git branching strategies that often cause more harm than good when managing schema changes across environments.

🔹 The problem:
Most teams use a separate Git branch for each environment (like devqaprod). While it seems structured, it often leads to merge conflicts, missed hotfixes, and environment drift — especially painful in DB deployments where rollback isn’t trivial.

🔹 What works better:
A trunk-based model with a single main branch and declarative promotion through pipelines. Instead of splitting branches per environment, you can use tools  to define environment-specific logic in the changelog itself.

🔹 GitOps and DBs:
Applying GitOps principles to database deployments — version-controlled, auditable, automated via CI/CD, goes a long way toward reducing fragility. Especially in teams scaling fast or operating in regulated environments.

If you're curious, I wrote a deeper blog post that outlines common pitfalls and tactical takeaways:
👉 Choosing the Right Branching Strategy for Database GitOps

Would love to hear how others are managing DB schemas in Git and your experience with GitOps for databases.

https://redd.it/1md2gw1
@r_devops
I built a local AI assistant like ChatGPT that runs 100% offline – No data leaks, no internet, just private intelligence

Hey folks,
I’m a developer and I recently built a **fully offline AI assistant** that you can run on your local computer — kind of like a private Jarvis, but with zero cloud dependencies.

It uses a locally hosted LLM (like LLaMA 3 or Mistral) + offline voice support + a simple desktop interface. You can talk to it, ask it to code, explain stuff, summarize files, or even build dev tools — all **without your data ever leaving your computer**.

# Features:

* 100% local: no internet or cloud access needed
* Can understand and generate code
* Voice input and output (like Jarvis)
* No OpenAI, no Google, no Gemini — just your own secure AI
* Perfect for devs, sysadmins, and companies that banned ChatGPT

I made this because I saw so many developers being blocked from using ChatGPT at work due to **data privacy rules**. This solves that.

Would love to hear what you think — and happy to give early access or even walk you through setting it up.



https://redd.it/1md3zkd
@r_devops
Managing Alpine Linux with Sparrow automation tool

https://asciinema.org/a/730670 - Sparrow is a lightweight alternative to Ansible for operations managing Linux boxes

https://redd.it/1md57ix
@r_devops
DevOps Engineer Interview with Apple

I have an upcoming interview tomorrow for a DevOps position there and would appreciate any tips about the interview process or insights or any topics

https://redd.it/1md7cek
@r_devops
CI/CD pipeline testing with file uploads - how do you generate consistent test data?

Running into an annoying issue with our CI/CD pipeline. We have microservices that handle file processing (image resizing, video transcoding, document parsing), and our tests keep failing inconsistently because of test data problems.

Current setup:

* Tests run in Docker containers
* Need various file types/sizes for boundary testing
* Some tests need exactly 10MB files, others need 100MB+
* Can't commit large binary files to repo (obvs)

What we've tried:

* wget random files from internet (unreliable, different sizes)
* Storing test files in S3 (works but adds external dependency)
* dd commands (creates files but wrong headers/formats)

The S3 approach works but feels heavy for simple unit tests. Plus some environments don't have internet access.

Built a simple solution that generates files in-browser with exact specs:

[https://filemock.com?utm\_source=reddit&utm\_medium=social&utm\_campaign=devops](https://filemock.com?utm_source=reddit&utm_medium=social&utm_campaign=devops)

Now thinking about integrating it into our pipeline with headless Chrome to generate test files on-demand. Anyone done something similar?

How do you handle test file generation in your pipelines? Looking for cleaner approaches that don't require external dependencies or huge repo sizes.

https://redd.it/1md8u5p
@r_devops
"Have you ever done any contributions to open source projects?"

No. I got a family and kids. Welp. Failed that interview.

Anybody got any open source projects I can add two or three features to so I can tick that off my bucket and have something to talk about in interviews?

These things feel like flippin marathons man! So many stages, so many non relevant questions,

https://redd.it/1mdbkh1
@r_devops