Need Help Integrating AWS ECS Cluster, Service & Task with LGTM Stack using Terraform
So I've been working on Integrating LGTM Stack into my current AWS Infrastructure Stack.
Let me first explain my current work I've done so far,
\######LGTM Infra :
\- Grafana = Using AWS Managed Grafana with Loki, Mimir and Tempo Data Source deployed using Terraform
\- Loki, Tempo and Mimir servers are hosted on EC2 using Docker Compose and using AWS S3 as Backend storage for all three.
\- To push my ECS Task Logs, Metrics and Traces, I've added Side-Cars with current Apps Task Definition which will run alongside with app container and push the data to Loki, Tempo and Mimir servers. For Logs I'm using __aws firelens__ log driver, for Metrics and Traces I'm using Grafana Alloy.
LGTM Server stack is running fine and all three data are being pushed to backend servers, now i'm facing issue with labeling like the metrics and traces are pushed to Mimir and Tempo backend servers but how will i identify from which Cluster, Service and Task i'm getting these Logs, Metrics and Traces.
For logs it was straight forward since i was using AWS Firelens log driver, the code was like this:
>log_configuration = {
>logDriver = "awsfirelens"
>options = {
>"Name" = "grafana-loki"
>"Url" = "${var.loki_endpoint}/loki/api/v1/push"
>"Labels" = "{job=\\"firelens\\"}"
>"RemoveKeys" = "ecs_task_definition,source,ecs_task_arn"
>"LabelKeys" = "container_id,container_name,ecs_cluster",
>"LineFormat" = "key_value"
>}
>}
as you can see in the below screenshots, ecs related details are getting populated on grafana,
: https://i.postimg.cc/HspwKRVW/loki.png
and for the same i was able to create dashboard as well with some basic filtering and search box,
: https://i.postimg.cc/tT36vNbV/loki-dashboard.png
Now comes the Metrics a.k.a Mimir part:
for this i used Grafana Alloy, and used below config.alloy config file:
>prometheus.exporter.unix "local_system" { }
>
>prometheus.scrape "scrape_metrics" {
>targets = prometheus.exporter.unix.local_system.targets
>forward_to = [prometheus.relabel.add_ecs_labels.receiver\]
>scrape_interval = "10s"
>}
>
>remote.http "ecs_metadata" {
>url = "ECS_METADATA_URI"
>}
>
>prometheus.relabel "add_ecs_labels" {
>rule {
>source_labels = ["__address__"\]
>target_label = "ecs_cluster_name"
>regex = "(.*)"
>replacement = "ECS_CLUSTER_NAME"
>}
>
>rule {
>source_labels = ["__address__"\]
>target_label = "ecs_service_name"
>regex = "(.*)"
>replacement = "ECS_SERVICE_NAME"
>}
>
>rule {
>source_labels = ["__address__"\]
>target_label = "ecs_container_name"
>regex = "(.*)"
>replacement = "ECS_CONTAINER_NAME"
>}
>
>forward_to = [prometheus.remote_write.metrics_service.receiver\]
>}
>
>prometheus.remote_write "metrics_service" {
>endpoint {
>url = "${local.mimir_endpoint}/api/v1/push"
>headers = {
>"X-Scope-OrgID" = "staging",
>}
>}
>}
I used AWS to create this config in Param store and added another app task side car which will load this config file, run a custom script which will fetch the ECS Cluster name from ECS_CONTAINER_METADATA_URI_V4 and passed Service Name and Container Name as ECS Task Definition Environment Variable.
so after all this, I was able to do the relabeling and populate the Cluster, Service and Task name on Mimir Data Source:
: https://i.postimg.cc/Gh8LchBX/mimir.png
Now when I was trying to use Node\_Exporter\_Full Grafana dashboard for the metrics, I was getting the metrics but for unix level filtering only,
: https://i.postimg.cc/Jn0wPPZp/mimir-dashboard-1.png
:
So I've been working on Integrating LGTM Stack into my current AWS Infrastructure Stack.
Let me first explain my current work I've done so far,
\######LGTM Infra :
\- Grafana = Using AWS Managed Grafana with Loki, Mimir and Tempo Data Source deployed using Terraform
\- Loki, Tempo and Mimir servers are hosted on EC2 using Docker Compose and using AWS S3 as Backend storage for all three.
\- To push my ECS Task Logs, Metrics and Traces, I've added Side-Cars with current Apps Task Definition which will run alongside with app container and push the data to Loki, Tempo and Mimir servers. For Logs I'm using __aws firelens__ log driver, for Metrics and Traces I'm using Grafana Alloy.
LGTM Server stack is running fine and all three data are being pushed to backend servers, now i'm facing issue with labeling like the metrics and traces are pushed to Mimir and Tempo backend servers but how will i identify from which Cluster, Service and Task i'm getting these Logs, Metrics and Traces.
For logs it was straight forward since i was using AWS Firelens log driver, the code was like this:
>log_configuration = {
>logDriver = "awsfirelens"
>options = {
>"Name" = "grafana-loki"
>"Url" = "${var.loki_endpoint}/loki/api/v1/push"
>"Labels" = "{job=\\"firelens\\"}"
>"RemoveKeys" = "ecs_task_definition,source,ecs_task_arn"
>"LabelKeys" = "container_id,container_name,ecs_cluster",
>"LineFormat" = "key_value"
>}
>}
as you can see in the below screenshots, ecs related details are getting populated on grafana,
: https://i.postimg.cc/HspwKRVW/loki.png
and for the same i was able to create dashboard as well with some basic filtering and search box,
: https://i.postimg.cc/tT36vNbV/loki-dashboard.png
Now comes the Metrics a.k.a Mimir part:
for this i used Grafana Alloy, and used below config.alloy config file:
>prometheus.exporter.unix "local_system" { }
>
>prometheus.scrape "scrape_metrics" {
>targets = prometheus.exporter.unix.local_system.targets
>forward_to = [prometheus.relabel.add_ecs_labels.receiver\]
>scrape_interval = "10s"
>}
>
>remote.http "ecs_metadata" {
>url = "ECS_METADATA_URI"
>}
>
>prometheus.relabel "add_ecs_labels" {
>rule {
>source_labels = ["__address__"\]
>target_label = "ecs_cluster_name"
>regex = "(.*)"
>replacement = "ECS_CLUSTER_NAME"
>}
>
>rule {
>source_labels = ["__address__"\]
>target_label = "ecs_service_name"
>regex = "(.*)"
>replacement = "ECS_SERVICE_NAME"
>}
>
>rule {
>source_labels = ["__address__"\]
>target_label = "ecs_container_name"
>regex = "(.*)"
>replacement = "ECS_CONTAINER_NAME"
>}
>
>forward_to = [prometheus.remote_write.metrics_service.receiver\]
>}
>
>prometheus.remote_write "metrics_service" {
>endpoint {
>url = "${local.mimir_endpoint}/api/v1/push"
>headers = {
>"X-Scope-OrgID" = "staging",
>}
>}
>}
I used AWS to create this config in Param store and added another app task side car which will load this config file, run a custom script which will fetch the ECS Cluster name from ECS_CONTAINER_METADATA_URI_V4 and passed Service Name and Container Name as ECS Task Definition Environment Variable.
so after all this, I was able to do the relabeling and populate the Cluster, Service and Task name on Mimir Data Source:
: https://i.postimg.cc/Gh8LchBX/mimir.png
Now when I was trying to use Node\_Exporter\_Full Grafana dashboard for the metrics, I was getting the metrics but for unix level filtering only,
: https://i.postimg.cc/Jn0wPPZp/mimir-dashboard-1.png
:
postimg.cc
loki — Postimages
https://i.postimg.cc/mD5vqCSB/mimir-dashboard-filter.png
so i did some dashboard JSON filtering and was able to get ECS Cluster Name, ECS Service Name & ECS Container Name for the same dashboard,
: https://i.postimg.cc/2yLsfyHv/mimir-dashboard-2.png
but now I'm not able to get the metrics on dashboard,
It's been only 2 Weeks since I've started the Observability and before that i didn't know much about these apart from the term Observability so i might be doing something wrong with the Metrics for my Custom Node Exporter Dashboard.
Do I need to relabel the exisitng labels like __job__ and __host__ and replace them with my added labels like ECS Service or Container Names to fetch the metrics on the basis of ECS Containers?
Since i'm doing this for the first time so not sure much about this.
If anyone here has done something like same, can you please help me with this implementation??
Next thing once this is done then I'll be going for like aggregated metrics based on ECS Services since there might be more than one task running for one ecs services and then i believe i'll be needing the something like same relabeling for tempo traces as well.
Please help me guys for this.
Thank you!!!
https://redd.it/1iheu0p
@r_devops
so i did some dashboard JSON filtering and was able to get ECS Cluster Name, ECS Service Name & ECS Container Name for the same dashboard,
: https://i.postimg.cc/2yLsfyHv/mimir-dashboard-2.png
but now I'm not able to get the metrics on dashboard,
It's been only 2 Weeks since I've started the Observability and before that i didn't know much about these apart from the term Observability so i might be doing something wrong with the Metrics for my Custom Node Exporter Dashboard.
Do I need to relabel the exisitng labels like __job__ and __host__ and replace them with my added labels like ECS Service or Container Names to fetch the metrics on the basis of ECS Containers?
Since i'm doing this for the first time so not sure much about this.
If anyone here has done something like same, can you please help me with this implementation??
Next thing once this is done then I'll be going for like aggregated metrics based on ECS Services since there might be more than one task running for one ecs services and then i believe i'll be needing the something like same relabeling for tempo traces as well.
Please help me guys for this.
Thank you!!!
https://redd.it/1iheu0p
@r_devops
postimg.cc
mimir dashboard filter — Postimages
Linux Server which can run Virtualbox for a month, where to go ? EU
Customer's client provided me a dev environment based on Vagrant. I'm not looking for alternatives for that, it's the way it is. That vagrant is running k3s. I tried with my old Intel MB Pro but I'm lacking memory. I need a server which can run Virtualbox, and with a short contract, max 2 months. Where should I go ?
Hope this post is ok with Mods, asking for vendors.
https://redd.it/1ihfug0
@r_devops
Customer's client provided me a dev environment based on Vagrant. I'm not looking for alternatives for that, it's the way it is. That vagrant is running k3s. I tried with my old Intel MB Pro but I'm lacking memory. I need a server which can run Virtualbox, and with a short contract, max 2 months. Where should I go ?
Hope this post is ok with Mods, asking for vendors.
https://redd.it/1ihfug0
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Cannot reach service by node ip and port from browser
I'm running Docker Desktop on a Windows 11 PC. I want to try the built-in Kubernetes based on Kind. It works, although I cannot reach the service by node ip and port. I tested the connection inside the cluster it works fine. I also tried disabling firewalls. When I tried Minikube with Hyper V driver it worked fine, using the docker driver gave me the same problems like Kind has. How to solve this?
https://redd.it/1ihhe59
@r_devops
I'm running Docker Desktop on a Windows 11 PC. I want to try the built-in Kubernetes based on Kind. It works, although I cannot reach the service by node ip and port. I tested the connection inside the cluster it works fine. I also tried disabling firewalls. When I tried Minikube with Hyper V driver it worked fine, using the docker driver gave me the same problems like Kind has. How to solve this?
https://redd.it/1ihhe59
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
I built an AI agent for website monitoring - looking for feedback
Hey everyone, I wanted to share [https://flowtest.ai/](https://flowtest.ai/), a product my 2 friends and I are working on. We’d love to hear your feedback and opinions.
Everything started, when we discovered that LLMs can be really good at browsing websites simply by following a chatGPT-like prompt. So, we built LLM agent and gave it tools like keyboard & mouse control. We parse the website and agent does actions you prompt it to do. This opens lots of opportunities for website monitoring and testing. It’s also a great alternative to Pingdom.
Instead of just pinging a website, you can now prompt an AI agent to visit and interact with a website as a real user. Even if the website is up, agent can identify other issues and immediately alert you if certain elements aren't functioning correctly e.g. 3rd party app crashes or features fail to load.
Once you set a frequency for the agent to run its monitoring flow, it will actually visit your website each time. LLMs are now smart enough and combined with our web parsing, if some web elements change, agent will adapt without asking your help.
**Here are a few more complex examples of how our first customers are using it:**
* Agent visits your site, enters a keyword in a search box, and verifies that relevant search results appear.
* Agent visits your login page, enters credentials, and confirms successful login into the correct account.
* Agent completes a purchasing flow by filling in all necessary fields and checks if the checkout process works correctly.
We initially launched it as a quality assurance testing automation agent but noticed that our early customers use it more as a website uptime monitoring service.
We offer 7 days free trial (no cc required), but if you’d like to try it for a longer period, just DM me, and I'll give you a month free of charge in exchange for your feedback.
We’d love to hear all your feedback and opinions.
https://redd.it/1ihhv45
@r_devops
Hey everyone, I wanted to share [https://flowtest.ai/](https://flowtest.ai/), a product my 2 friends and I are working on. We’d love to hear your feedback and opinions.
Everything started, when we discovered that LLMs can be really good at browsing websites simply by following a chatGPT-like prompt. So, we built LLM agent and gave it tools like keyboard & mouse control. We parse the website and agent does actions you prompt it to do. This opens lots of opportunities for website monitoring and testing. It’s also a great alternative to Pingdom.
Instead of just pinging a website, you can now prompt an AI agent to visit and interact with a website as a real user. Even if the website is up, agent can identify other issues and immediately alert you if certain elements aren't functioning correctly e.g. 3rd party app crashes or features fail to load.
Once you set a frequency for the agent to run its monitoring flow, it will actually visit your website each time. LLMs are now smart enough and combined with our web parsing, if some web elements change, agent will adapt without asking your help.
**Here are a few more complex examples of how our first customers are using it:**
* Agent visits your site, enters a keyword in a search box, and verifies that relevant search results appear.
* Agent visits your login page, enters credentials, and confirms successful login into the correct account.
* Agent completes a purchasing flow by filling in all necessary fields and checks if the checkout process works correctly.
We initially launched it as a quality assurance testing automation agent but noticed that our early customers use it more as a website uptime monitoring service.
We offer 7 days free trial (no cc required), but if you’d like to try it for a longer period, just DM me, and I'll give you a month free of charge in exchange for your feedback.
We’d love to hear all your feedback and opinions.
https://redd.it/1ihhv45
@r_devops
flowtest.ai
Flowtest.ai - AI Agent for website uptime monitoring.
Flowtest.ai is an AI Agent for website uptime monitoring.
Get website issues alerts from AI Agent that continuously browse & monitor your website like a real user.
Get website issues alerts from AI Agent that continuously browse & monitor your website like a real user.
Alternatives to Yor
Looks like Yor (https://github.com/bridgecrewio/yor) is not really active anymore. Last PR was over 7 months ago and no releases since August 24. Their slack is pretty dead as well.
Most PRs are closed without comment.
So is anyone aware of an alternative?
https://redd.it/1ihjvyq
@r_devops
Looks like Yor (https://github.com/bridgecrewio/yor) is not really active anymore. Last PR was over 7 months ago and no releases since August 24. Their slack is pretty dead as well.
Most PRs are closed without comment.
So is anyone aware of an alternative?
https://redd.it/1ihjvyq
@r_devops
GitHub
GitHub - bridgecrewio/yor: Extensible auto-tagger for your IaC files. The ultimate way to link entities in the cloud back to the…
Extensible auto-tagger for your IaC files. The ultimate way to link entities in the cloud back to the codified resource which created it. - bridgecrewio/yor
Hyperping vs. Better Stack vs. OneUptime for observability
Which one is better? Pricing is not the problem.
I am specifically interested in synthetic monitoring with playwright.
https://redd.it/1ihkrew
@r_devops
Which one is better? Pricing is not the problem.
I am specifically interested in synthetic monitoring with playwright.
https://redd.it/1ihkrew
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Looking to get back into a DevOps role.
Looking for any tips on what I need to focus on when interviewing. I’ve worked in IT for 20+ years, I’ve been a team lead on Linux and Virtualization teams, have worked with most automation tools, and have sold some of these products. It’s been a while since I’ve sat in this role and looking for help on anything new in the market, what I should focus on now and what to expect from the interviews.
https://redd.it/1ihm7vc
@r_devops
Looking for any tips on what I need to focus on when interviewing. I’ve worked in IT for 20+ years, I’ve been a team lead on Linux and Virtualization teams, have worked with most automation tools, and have sold some of these products. It’s been a while since I’ve sat in this role and looking for help on anything new in the market, what I should focus on now and what to expect from the interviews.
https://redd.it/1ihm7vc
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Best way to sync a private GitHub repo to a shared remote machine without shared credentials?
My team and I have a remote desktop machine connected to a PLC, conveyor belt, and sensors. We need to clone and pull updates from our private GitHub repository to this machine. However, we’re stuck on how to do this efficiently without creating a shared user account on the machine (which would require sharing credentials).
Here’s the issue:
\- We can’t create a GitHub account for the machine because it doesn’t have an official organization email.
\- Sharing a single user account on the machine isn’t ideal and goes against best practices.
\- We need to be able to:
\- Clone and pull the latest changes to the machine.
\- Push changes made on the remote machine back to the repo using our individual GitHub credentials.
**Options we’re considering:**
1. Use tools like TeamViewer or SSH tunnels to transfer files between our local machines (which are already set up) and the remote machine.
2. Set up GitHub on the remote machine but deal with the inefficiency of constantly asking for user credentials to push changes.
What’s the best practice here? Are there tools or workflows (deploy keys, GitHub Actions?) designed for this kind of scenario? Any advice or recommendations would be greatly appreciated!
https://redd.it/1ihjvoj
@r_devops
My team and I have a remote desktop machine connected to a PLC, conveyor belt, and sensors. We need to clone and pull updates from our private GitHub repository to this machine. However, we’re stuck on how to do this efficiently without creating a shared user account on the machine (which would require sharing credentials).
Here’s the issue:
\- We can’t create a GitHub account for the machine because it doesn’t have an official organization email.
\- Sharing a single user account on the machine isn’t ideal and goes against best practices.
\- We need to be able to:
\- Clone and pull the latest changes to the machine.
\- Push changes made on the remote machine back to the repo using our individual GitHub credentials.
**Options we’re considering:**
1. Use tools like TeamViewer or SSH tunnels to transfer files between our local machines (which are already set up) and the remote machine.
2. Set up GitHub on the remote machine but deal with the inefficiency of constantly asking for user credentials to push changes.
What’s the best practice here? Are there tools or workflows (deploy keys, GitHub Actions?) designed for this kind of scenario? Any advice or recommendations would be greatly appreciated!
https://redd.it/1ihjvoj
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Learning GCP and Terraform at the same time?
I'm confident at frontend development. I know the basics of Node and Postgres. But I'm weak on DevOps.
I've traditionally been a freelancer and used tools like Vercel and Supabase. However now I have a job with a startup and I need to learn GCP.
I've only spent 1/2 a day but I find using Google Cloud's website and `gcloud` in the terminal quite awkward.
Does it make sense to use something like Terraform from the start? I like the idea of a code-first approach, and being able to switch providers in the future is also nice (we're on GCP as we got a bunch of free credits).
https://redd.it/1ihpl7g
@r_devops
I'm confident at frontend development. I know the basics of Node and Postgres. But I'm weak on DevOps.
I've traditionally been a freelancer and used tools like Vercel and Supabase. However now I have a job with a startup and I need to learn GCP.
I've only spent 1/2 a day but I find using Google Cloud's website and `gcloud` in the terminal quite awkward.
Does it make sense to use something like Terraform from the start? I like the idea of a code-first approach, and being able to switch providers in the future is also nice (we're on GCP as we got a bunch of free credits).
https://redd.it/1ihpl7g
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
How Much Do You Spend on Databases? (2-Min Survey)
Hey all,
We’re doing a quick research study on **database costs & infrastructure**—figuring out how developers & companies use PostgreSQL, InfluxDB, ClickHouse, and managed DBaaS.
**Common problems we hear:**
* 💸 AWS RDS costs way more than expected
* 😩 Managing **high availability & scaling** is painful
* 🔗 Vendor lock-in sucks
🔥 If you run databases, we’d **love your insights!**
👉 **Survey Link (2 mins, no email required):** [https://app.formbricks.com/s/cm6r296dm0007l203s8953ph4](https://app.formbricks.com/s/cm6r296dm0007l203s8953ph4)
(Results will be shared back with the community!)
https://redd.it/1ihvket
@r_devops
Hey all,
We’re doing a quick research study on **database costs & infrastructure**—figuring out how developers & companies use PostgreSQL, InfluxDB, ClickHouse, and managed DBaaS.
**Common problems we hear:**
* 💸 AWS RDS costs way more than expected
* 😩 Managing **high availability & scaling** is painful
* 🔗 Vendor lock-in sucks
🔥 If you run databases, we’d **love your insights!**
👉 **Survey Link (2 mins, no email required):** [https://app.formbricks.com/s/cm6r296dm0007l203s8953ph4](https://app.formbricks.com/s/cm6r296dm0007l203s8953ph4)
(Results will be shared back with the community!)
https://redd.it/1ihvket
@r_devops
Formbricks
Start from scratch
Thanks a lot for your time 🙏
What do you think of the idea your life partner is a DevOps engineer as well?
Any thoughts? Cons pros ?
https://redd.it/1ihx10i
@r_devops
Any thoughts? Cons pros ?
https://redd.it/1ihx10i
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Roast my Resume!
Ok so, my contract at a startup ended a few weeks ago and I'm currently looking for my next adventure, the DevOps job market has been really tough lately, I've sent lots of CVs here and there, so far I made it to a couple interviews, second interviews, technical interviews but then I'm left in radio silence or the position just closes.
I come from a system administration and infrastructure background, not really much of a coder other than python automation scripts, terraform files and CI/CD configuration files.
I do not consider myself a senior engineer, I'm more like intermediate-level.
Roast my resume and let me know what should I improve/learn next in my career.
https://imgur.com/a/Y0ukRkM
Edit: I'm not from the US, so that could make it a little bit harder to get hired.
https://redd.it/1ihvbxm
@r_devops
Ok so, my contract at a startup ended a few weeks ago and I'm currently looking for my next adventure, the DevOps job market has been really tough lately, I've sent lots of CVs here and there, so far I made it to a couple interviews, second interviews, technical interviews but then I'm left in radio silence or the position just closes.
I come from a system administration and infrastructure background, not really much of a coder other than python automation scripts, terraform files and CI/CD configuration files.
I do not consider myself a senior engineer, I'm more like intermediate-level.
Roast my resume and let me know what should I improve/learn next in my career.
https://imgur.com/a/Y0ukRkM
Edit: I'm not from the US, so that could make it a little bit harder to get hired.
https://redd.it/1ihvbxm
@r_devops
Imgur
Roast my Resume - Album on Imgur
Discover the magic of the internet at Imgur, a community powered entertainment destination. Lift your spirits with funny jokes, trending memes, entertaining gifs, inspiring stories, viral videos, and so much more from users like driconrikus.
Helping folks understand what EKS Auto Mode is and how to set it up in Terraform
Hi all. Purely as an educational tool for knowledge sharing, I created an all-in-one repo that deploys a Kubernetes cluster using Auto Mode enabled, and shows how Auto Mode is different than not using Auto Mode.
https://github.com/setheliot/eks\_auto\_mode
Let me know what you think of it? PR to improve it are always welcome.
https://redd.it/1ii2wx4
@r_devops
Hi all. Purely as an educational tool for knowledge sharing, I created an all-in-one repo that deploys a Kubernetes cluster using Auto Mode enabled, and shows how Auto Mode is different than not using Auto Mode.
https://github.com/setheliot/eks\_auto\_mode
Let me know what you think of it? PR to improve it are always welcome.
https://redd.it/1ii2wx4
@r_devops
GitHub
GitHub - setheliot/eks_auto_mode: Demonstrating how to setup AWS EKS Auto Mode
Demonstrating how to setup AWS EKS Auto Mode. Contribute to setheliot/eks_auto_mode development by creating an account on GitHub.
How do you handle downtime communication with your users?
Hey everyone,
I’m working on a simple tool that helps businesses display custom messages directly in their apps when things go wrong (e.g., API failures, server downtime) aka incidents.
The idea is to let you quickly toggle a message to your users without relying on a full status page or complex integrations. The main point is that this error status message is hosted outside your apps infrastructure and your client side can easily check the api regularly to see if there is any need to display the message.
If you run a SaaS, fintech, or e-commerce platform, I’d love to hear your thoughts:
How do you currently notify users when your services are down?
Would something like this be useful for your business?
Looking forward to hearing your feedback!
https://redd.it/1iia4k9
@r_devops
Hey everyone,
I’m working on a simple tool that helps businesses display custom messages directly in their apps when things go wrong (e.g., API failures, server downtime) aka incidents.
The idea is to let you quickly toggle a message to your users without relying on a full status page or complex integrations. The main point is that this error status message is hosted outside your apps infrastructure and your client side can easily check the api regularly to see if there is any need to display the message.
If you run a SaaS, fintech, or e-commerce platform, I’d love to hear your thoughts:
How do you currently notify users when your services are down?
Would something like this be useful for your business?
Looking forward to hearing your feedback!
https://redd.it/1iia4k9
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Wednesday humor? Ops-abet Soup (Like alphabet soup)
During my last meeting someone mentioned in stand-up...
BreakOps
ClickOps
DevOps
EphemeralOps
HelpdeskOps
RebootOps
What else can we add to this list? lol
https://redd.it/1iichz0
@r_devops
During my last meeting someone mentioned in stand-up...
BreakOps
ClickOps
DevOps
EphemeralOps
HelpdeskOps
RebootOps
What else can we add to this list? lol
https://redd.it/1iichz0
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Survey on multi factor authentication in cloud for academic
Hi everyone! 🙋♂🙋♀
I am working on a study about - usability and usage of Multi Factor Authentication systems in cloud and I would love your support. Your honest feedback will help us collect real data for meaningful analysis.
📝 The survey is quick and easy – it will only take 5-7 minutes!
👉 Click here to participate: https://forms.gle/qeszv1JER6TtEA5W9
Your participation means a lot to us. Thank you for your time and support!
https://redd.it/1iic05x
@r_devops
Hi everyone! 🙋♂🙋♀
I am working on a study about - usability and usage of Multi Factor Authentication systems in cloud and I would love your support. Your honest feedback will help us collect real data for meaningful analysis.
📝 The survey is quick and easy – it will only take 5-7 minutes!
👉 Click here to participate: https://forms.gle/qeszv1JER6TtEA5W9
Your participation means a lot to us. Thank you for your time and support!
https://redd.it/1iic05x
@r_devops
Google Docs
MFA QUESTIONNAIRES SURVEY
Hello guys, i am working on a thesis as part of my academic work. My research looks into the usability and usage of multi factor authentication systems in cloud among people of different ages. How it affects them. Now i am counting on you guys to help me…
Brainstorming Ideas for University Final Year Project
Hello everyone,
I'm about to start my senior year at university and want to pursue a final-year project that blends DevOps and AI, (most likely a dev tool). With 2 years of experience in DevOps, my skill set includes automation, observability, IaC, Kubernetes, containerization, AWS, Azure, with strong emphasis on security.
I'm looking for project ideas that align with these strengths and also resonate with recent Y Combinator's Request for Startups. Would love to hear thoughts from this amazing sub, which has been super helpful throughout my career.
Thank you!
https://redd.it/1iig1wy
@r_devops
Hello everyone,
I'm about to start my senior year at university and want to pursue a final-year project that blends DevOps and AI, (most likely a dev tool). With 2 years of experience in DevOps, my skill set includes automation, observability, IaC, Kubernetes, containerization, AWS, Azure, with strong emphasis on security.
I'm looking for project ideas that align with these strengths and also resonate with recent Y Combinator's Request for Startups. Would love to hear thoughts from this amazing sub, which has been super helpful throughout my career.
Thank you!
https://redd.it/1iig1wy
@r_devops
Y Combinator
Requests for Startups | Y Combinator
Y Combinator is looking for startups working on these ideas. If you're working on something we're interested in, we'd love to hear from you.
Starting a Systems Specialist role Monday...any advice?
Hey all!
Hope everyone is well. So I am starting a new job this coming Monday. Title is 'System Specialist'. Currently a Junior SysAdmin at a health organization, dealing with a lot of Azure, O365, AD, Veeam, Wasabi administration, along with other technologies.
My new job is at a pharmaceutical 'start-up' but it appears they're successful enough to be past the 'start-up' stage. This job I've been told deals with Azure/AD for directory usage, and a bunch of Linux/AWS/Docker stuff with other projects. Super excited for the opportunity, as I've been trying to get into more Cloud/DevOps stuff the past year or so. Also a 60% salary increase compared to my current role.
My last day at my current employer is tomorrow, taking Friday off to clear the head and be fresh for Monday.
Pretty bittersweet, aside from the obvious pay upgrade and upgrade in technologies used, as I really enjoyed working with my IT team at my current place.
Anything I should be worried about? Any advice adjusting to the new job? Or just any advice in general? Getting first day anxiety at this point lol, just want to get started already.
Thanks all and have a blessed weekend!
https://redd.it/1iifho0
@r_devops
Hey all!
Hope everyone is well. So I am starting a new job this coming Monday. Title is 'System Specialist'. Currently a Junior SysAdmin at a health organization, dealing with a lot of Azure, O365, AD, Veeam, Wasabi administration, along with other technologies.
My new job is at a pharmaceutical 'start-up' but it appears they're successful enough to be past the 'start-up' stage. This job I've been told deals with Azure/AD for directory usage, and a bunch of Linux/AWS/Docker stuff with other projects. Super excited for the opportunity, as I've been trying to get into more Cloud/DevOps stuff the past year or so. Also a 60% salary increase compared to my current role.
My last day at my current employer is tomorrow, taking Friday off to clear the head and be fresh for Monday.
Pretty bittersweet, aside from the obvious pay upgrade and upgrade in technologies used, as I really enjoyed working with my IT team at my current place.
Anything I should be worried about? Any advice adjusting to the new job? Or just any advice in general? Getting first day anxiety at this point lol, just want to get started already.
Thanks all and have a blessed weekend!
https://redd.it/1iifho0
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Integrating Fastlane Match with Azure Storage Container – Has Anyone Done This?
Hello everyone!
I've recently started automating iOS certificate and profile management at my new job using Fastlane Match, and it's been functioning quite well from an Azure DevOps Repo setup. I had done something like this at a previous job I worked at as well.
However, my superior wants to migrate this setup to Azure, possibly utilizing an Azure Storage Container. He specifically doesn't want it to run out of a repository even though it would be encrypted. From what I've researched, Fastlane Match's documentation suggests compatibility with Google Cloud Storage or an Amazon S3 bucket, but Azure is not mentioned.
I'm contemplating whether anyone in this community has experience or has tackled integration with Azure. Building custom scripts might be an option, but before I head down that route, I wanted to reach out and see if anyone else has successfully implemented this or explored the possibility.
Any insights or recommendations would be highly appreciated. Looking forward to your valuable input on this!
https://redd.it/1iij4p8
@r_devops
Hello everyone!
I've recently started automating iOS certificate and profile management at my new job using Fastlane Match, and it's been functioning quite well from an Azure DevOps Repo setup. I had done something like this at a previous job I worked at as well.
However, my superior wants to migrate this setup to Azure, possibly utilizing an Azure Storage Container. He specifically doesn't want it to run out of a repository even though it would be encrypted. From what I've researched, Fastlane Match's documentation suggests compatibility with Google Cloud Storage or an Amazon S3 bucket, but Azure is not mentioned.
I'm contemplating whether anyone in this community has experience or has tackled integration with Azure. Building custom scripts might be an option, but before I head down that route, I wanted to reach out and see if anyone else has successfully implemented this or explored the possibility.
Any insights or recommendations would be highly appreciated. Looking forward to your valuable input on this!
https://redd.it/1iij4p8
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Effective Usage of AI Code Reviewers on GitHub
The article discusses the effective use of AI code reviewers on GitHub, highlighting their role in enhancing the code review process within software development: How to Effectively Use AI Code Reviewers on GitHub
It outlines the traditional manual code review process, emphasizing its importance in maintaining coding standards, identifying vulnerabilities, and ensuring architectural integrity.
https://redd.it/1iih1vj
@r_devops
The article discusses the effective use of AI code reviewers on GitHub, highlighting their role in enhancing the code review process within software development: How to Effectively Use AI Code Reviewers on GitHub
It outlines the traditional manual code review process, emphasizing its importance in maintaining coding standards, identifying vulnerabilities, and ensuring architectural integrity.
https://redd.it/1iih1vj
@r_devops
Qodo
How to Effectively Use AI Code Reviewers on GitHub | Qodo
AI code reviewers can analyze code for potential bugs. Sometimes, they can capture bugs that humans would miss.