No calls in DevOps despite 1.5 YOE feels low
Hi everyone 🙌
I am having 1.5 years experience in AWS DevOps, applied to so many company still no calls. My current company don’t have good clients and projects. Help me to land a job what can u do in my cv ? Is my cv that bad for DevOps role 🙂↕️.
Here is my CV 👇
Technical Skills
Tools: Ansible, Docker, Kubernetes, Terraform(IaC), Maven, CICD(Jenkins), Argo CD, Git & Git-Hub, ELK(Elastic, Logstash,
Kibana), Prometheus & Grafana
Scripting Languages: Bash & Python
AWS Services: AWS Route 53, EKS, IAM, RDS, DynamoDB, ASG, CloudWatch, SNS, S3, AWS Lambda, EC2
Experience
Loomtex Exports July 2022- May 2023
- Managment trainee
Fusion5 August 2023 – Present
- Junior DevOps Engineer Hyderabad, India
Projects
Multi-tier web application
• Deployed a 3-tier application (Front-end, Back-end, Database) on an EKS cluster using Terraform for IaC.
• Configured CI/CD pipelines using Jenkins, integrated with SonarQube for code quality and Nexus for artifact storage.
• Automated environment setup with Ansible playbooks for Jenkins, SonarQube, and Nexus.
• Managed source code with GitHub and utilized Maven and NodeJS for building and packaging application artifacts.
• Enhanced project security using SonarQube and Trivy to detect and mitigate vulnerabilities.
• Built Docker images, stored them on Docker Hub, and set up comprehensive monitoring for system and website
metrics.
Microservice application
• Engineered the implementation of EKS via Terraform and configured Jenkins and SonarQube using Ansible, boosting
deployment efficiency by 35%.
• Established 12 different Jenkins multibranch pipelines to streamline CI/CD processes.
• Implemented Webhooks to increase automation and minimize manual work.
• Created and integrated application components using build tools specified in the pipeline.
• Utilized Docker to create images, transferring them to the Docker registry, and employed Trivy for enhanced security.
• Launched the application on an EKS cluster, using Prometheus and Grafana for performance tracking, achieving a
30% increase in uptime and improving resource utilization by 20%.
AWS Cost Optimization
• Developed and deployed an automated solution using Boto3 and AWS Lambda to remove obsolete EBS snapshots
exceeding 30 days, reducing storage expenses by 20%.
Certifications
AWS Certified Cloud Practitioner Dec 2023 - Dec 2026
Education
Institute of Chemical Technology Bachelor of Technology in Fibers and Textile Processing Technology June 2022
Mumbai, Maharashtra
https://redd.it/1g3aubb
@r_devops
Hi everyone 🙌
I am having 1.5 years experience in AWS DevOps, applied to so many company still no calls. My current company don’t have good clients and projects. Help me to land a job what can u do in my cv ? Is my cv that bad for DevOps role 🙂↕️.
Here is my CV 👇
Technical Skills
Tools: Ansible, Docker, Kubernetes, Terraform(IaC), Maven, CICD(Jenkins), Argo CD, Git & Git-Hub, ELK(Elastic, Logstash,
Kibana), Prometheus & Grafana
Scripting Languages: Bash & Python
AWS Services: AWS Route 53, EKS, IAM, RDS, DynamoDB, ASG, CloudWatch, SNS, S3, AWS Lambda, EC2
Experience
Loomtex Exports July 2022- May 2023
- Managment trainee
Fusion5 August 2023 – Present
- Junior DevOps Engineer Hyderabad, India
Projects
Multi-tier web application
• Deployed a 3-tier application (Front-end, Back-end, Database) on an EKS cluster using Terraform for IaC.
• Configured CI/CD pipelines using Jenkins, integrated with SonarQube for code quality and Nexus for artifact storage.
• Automated environment setup with Ansible playbooks for Jenkins, SonarQube, and Nexus.
• Managed source code with GitHub and utilized Maven and NodeJS for building and packaging application artifacts.
• Enhanced project security using SonarQube and Trivy to detect and mitigate vulnerabilities.
• Built Docker images, stored them on Docker Hub, and set up comprehensive monitoring for system and website
metrics.
Microservice application
• Engineered the implementation of EKS via Terraform and configured Jenkins and SonarQube using Ansible, boosting
deployment efficiency by 35%.
• Established 12 different Jenkins multibranch pipelines to streamline CI/CD processes.
• Implemented Webhooks to increase automation and minimize manual work.
• Created and integrated application components using build tools specified in the pipeline.
• Utilized Docker to create images, transferring them to the Docker registry, and employed Trivy for enhanced security.
• Launched the application on an EKS cluster, using Prometheus and Grafana for performance tracking, achieving a
30% increase in uptime and improving resource utilization by 20%.
AWS Cost Optimization
• Developed and deployed an automated solution using Boto3 and AWS Lambda to remove obsolete EBS snapshots
exceeding 30 days, reducing storage expenses by 20%.
Certifications
AWS Certified Cloud Practitioner Dec 2023 - Dec 2026
Education
Institute of Chemical Technology Bachelor of Technology in Fibers and Textile Processing Technology June 2022
Mumbai, Maharashtra
https://redd.it/1g3aubb
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Seeking Advice: Implementing a Container Image Proxy - What Do You Wish You Knew Before?
Hello there,
We're planning to implement a container image proxy in our environment, and I wanted to reach out to see what advice you all might have.
For those of you who have already set this up, I’m curious:
1. What are the biggest challenges you faced when implementing your container image proxy?
2. Were there any "gotchas" or pitfalls you wish you had known beforehand?
3. What tools or approaches did you find most helpful?
Any insights would be greatly appreciated! We’re currently assessing potential proxies (Harbor, Nexus, etc.) and planning how to integrate this with our existing CI/CD pipelines and Kubernetes clusters.
Thanks in advance for your help!
https://redd.it/1g3curl
@r_devops
Hello there,
We're planning to implement a container image proxy in our environment, and I wanted to reach out to see what advice you all might have.
For those of you who have already set this up, I’m curious:
1. What are the biggest challenges you faced when implementing your container image proxy?
2. Were there any "gotchas" or pitfalls you wish you had known beforehand?
3. What tools or approaches did you find most helpful?
Any insights would be greatly appreciated! We’re currently assessing potential proxies (Harbor, Nexus, etc.) and planning how to integrate this with our existing CI/CD pipelines and Kubernetes clusters.
Thanks in advance for your help!
https://redd.it/1g3curl
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Suggestions for a tool that can perform deployments from a monorepo
I worked at a large org some time ago. Their cloud was deployed by an Azure devops pipeline which ran a powershell script. The script would calculate files from the commit(s) and based on directory and file paths of said files, perform relevant actions, e.g. terraform apply, run a powershell script, apply an Azure policy. The had been developed in-house organicly (=a mess), and my question today is if there are modern open source tools that can perform something similar? E.g. orchestrate shit, based on some kind of rule set, but in a well defined framework.
https://redd.it/1g3e9gx
@r_devops
I worked at a large org some time ago. Their cloud was deployed by an Azure devops pipeline which ran a powershell script. The script would calculate files from the commit(s) and based on directory and file paths of said files, perform relevant actions, e.g. terraform apply, run a powershell script, apply an Azure policy. The had been developed in-house organicly (=a mess), and my question today is if there are modern open source tools that can perform something similar? E.g. orchestrate shit, based on some kind of rule set, but in a well defined framework.
https://redd.it/1g3e9gx
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
What is the most reliable way to deploy a react application in production?
I'm trying to deploy a docker container of a create react app but the environment variables sometimes are not set by the github workflow.
Dockerfile and github workflow
# Use node 21.7.1 as the base image
FROM node:21.7.1
# Set the working directory in the Docker image
WORKDIR /app
# Accept REACT_APP_HOST_API_URL as a build argument
RUN echo "The environment variable REACT_APP_HOST_API_URL is https://20.0.0.120:8080"
RUN echo "The environment variable REACT_APP_ENV is development"
# Set the environment variable so it's available during the build and runtime
ENV REACT_APP_HOST_API_URL=https://20.0.0.120:8080
ENV REACT_APP_ENV=development
ENV NODE_ENV=production
COPY package*.json ./
RUN npm install
RUN npm install -g serve
COPY . .
RUN npm run build
CMD ["serve", "-s", "build", "-l", "3000"]
EXPOSE 3000
name: Build and Push Docker Image
on:
push:
branches:
- main # You can change this to the branch you want to trigger the workflow on
jobs:
build:
runs-on: ubuntu-latest # Use the latest Ubuntu environment for the build
steps:
# Step 1: Checkout the code from the repository
- name: Checkout code
uses: actions/checkout@v3
# Step 2: Set up cache for npm dependencies
- name: Cache npm dependencies
uses: actions/cache@v3
with:
path: ~/.npm # Cache path for npm
key: ${{ runner.os }}-node-${{ hashFiles('package-lock.json') }}
restore-keys: |
${{ runner.os }}-node-
# Step 3: Install dependencies
- name: Install dependencies
run: npm install
# Step 4: Log in to Docker Hub
- name: Log in to Docker Hub
uses: docker/login-action@v2
with:
username: ${{ secrets.DOCKERHUB_USERNAME }} # Your Docker Hub username
password: ${{ secrets.DOCKERHUB_TOKEN }} # Your Docker Hub token or password
# Step 5: Build the Docker image using a custom Dockerfile (Dockerfile-dev.yml)
- name: Build Docker image
run: docker build -f Dockerfile-dev.yml -t my-user-name/react-app:latest .
# Step 6: Push the Docker image to Docker Hub
- name: Push Docker image to Docker Hub
run: docker push my-user-name/react-app:latest
they env are not set when executing the docker image:
`docker exec -it agent-react-dev-react-agent-app-1 /bin/sh`
`# env`
`NODE_VERSION=21.7.1`
`HOSTNAME=db4d9df42f42`
`YARN_VERSION=1.22.19`
`HOME=/root`
`TERM=xterm`
`PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin`
`REACT_APP_HOST_API_URL=`
`REACT_APP_ENV=`
`PWD=/app`
`# exit`
https://redd.it/1g3d1p6
@r_devops
I'm trying to deploy a docker container of a create react app but the environment variables sometimes are not set by the github workflow.
Dockerfile and github workflow
# Use node 21.7.1 as the base image
FROM node:21.7.1
# Set the working directory in the Docker image
WORKDIR /app
# Accept REACT_APP_HOST_API_URL as a build argument
RUN echo "The environment variable REACT_APP_HOST_API_URL is https://20.0.0.120:8080"
RUN echo "The environment variable REACT_APP_ENV is development"
# Set the environment variable so it's available during the build and runtime
ENV REACT_APP_HOST_API_URL=https://20.0.0.120:8080
ENV REACT_APP_ENV=development
ENV NODE_ENV=production
COPY package*.json ./
RUN npm install
RUN npm install -g serve
COPY . .
RUN npm run build
CMD ["serve", "-s", "build", "-l", "3000"]
EXPOSE 3000
name: Build and Push Docker Image
on:
push:
branches:
- main # You can change this to the branch you want to trigger the workflow on
jobs:
build:
runs-on: ubuntu-latest # Use the latest Ubuntu environment for the build
steps:
# Step 1: Checkout the code from the repository
- name: Checkout code
uses: actions/checkout@v3
# Step 2: Set up cache for npm dependencies
- name: Cache npm dependencies
uses: actions/cache@v3
with:
path: ~/.npm # Cache path for npm
key: ${{ runner.os }}-node-${{ hashFiles('package-lock.json') }}
restore-keys: |
${{ runner.os }}-node-
# Step 3: Install dependencies
- name: Install dependencies
run: npm install
# Step 4: Log in to Docker Hub
- name: Log in to Docker Hub
uses: docker/login-action@v2
with:
username: ${{ secrets.DOCKERHUB_USERNAME }} # Your Docker Hub username
password: ${{ secrets.DOCKERHUB_TOKEN }} # Your Docker Hub token or password
# Step 5: Build the Docker image using a custom Dockerfile (Dockerfile-dev.yml)
- name: Build Docker image
run: docker build -f Dockerfile-dev.yml -t my-user-name/react-app:latest .
# Step 6: Push the Docker image to Docker Hub
- name: Push Docker image to Docker Hub
run: docker push my-user-name/react-app:latest
they env are not set when executing the docker image:
`docker exec -it agent-react-dev-react-agent-app-1 /bin/sh`
`# env`
`NODE_VERSION=21.7.1`
`HOSTNAME=db4d9df42f42`
`YARN_VERSION=1.22.19`
`HOME=/root`
`TERM=xterm`
`PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin`
`REACT_APP_HOST_API_URL=`
`REACT_APP_ENV=`
`PWD=/app`
`# exit`
https://redd.it/1g3d1p6
@r_devops
What methodology/best practice would you suggest for devops regarding an Angular Project?
I am new and looking to get some information in CI/CD area for a angular project. I use Gitlab as my repo manager
https://redd.it/1g3kqy9
@r_devops
I am new and looking to get some information in CI/CD area for a angular project. I use Gitlab as my repo manager
https://redd.it/1g3kqy9
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
I believe our devops team under new leadership 90% participates in self manufactured beurocratic process.
We now have a terraform architecture where every piece of infrastructure is a module and every account is a root config (read: one github repo for each module and root config). We are a relatively small actor in our enterprise and this has already led to at least 100 git repos. Every PR to every git repo requires a Jira ticket, a pull request labeled with that jira ticket, and a PR review. All modules are semantic versioned. Realize that any patch to any module requires N PR's to all its upstream modules which import it, which is N jira tickets too, N runs of CICD, and then those modules need to be updated on M other repos, until eventually you get to the root configs, where then you have to do a terraform apply on each of those. One bug, in a downstream module, can take a week to get fixed upstream with this approach. We have set up CICD to do some of the automatic jira ticket creation and PR update creation but we wouldn't need that if we had a more scalable system. I advised the new manager of this problem when they joined and they just said that that's how they did it in their last company and it worked. If we keep going like this it'll move into the thousands quickly I fear. I'm getting carpel tunnel just closing jira tickets and merging PRs for bugs.
I used a monorepo following gitops principles when I was in charge (story is I moved out of management). None of this would be a problem under that paradigm. I'm sure my process could have been improved, but this process is insane.
https://redd.it/1g3mxs2
@r_devops
We now have a terraform architecture where every piece of infrastructure is a module and every account is a root config (read: one github repo for each module and root config). We are a relatively small actor in our enterprise and this has already led to at least 100 git repos. Every PR to every git repo requires a Jira ticket, a pull request labeled with that jira ticket, and a PR review. All modules are semantic versioned. Realize that any patch to any module requires N PR's to all its upstream modules which import it, which is N jira tickets too, N runs of CICD, and then those modules need to be updated on M other repos, until eventually you get to the root configs, where then you have to do a terraform apply on each of those. One bug, in a downstream module, can take a week to get fixed upstream with this approach. We have set up CICD to do some of the automatic jira ticket creation and PR update creation but we wouldn't need that if we had a more scalable system. I advised the new manager of this problem when they joined and they just said that that's how they did it in their last company and it worked. If we keep going like this it'll move into the thousands quickly I fear. I'm getting carpel tunnel just closing jira tickets and merging PRs for bugs.
I used a monorepo following gitops principles when I was in charge (story is I moved out of management). None of this would be a problem under that paradigm. I'm sure my process could have been improved, but this process is insane.
https://redd.it/1g3mxs2
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Candidates Using AI Assistants in Interviews
This is a bit of a doozy — I am interviewing candidates for a senior DevOps role, and all of them have great experience on paper. However, literally 4/6 of them have obviously been using AI resources very blatantly in our interviews (clearly reading from their second monitor, creating very perfect solutions without an ability to adequately explain motivations behind specifics, having very deep understanding of certain concepts while not even being able to indent code properly, etc.)
I’m honestly torn on this issue. On one hand, I use AI tools daily to accelerate my workflow. I understand why someone would use these, and theoretically, their answers to my very basic questions are perfect. My fear is that if they’re using AI tools as a crutch for basic problems, what happens when they’re given advanced ones?
And do we constitute use of AI tools in an interview as cheating? I think the fact that these candidates are clearly trying to act as though they are giving these answers rather than an assistant (or are at least not forthright in telling me they are using an assistant) is enough to suggest they think it’s against the rules.
I am getting exhausted by it, honestly. It’s making my time feel wasted, and I’m not sure if I’m overreacting.
https://redd.it/1g3np7t
@r_devops
This is a bit of a doozy — I am interviewing candidates for a senior DevOps role, and all of them have great experience on paper. However, literally 4/6 of them have obviously been using AI resources very blatantly in our interviews (clearly reading from their second monitor, creating very perfect solutions without an ability to adequately explain motivations behind specifics, having very deep understanding of certain concepts while not even being able to indent code properly, etc.)
I’m honestly torn on this issue. On one hand, I use AI tools daily to accelerate my workflow. I understand why someone would use these, and theoretically, their answers to my very basic questions are perfect. My fear is that if they’re using AI tools as a crutch for basic problems, what happens when they’re given advanced ones?
And do we constitute use of AI tools in an interview as cheating? I think the fact that these candidates are clearly trying to act as though they are giving these answers rather than an assistant (or are at least not forthright in telling me they are using an assistant) is enough to suggest they think it’s against the rules.
I am getting exhausted by it, honestly. It’s making my time feel wasted, and I’m not sure if I’m overreacting.
https://redd.it/1g3np7t
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Will AI take DevOps roles?
One of the main reasons I decided to transition from software engineering to DevOps a couple of years ago is because i think that the SWE field may become increasingly saturated. As tools like ChatGPT continue to improve, more people will be able to rely on AI to complete tasks and potentially secure engineering roles. I see a future where AI significantly reduces the need for traditional software engineers.
DevOps, will be safer from this trend or more difficult to fully automate but do you think at some point this will happen also?
https://redd.it/1g3ttsd
@r_devops
One of the main reasons I decided to transition from software engineering to DevOps a couple of years ago is because i think that the SWE field may become increasingly saturated. As tools like ChatGPT continue to improve, more people will be able to rely on AI to complete tasks and potentially secure engineering roles. I see a future where AI significantly reduces the need for traditional software engineers.
DevOps, will be safer from this trend or more difficult to fully automate but do you think at some point this will happen also?
https://redd.it/1g3ttsd
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Connect Cloud Build and Bitbucket Cloud
Hey guys, devops newbie here. Currently I’m trying to find an alternative to Bitbucket Pipelines due to some limitations related to its self hosted runners (lack of build concurrency, build step timeout only 2 hours).
I am trying to see if Cloud Build is a viable alternative due to its private worker pool option (it’s managed as well).
My company’s repositories are hosted in a workspace on Bitbucket Cloud behind an IP Allowlist.
I’m having trouble trying to connect Cloud Build to a repository since during the link repository process GCP uses an external IP address from a range it has allocated for itself. Google publishes the allocated ranges in a webpage as a json and updates them frequently.
However adding these ranges to our IP Allowlist cannot be a safe choice since this is a public IP range.
Before I move on to another CICD solution, is there something I’m missing to make Cloud Build work?
Please let me know if I need to provide more information.
https://redd.it/1g425zj
@r_devops
Hey guys, devops newbie here. Currently I’m trying to find an alternative to Bitbucket Pipelines due to some limitations related to its self hosted runners (lack of build concurrency, build step timeout only 2 hours).
I am trying to see if Cloud Build is a viable alternative due to its private worker pool option (it’s managed as well).
My company’s repositories are hosted in a workspace on Bitbucket Cloud behind an IP Allowlist.
I’m having trouble trying to connect Cloud Build to a repository since during the link repository process GCP uses an external IP address from a range it has allocated for itself. Google publishes the allocated ranges in a webpage as a json and updates them frequently.
However adding these ranges to our IP Allowlist cannot be a safe choice since this is a public IP range.
Before I move on to another CICD solution, is there something I’m missing to make Cloud Build work?
Please let me know if I need to provide more information.
https://redd.it/1g425zj
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
1 Month Until I Start My First Full-Time DevOps Role – Any Advice?
Hey everyone,
I’ve been working in IT/Cloud/Security for around 4 years, and recently started taking on some DevOps responsibilities at my current job. I’m really happy to share that I’ve just landed a full-time DevOps Engineer role at a great company. My official start date is exactly one month from today.
I’ve offered to visit the office before my start date to better get to know the team and familiarize myself with the projects they have going on. This way, I can gauge where I stand and identify any areas I might need to catch up on.
I’d really appreciate any advice or suggestions on how to best prepare for my first day. This is a big opportunity that I’ve worked incredibly hard to achieve, and I want to make sure I hit the ground running.
Small story time.... just a year ago I was feeling pretty lost. I was out of work, unsure of my next steps, and burned out from my previous role. I even questioned whether I wanted to keep pursuing the engineering path. I decided to take a break, regroup, and commit myself to turning things around. I hit the books, worked on projects, kept my public GitHub active, and sent out around 10 job applications every day. After countless rejections, I finally got the “yes” I had been waiting for.
Thanks for reading, and I’d love to hear any thoughts or advice.
https://redd.it/1g4440l
@r_devops
Hey everyone,
I’ve been working in IT/Cloud/Security for around 4 years, and recently started taking on some DevOps responsibilities at my current job. I’m really happy to share that I’ve just landed a full-time DevOps Engineer role at a great company. My official start date is exactly one month from today.
I’ve offered to visit the office before my start date to better get to know the team and familiarize myself with the projects they have going on. This way, I can gauge where I stand and identify any areas I might need to catch up on.
I’d really appreciate any advice or suggestions on how to best prepare for my first day. This is a big opportunity that I’ve worked incredibly hard to achieve, and I want to make sure I hit the ground running.
Small story time.... just a year ago I was feeling pretty lost. I was out of work, unsure of my next steps, and burned out from my previous role. I even questioned whether I wanted to keep pursuing the engineering path. I decided to take a break, regroup, and commit myself to turning things around. I hit the books, worked on projects, kept my public GitHub active, and sent out around 10 job applications every day. After countless rejections, I finally got the “yes” I had been waiting for.
Thanks for reading, and I’d love to hear any thoughts or advice.
https://redd.it/1g4440l
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
What are your best tips for automation without over-complicating things?
I've been working on automating a few tasks in our DevOps workflow, but sometimes I'm adding unnecessary complexity rather than simplifying things. I'm curious to hear from the community:
* How do you decide what’s worth automating and what’s the better-left manual?
* What tools or techniques help you keep automation simple yet effective?
* Any lessons learned from over-complicating automation in the past?
Thanks in Advance!
https://redd.it/1g44mox
@r_devops
I've been working on automating a few tasks in our DevOps workflow, but sometimes I'm adding unnecessary complexity rather than simplifying things. I'm curious to hear from the community:
* How do you decide what’s worth automating and what’s the better-left manual?
* What tools or techniques help you keep automation simple yet effective?
* Any lessons learned from over-complicating automation in the past?
Thanks in Advance!
https://redd.it/1g44mox
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
OS factory : time and load for such a project
Hello DevOps around the world! 👋
I'm starting a new position at a large firm that primarily uses GCP and Azure for IaaS. Due to the nature of the business, we have a lot of software that isn't compatible with PaaS, so we're stuck with a lot of manual installations.
My team currently installs OS/Agents/Middleware/Software directly on machines, which is a huge time-sink and slows down our time-to-market.
To address this, I want to propose an OS factory project to my management. While I haven't built one before, here's my initial thinking:
* **Core Technologies to be used:**
* **Packer:** For image creation and management.
* **Ansible:** For configuration management and provisioning within the images.
* **GitLab:** To manage our CI/CD pipelines for the entire process.
* **Phased Approach (i want to keep it simple, and build small victories ✌)**
* **Phase 1:** Build a pipeline to create and push patched OS images daily.
* **Phase 2:** Pre-install monitoring agents on the images.
* **Phase 3:** Pre-install security agents on the images.
* **Phase 4:** Work with the security team to apply CIS benchmarks for hardening.
**My questions for the community are:**
* **Existing Open Source Projects:** Are there any open-source projects that could help accelerate this effort?
* **Experience and Effort:** For those who have implemented OS factories:
* How long did it take, and what were the resource requirements (people, time)?
* What were your biggest challenges, and what advice would you give?
I believe this is a relevant topic for many DevOps professionals. Thanks in advance for your insights!
https://redd.it/1g43r72
@r_devops
Hello DevOps around the world! 👋
I'm starting a new position at a large firm that primarily uses GCP and Azure for IaaS. Due to the nature of the business, we have a lot of software that isn't compatible with PaaS, so we're stuck with a lot of manual installations.
My team currently installs OS/Agents/Middleware/Software directly on machines, which is a huge time-sink and slows down our time-to-market.
To address this, I want to propose an OS factory project to my management. While I haven't built one before, here's my initial thinking:
* **Core Technologies to be used:**
* **Packer:** For image creation and management.
* **Ansible:** For configuration management and provisioning within the images.
* **GitLab:** To manage our CI/CD pipelines for the entire process.
* **Phased Approach (i want to keep it simple, and build small victories ✌)**
* **Phase 1:** Build a pipeline to create and push patched OS images daily.
* **Phase 2:** Pre-install monitoring agents on the images.
* **Phase 3:** Pre-install security agents on the images.
* **Phase 4:** Work with the security team to apply CIS benchmarks for hardening.
**My questions for the community are:**
* **Existing Open Source Projects:** Are there any open-source projects that could help accelerate this effort?
* **Experience and Effort:** For those who have implemented OS factories:
* How long did it take, and what were the resource requirements (people, time)?
* What were your biggest challenges, and what advice would you give?
I believe this is a relevant topic for many DevOps professionals. Thanks in advance for your insights!
https://redd.it/1g43r72
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
DevOps online courses help
Hey everyone,
I'm coming from a JavaScript Full Stack background and looking to transition into DevOps. I found the "DevOps Beginners to Advanced with Projects" course on Udemy ( https://www.udemy.com/course/decodingdevops/?couponCode=2021PM20 ) and was wondering if it's a good starting point to be able to pursue a junior position in the field. Has anyone taken this course? Would you recommend it or suggest something else?
I was also recommended a more specific aws and cka courses but I'm aiming to accomplish them after going through a complete DevOps course.
Thanks in advance!
https://redd.it/1g440rd
@r_devops
Hey everyone,
I'm coming from a JavaScript Full Stack background and looking to transition into DevOps. I found the "DevOps Beginners to Advanced with Projects" course on Udemy ( https://www.udemy.com/course/decodingdevops/?couponCode=2021PM20 ) and was wondering if it's a good starting point to be able to pursue a junior position in the field. Has anyone taken this course? Would you recommend it or suggest something else?
I was also recommended a more specific aws and cka courses but I'm aiming to accomplish them after going through a complete DevOps course.
Thanks in advance!
https://redd.it/1g440rd
@r_devops
Udemy
DevOps Beginners to Advanced with Projects
Begin Your DevOps Career As a Newbie | AWS, Linux, Scripting, Jenkins, Ansible, GitOps, Docker, Kubernetes, & Terraform.
Ideas for creating Dead Man's Switch emailing system
Hey guys.
I am not sure if this is the right sub for this, but I feel like you all are my best bet.
Well I am looking to setup a system that basically functions as a Dead Man's Switch that will send out an email to my family members in case I pass away or something. I have seen services like deadmansswitch.net, but there are a few reasons why I am not using their service.
Basically, it would have to work such that the system sends you reminders by email every now and then, and you have to click on a link. If you don't click the link within a predetermined period, the system will trigger and send out a predefined email to your recipients.
I am not a hardcore DevOp like most of you guys, but I know some basic programming. What would be the easiest way to go about building a homemade solution like this?
https://redd.it/1g487tb
@r_devops
Hey guys.
I am not sure if this is the right sub for this, but I feel like you all are my best bet.
Well I am looking to setup a system that basically functions as a Dead Man's Switch that will send out an email to my family members in case I pass away or something. I have seen services like deadmansswitch.net, but there are a few reasons why I am not using their service.
Basically, it would have to work such that the system sends you reminders by email every now and then, and you have to click on a link. If you don't click the link within a predetermined period, the system will trigger and send out a predefined email to your recipients.
I am not a hardcore DevOp like most of you guys, but I know some basic programming. What would be the easiest way to go about building a homemade solution like this?
https://redd.it/1g487tb
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
How common is it for companies that host hackathons to forbid contractors from participating?
Understand there are a variety of opinions on hackathons.
I work at a place that forbids full time contractors from participating in them. I'm trying to understand if the policy has a legal basis, if it's financially driven or has other motivations that aren't apparent to me.
https://redd.it/1g49sjx
@r_devops
Understand there are a variety of opinions on hackathons.
I work at a place that forbids full time contractors from participating in them. I'm trying to understand if the policy has a legal basis, if it's financially driven or has other motivations that aren't apparent to me.
https://redd.it/1g49sjx
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Advice on new architecture. No more AWS ECS, Ansible to orchestrate docker on EC2 instead. Am I insane?
The problem
I currently use AWS ECS Fargate and ALB to serve my api and run my background workers. I don't like this setup because 1) for what we're getting out of it it's too expensive, 2) I can't truly replicate it locally¹, and 3) most importantly, squeezing in DB migrations into an ECS deployment has been quite painful. DB migrations specifically highlighted to me that ECS seems to be the wrong tool here.
So l'd like for some thoughts on if I'm thinking right about completely switching away from ECS and the tools I'm picking to do so.
Some context on the application first though. It's a B2B app, so sudden demand increase is unlikely. I will scale up slowly when needed, but I do want to have a plan for when that need comes. Zero downtime deployments are also not a requirement, a few minutes of downtime at night are fine.
My plan
A single EC2 instance as beefy as needed. It's got the app running in containers and the way to scale is to increase the number of containers. I'll use traefik as a reverse proxy in front of the api containers. I don't know what the appropriate load balancing algorithm would be here, but I didn't think this is that big of a deal either, right? As in, would the answer to this question affect the architecture I'm deciding on? Or can I just revisit this after I've implemented everything to better optimize the load balancer?
When code is merged to master, GitHub Actions will build a new app image, and then I'll use Ansible to automate the deployment in that EC2 instance.
I picked Ansible instead of writing a custom bash script because it seemed that I can use Ansible to declare what I want to happen in some sense, but I can still also imperatively write how it should be done. Is that correct?
This is the most vague area to me to be honest, so feedback here is greatly appreciated. I have never used Ansible before.
Another relevant point to mention here as well is that my custom bash script would be annoyingly stateful, which seemed too error prone. For example I have to check and ensure the state of the machine first, like spinning up traefik if it's not already running, and checking the db connection, etc. Ansible seemed like it had a good approach to this issue. It's a complex issue though, so I'm keeping my expectations low. Experience with things like this is appreciated.
I'm planning on using parameterized docker-compose files to configure the containers and set their network, static env vars, entrypoints, etc., in addition to passing in dynamic configuration (i.e. from AWS SSM) as env vars as well.
Running the entire app locally would simply be `docker-compose up -d` or I can even simulate a deployment process exactly by running the Ansible playbook targeting my local machine.
Here's the deployment logic if it's useful to know:
(I will use a GitHub Actions workflow to run the Ansible playbook directly into the instance to update it)
- Pull the new container image.
- Add a higher priority route in Traefik to point api traffic to a maintenance page.
- Wait 30 seconds for any ongoing requests to complete since Traefik doesn't support connection draining.
- Stop (not remove) old app containers.
- Run DB migration using new app image. Migration atomicity will be ensured by a few steps that are irrelevant here, but what's relevant to know is that it will be done using direct access to the PostgreSQL CLI.
- If db migration is successful, spin up containers for new app version, remove maintenance page and clean up old containers. Deployment is done!
- If db migration fails, start old app containers, remove maintenance page and send alerts.
As for observability, Prometheus would scrape the local collectors (cAdvisor, OTel, etc.) for logs and metrics.
When I need to scale up the api or the workers I'll add more containers, and when I reach hardware limits I can upgrade the instance type. Any gotchas I should know about
The problem
I currently use AWS ECS Fargate and ALB to serve my api and run my background workers. I don't like this setup because 1) for what we're getting out of it it's too expensive, 2) I can't truly replicate it locally¹, and 3) most importantly, squeezing in DB migrations into an ECS deployment has been quite painful. DB migrations specifically highlighted to me that ECS seems to be the wrong tool here.
So l'd like for some thoughts on if I'm thinking right about completely switching away from ECS and the tools I'm picking to do so.
Some context on the application first though. It's a B2B app, so sudden demand increase is unlikely. I will scale up slowly when needed, but I do want to have a plan for when that need comes. Zero downtime deployments are also not a requirement, a few minutes of downtime at night are fine.
My plan
A single EC2 instance as beefy as needed. It's got the app running in containers and the way to scale is to increase the number of containers. I'll use traefik as a reverse proxy in front of the api containers. I don't know what the appropriate load balancing algorithm would be here, but I didn't think this is that big of a deal either, right? As in, would the answer to this question affect the architecture I'm deciding on? Or can I just revisit this after I've implemented everything to better optimize the load balancer?
When code is merged to master, GitHub Actions will build a new app image, and then I'll use Ansible to automate the deployment in that EC2 instance.
I picked Ansible instead of writing a custom bash script because it seemed that I can use Ansible to declare what I want to happen in some sense, but I can still also imperatively write how it should be done. Is that correct?
This is the most vague area to me to be honest, so feedback here is greatly appreciated. I have never used Ansible before.
Another relevant point to mention here as well is that my custom bash script would be annoyingly stateful, which seemed too error prone. For example I have to check and ensure the state of the machine first, like spinning up traefik if it's not already running, and checking the db connection, etc. Ansible seemed like it had a good approach to this issue. It's a complex issue though, so I'm keeping my expectations low. Experience with things like this is appreciated.
I'm planning on using parameterized docker-compose files to configure the containers and set their network, static env vars, entrypoints, etc., in addition to passing in dynamic configuration (i.e. from AWS SSM) as env vars as well.
Running the entire app locally would simply be `docker-compose up -d` or I can even simulate a deployment process exactly by running the Ansible playbook targeting my local machine.
Here's the deployment logic if it's useful to know:
(I will use a GitHub Actions workflow to run the Ansible playbook directly into the instance to update it)
- Pull the new container image.
- Add a higher priority route in Traefik to point api traffic to a maintenance page.
- Wait 30 seconds for any ongoing requests to complete since Traefik doesn't support connection draining.
- Stop (not remove) old app containers.
- Run DB migration using new app image. Migration atomicity will be ensured by a few steps that are irrelevant here, but what's relevant to know is that it will be done using direct access to the PostgreSQL CLI.
- If db migration is successful, spin up containers for new app version, remove maintenance page and clean up old containers. Deployment is done!
- If db migration fails, start old app containers, remove maintenance page and send alerts.
As for observability, Prometheus would scrape the local collectors (cAdvisor, OTel, etc.) for logs and metrics.
When I need to scale up the api or the workers I'll add more containers, and when I reach hardware limits I can upgrade the instance type. Any gotchas I should know about
when upgrading instance types? From what I read in the docs, upgrading within the same family seems to be a simple task, no?
I know this is all over the place, there's a lot of things that need to fit in together properly, but I tried to only mention what's relevant. If there are any points I forgot to mention, I'll be happy to answer.
If what I'm asking for isn't clear I can also try and clarify that further.
Please mention any pitfalls I might fall into, even if you don't think they apply to my situation.
---
¹: I know that as some point replicating a scalable system locally is not a realistic expectation, but in our current state I see no reason why I can't spin up the same containers and reverse proxy locally and get the same deployed setup on my local machine. When the need comes for something like a hosted global service, I'll drop that requirement, but for now I don't see why I shouldn't be able to do so.
https://redd.it/1g4alqw
@r_devops
I know this is all over the place, there's a lot of things that need to fit in together properly, but I tried to only mention what's relevant. If there are any points I forgot to mention, I'll be happy to answer.
If what I'm asking for isn't clear I can also try and clarify that further.
Please mention any pitfalls I might fall into, even if you don't think they apply to my situation.
---
¹: I know that as some point replicating a scalable system locally is not a realistic expectation, but in our current state I see no reason why I can't spin up the same containers and reverse proxy locally and get the same deployed setup on my local machine. When the need comes for something like a hosted global service, I'll drop that requirement, but for now I don't see why I shouldn't be able to do so.
https://redd.it/1g4alqw
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
I launched my DevOps PostgreSQL platform today - feedback?
My name is Elliott, I’ve been building a DevOps platform the last three years on the top best in class open source platforms (Kubernetes, Elixir, PostgreSQL, Grafana, etc). The goal is to give
engineering teams access to a modern DevOps infrastructure without needing to have full SRE/DevOps committed resourcing.
It’s also open source/fair source - all the source code is here → [https://github.com/batteries-included/batteries-included](https://github.com/batteries-included/batteries-included)
I just shipped a public beta today and would love to hear initial reactions, thoughts, feedback.
Here’s some of the specific details of the platform:
* The platform features a user-friendly suggestion-based interface that guides users on topics like PostgreSQL cluster memory/CPU ratios, serverless web hosting, and secure secret sharing. Advanced users can quickly access full control over their data.
* It’s an Elixir-based UI on a database-driven, self-hosted Kubernetes platform. It can automatically deploy a scalable cloud installation (currently on AWS, with more options to follow) without the need for YAML or Terraform configurations. Alternatively, it can set up a development instance using Kind and Docker or Podman, facilitating a smooth transition from local to production environments.
* The platform supports easy AI project hosting for various workloads. Use Ollama embedding models for text embedding, eliminating OpenAI costs and data leakage risks. With PGVector and Cloud Native PG for vector databases, you can achieve near-state-of-the-art performance without exposing your data to third-party APIs. Experiment with Jupyter Notebooks, featuring optional Nvidia Plugin batteries for no DevOps-required experimentation.
* Single Sign-On is streamlined via Keycloak, Istio Ingress, and OAuth Proxy, securely hosted on your machine or cloud account. We've simplified security with full mTLS, Istio, SSL generation, and automated routing with Let's Encrypt and Acme for HTTP2. Istio Ingress services are seamlessly configured down to the contents of config maps.
* Grafana and Victoria Metrics can be auto-configured with just a few clicks for easy installation.
Here’s also a look at the demo of the database deploy [https://www.youtube.com/watch?v=YbvkWja3VIQ](https://www.youtube.com/watch?v=YbvkWja3VIQ)
The platform follows all the best practices learned for configuring and running a maintainable system without Kubenete's GitOps pain.
If you want to check it out, here are links to docs, site, repo, and join:
* [https://www.batteriesincl.com/](https://www.batteriesincl.com/)
* [https://home.batteriesincl.com/signup](https://home.batteriesincl.com/signup)
* [https://github.com/batteries-included](https://github.com/batteries-included)
* [https://www.batteriesincl.com/docs](https://www.batteriesincl.com/docs)
https://redd.it/1g49t4q
@r_devops
My name is Elliott, I’ve been building a DevOps platform the last three years on the top best in class open source platforms (Kubernetes, Elixir, PostgreSQL, Grafana, etc). The goal is to give
engineering teams access to a modern DevOps infrastructure without needing to have full SRE/DevOps committed resourcing.
It’s also open source/fair source - all the source code is here → [https://github.com/batteries-included/batteries-included](https://github.com/batteries-included/batteries-included)
I just shipped a public beta today and would love to hear initial reactions, thoughts, feedback.
Here’s some of the specific details of the platform:
* The platform features a user-friendly suggestion-based interface that guides users on topics like PostgreSQL cluster memory/CPU ratios, serverless web hosting, and secure secret sharing. Advanced users can quickly access full control over their data.
* It’s an Elixir-based UI on a database-driven, self-hosted Kubernetes platform. It can automatically deploy a scalable cloud installation (currently on AWS, with more options to follow) without the need for YAML or Terraform configurations. Alternatively, it can set up a development instance using Kind and Docker or Podman, facilitating a smooth transition from local to production environments.
* The platform supports easy AI project hosting for various workloads. Use Ollama embedding models for text embedding, eliminating OpenAI costs and data leakage risks. With PGVector and Cloud Native PG for vector databases, you can achieve near-state-of-the-art performance without exposing your data to third-party APIs. Experiment with Jupyter Notebooks, featuring optional Nvidia Plugin batteries for no DevOps-required experimentation.
* Single Sign-On is streamlined via Keycloak, Istio Ingress, and OAuth Proxy, securely hosted on your machine or cloud account. We've simplified security with full mTLS, Istio, SSL generation, and automated routing with Let's Encrypt and Acme for HTTP2. Istio Ingress services are seamlessly configured down to the contents of config maps.
* Grafana and Victoria Metrics can be auto-configured with just a few clicks for easy installation.
Here’s also a look at the demo of the database deploy [https://www.youtube.com/watch?v=YbvkWja3VIQ](https://www.youtube.com/watch?v=YbvkWja3VIQ)
The platform follows all the best practices learned for configuring and running a maintainable system without Kubenete's GitOps pain.
If you want to check it out, here are links to docs, site, repo, and join:
* [https://www.batteriesincl.com/](https://www.batteriesincl.com/)
* [https://home.batteriesincl.com/signup](https://home.batteriesincl.com/signup)
* [https://github.com/batteries-included](https://github.com/batteries-included)
* [https://www.batteriesincl.com/docs](https://www.batteriesincl.com/docs)
https://redd.it/1g49t4q
@r_devops
GitHub
GitHub - batteries-included/batteries-included: Batteries Included is a Kubernetes based software platform for database, ai, web…
Batteries Included is a Kubernetes based software platform for database, ai, web, monitoring, and more. - batteries-included/batteries-included
Starting Devops (no cs background
Hey everyone I’m starting to learn devops buying a course on udemy by Imran teli I’m seeking for advice and suggestions about things while learning devops also will my no it/cs backround affect my hiring process once i’m ready to work??
https://redd.it/1g4k5xp
@r_devops
Hey everyone I’m starting to learn devops buying a course on udemy by Imran teli I’m seeking for advice and suggestions about things while learning devops also will my no it/cs backround affect my hiring process once i’m ready to work??
https://redd.it/1g4k5xp
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Is there an easy way to see which containers triggered an error without explicitly sending the errors to a logging service?
I have 25 containers constantly sending messages to one another and sometimes one of them gets an error, but I have no idea which container got it. Is there a way to listen for errors on every docker container and centralize logging without explicitly writing code to send error to a microservice? I am using a local docker environment.
https://redd.it/1g4ljef
@r_devops
I have 25 containers constantly sending messages to one another and sometimes one of them gets an error, but I have no idea which container got it. Is there a way to listen for errors on every docker container and centralize logging without explicitly writing code to send error to a microservice? I am using a local docker environment.
https://redd.it/1g4ljef
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
How much of a challenge are telemetry (metrics, logs, traces) storage costs for your team / company?
Real-life cases of overpriced/inefficient used telemetry storage are very welcome in comments!
View Poll
https://redd.it/1g4iysh
@r_devops
Real-life cases of overpriced/inefficient used telemetry storage are very welcome in comments!
View Poll
https://redd.it/1g4iysh
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community