Innovating the docs experience Like most in the cloud-native community, we’re passionate about great documentation. We’ve felt the joy when discovering a clear, thoughtful guide that helps us get up and running with a new project in our homelabs… and we’ve certainly felt the pain that bad technical writing can cause. So you can understand why our docs team here at Spectro Cloud are always working to improve the experience we offer our users (that’s you). Of course, that includes expanding and refining our content itself — but also innovating how we make that content accessible. SpectroMate is an open-source project I created while working at Spectro Cloud. SpectroMate is an API server with extended functionality designed for Slack integration in the form of a bot. You can use SpectroMate to handle slash commands, and message actions. This article was originally published in Spectro Cloud’s blog. Click on the Article Link to read the original source.
You have assumed the leadership of a team that is operating in a cloud environment. It’s a new beginning, you are excited about the future (hopefully), the team members, and most of all, the thrill of a new challenge. After the excitement settles down you start asking questions to better understand the work and the team. Among the list of questions you have, you should include questions pertaining to cloud cost and cost optimization. This article was originally published on Medium. Link to the Medium article can be found here. In this article, you will find a set of questions that are beneficial for you and your team to further explore. These are questions I have found beneficial in the past and I believe they will be beneficial to you too. Without further ado, let’s dive into it. Q: Do we have any budget alarms established? This is a simple question but the answer will reveal a lot of information about the team, the organization, and the emphasis placed on cost management.
Two approaches to injecting variability into your Nomad batch job template without having to modify the template in the future. This article was originally published in HashiCorp’s blog while I was an employee. Click on the Article Link to learn how to inject variability into Nomad jobs.
If you find yourself authoring several AWS lambdas for a serverless application architecture, you might have encountered this error: 1 2 3 4 An error occurred: mySweetLambda– Code storage limit exceeded. (Service: AWSLambda; Status Code: 400; Error Code: CodeStorageExceededException; Request ID: 05d3ae68-a7c2-a3e8-948e-41c2739638af). The first time I encountered this error I wasn’t quite sure what was happening, but after some quick web searches I learned that AWS has a limit on Lambda storage that maxes out at 75Gb. Additionally, I also learned that AWS retains all the previous versions of all my lambdas. That’s all fine, I should probably go do some “spring cleaning” and remove the unused versions. AWS does expose the functionality to remove former versions through the console. However, in my scenario I had over 500+ versions for some of my older lambdas. Clicking through 500+ versions is not how I want to spend my time. So what options are available?
Chances are most of us have unique situations for wanting to interact with DynamoDB locally, maybe it’s to develop and test different data models, perhaps it’s to develop programmatic functions to interact with the database, perhaps you want to reduce development expenses, or perhaps you’re just doing research. Regardless of your reasons, I want to help you by showing you how to leverage DynamoDB locally. We will use the following tools. Medium Link Localstack Terraform Go AWS CLI noSQL Workbench for DynamoDB We will walk through setting up the local environment, generating data, uploading data, interacting with the noSQL Workbench, and some neat tips to keep in mind. So with that being said, let’s dive into into it! Note: If you get lost, simply visit https://github.com/karl-cardenas-coding/dynamodb-local-example to view the end solution. Also, feel free to fork this template project and use it as a starting point. Setting up the environment First thing first, ensure that you have Terraform (> v0.
It’s tough to find an organization that is not leveraging a public cloud platform. With all the SaaS, PaaS, and IaaS providers out there, chances are you’ve used cloud services in some manner. As public cloud utilization continues to ramp up in many industries (e.g., insurance, financial, business), organizations are encountering a new challenge: how to correctly consume these new platforms and technologies. It’s a workforce challenge more so than it is a technical challenge. This article was originally published in the State Farm Engineering Blog. Link to the original blog can be found here. It’s a workforce challenge because, at the end of the day, it comes down to technical education and experience. Organizations with a workforce skilled in cloud technologies will have a smoother cloud consumption experience compared to one lacking a workforce with cloud experience. Every organization is competing for talent - we all want that DevSecOps engineer and/or Cloud Architect that can do it all (CI/CD, Programming, Security, Architecting, Data, Test, etc.