These are my links for 9 Dic 2015 from 12:09 to 16:26:
- One Thing Well | Let’s Encrypt – Let’s Encrypt is now in public beta and offers a command line tool that makes the process of getting and renewing certificates easy, but you have to run it as root, and it’s designed to rewrite your web server’s configuration files. Here’s a selection of alternative tools and clients:
- Tsung – It can be used to stress HTTP, WebDAV, SOAP, PostgreSQL, MySQL, LDAP and Jabber/XMPP servers. Tsung is a free software released under the GPLv2 license. The purpose of Tsung is to simulate users in order to test the scalability and performance of IP based client/server applications. You can use it to do load and stress testing of your servers. Many protocols have been implemented and tested, and it can be easily extended. It can be distributed on several client machines and is able to simulate hundreds of thousands of virtual users concurrently (or even millions if you have enough hardware …). Tsung is developed in Erlang, an open-source language made by Ericsson for building robust fault-tolerant distributed applications. [ via http://onethingwell.org/post/134852940551/tsung ]
- Internet Redundancy with ASA SLA and IPSec – PacketU – I’ve seen a lot of examples of redundant Internet connections that use SLA to track a primary connection. The logic is that the primary Internet connection is constantly being validated by pinging something on that ISP’s network and routing floats over to a secondary service provider in the event of a failure. I was recently challenged with how this interacted with IPSec. As a result I built out this configuration and performed some fairly extensive testing.
These are my links for 16 nov 2015 through 17 nov 2015:
- Dkron – Distributed, fault tolerant job scheduling system – Dkron is a system service that runs scheduled jobs at given intervals or times, just like the cron unix service but distributed in several machines in a cluster. If a machine fails (the leader), a follower will take over and keep running the scheduled jobs without human intervention. Dkron is Open Source and freely available.
- apache tuneup – Important to understand, this apache servers only one static image we use for the BI team. It does recieve and log a huge cookie and sends a 43 byte 1×1.gif image. Atention: It may not work right for dynamic web servers, as I use the mod_cache.
- Gogs – Go Git Service – a painless self-hosted Git service – A painless self-hosted Git service. The goal of this project is to make the easiest, fastest, and most painless way of setting up a self-hosted Git service. With Go, this can be done with an independent binary distribution across ALL platforms that Go supports, including Linux, Mac OS X, Windows and ARM.
These are my links for 21 set 2015 through 24 set 2015:
- IOWait. (Sysadmin’s bedtime horror story) – […] hope this helps someone out there. The last 3 days, I have had my server crash on me every 2-3 hours. At first I thought it would be a spike in the traffic, since I couldn’t find any crash reports from Apache, and there was a spike in the traffic at this time. So I increased the resources on the server. It crashed again, every 2-3 hours […] It wasn't my case but… who knows in the future?
- Apache mod_deflate and mod_cache issues | Devon Hillard’s Digital Sanctuary – The Problem: Using Apache mod_deflate and mod_disk_cache (or other mod_cache) together can create far too many cached files. The Background: Apache is a web server with many different modules you can load in to enhance it. Two common ones are mod_deflate and mod_cache (or mod_disk_cache).
- haskellcamargo/skype-unofficial-client · GitHub – (unofficial) Skype client built on top of node webkit [ via http://www.lffl.org/2015/09/skype-web-client-linux.html ]
- dfletcher/tsws · GitHub – TSWS, A Totally Simple Web Server
- Home | Lattice – Lattice aspires to make clustering containers easy. Lattice includes a cluster scheduler, http load balancing, log aggregation and health management. Lattice containers can be long running or temporary tasks which get dynamically scaled and balanced across a cluster. Lattice packages components from Cloud Foundry to provide a cloud native platform for individual developers and small teams.
These are my links for 29 mag 2015 through 10 giu 2015:
- My Blog: AWS EC2 Auto Scaling: Basic Configuration – Our goal: Create an Auto Scaling EC2 Group in a single Availability Zone and use a HTTP status page as a Health Monitor for our Load Balancer and the Auto Scaling group instances. This exercise will show us some Auto Scaling basics and will be useful to understand the concepts beneath but the Auto Scaling Group will not automatically "scale" responding to external influence like Average CPU Usage or Total Apache Connections (This aspect is covered in this post: AWS EC2 Auto Scaling: External CloudWatch Metric). With the Auto Scaling configuration described here, we will obtain a web server cluster that can be increased and decreased in members with a simple Auto Scaling API call and we will transfer the monitoring role to the ELB to automatically replace failed EC2 instances or web servers.
- Autoscaling with custom metrics « That’s Geeky – One of the appeals of cloud computing is the idea of using what you need when you need. One of the ways that Amazon provides for this is through autoscaling. In essence, this allows you to vary the number of (related) running instances according to some metric that is being tracked. In this article, we look at how you can trigger a change in the number of running instances using a custom Cloudwatch metric – including the setup of said metric, and a brief look at the interactions between the various autoscaling commands used.
- Painless AWS Auto Scaling With EBS Snapshots And Capistrano – Boom – AWS (Amazon Web Services) auto scaling is a simple concept on the surface: You get an AMI, set up rules, and the load balancer takes care of the rest. However, actually getting it done is more complicated. Some choices are worse than others: you could bake an AMI (Amazon Machine Image) before you deploy, but that could add 10 minutes or more to each deployment. Some are dangerous: you could create an AMI after each deploy, but you run the risk that an auto scale even happens before your AMIs are done. Plus, you have a whole variety of AMIs deployed in at any given time. Some are similar to what we propose in this tutorial: you could push your code to S3 on each deploy and have user-data scripts that pull it down on each auto scaling event. However you slice it, to get auto scaling to fit into your development work flow in a transparent way takes careful thought and planning. We recently rolled out the following solution at CodePen. It keeps our AMIs static and our application ready for scaling on EBS (Elastic Block Store) snapshots. We can push code using Capistrano and let a few scripts distribute the ever-changing code base to our fleet of servers. I’d like to share the steps required to make it work. This series of posts will walk you through the steps required to build an auto-scaling infrastructure that stays out of your way.
- coderwall.com : establishing geek cred since 1305712800 – Did you accidentally set node.normal[:foo][:bar] = 'something bad' in your chef recipe? Then you found that the node's normal attributes persisted between chef runs, and you really wanted to use the default attribute precedence level in your cookbook's attributes/default.rb file?