AWS offers a range of EC2 instance types optimized for various purposes. It’s great that they provide so much variety, but of course, it means one more thing that you have to learn. It’s worth taking the time to do so, as ⅔ of IaaS spend goes toward compute – that’s a lot of EC2.
Check out a brief breakdown in this video, which also compares EC2 purchasing options. Check it out here:
Or, read on for a look into each of the AWS instance types. Remember that within each type, you’ll still need to choose the AWS instance sizes that suit your specific needs. Additionally, older generations within each instance types are available for purchase – for example, c5 is the latest “c” instance, but c4 and c3 are still available – but as the newer types tend to perform better at a cheaper price, you’ll only want to use the older types if you have an AMI or other dependency. The differences matter for some users… but you probably already know who you are.
Note: a version of this blog was originally published in July 2018. It has been rewritten and updated for 2020. New EC2 instance types since our last writeup include A1, T3, z1d, high memory, R5, G4, and F1.
Quick EC2 Instance Info
This chart shows a quick summary of what we’ll cover. We’re including a brief description and mnemonic for each (hopefully helpful if you’re studying for an AWS certification!)
If you’ve taken a look at AWS training materials, you may have seen a couple of overall acronyms to remember all of these – perhaps Dr McGiFT Px or FIGHT Dr McPX. Whether these acronyms are useful at all is perhaps a point of discussion, but to ensure that all the instance types above are in your list, we suggest:
Fight Czar MXPD
Fright Camp DXZ
March Gift PZXD
(and don’t forget high memory and Inf!)
These general purpose AWS EC2 instance types are a good place to start, particularly if you’re not sure what type to use. There are three general purpose types.
t instance type
The t3 family is a burstable instance type. If you have an application that needs to run with some basic CPU and memory usage, you can choose t3. It also works well if you have an application that gets used sometimes but not others. When the resource is idle, you’ll generate CPU credit, which you’ll utilize when the resource is used. It’s useful for things that come and go a lot, such as websites or development environments, and while generally inexpensive, make sure you understand how the CPU credits work before deploying these. There’s a little bit of math and they may not be as cheap as they look at first glance.
Make sure you also understand the difference between t3 and the older t2 – t3 are in “unlimited mode” by default, so instead of throttling down to baseline CPU when your instance runs out of credits, you pay for overages.
For each of the EC2 types we cover here, we’ll also add a mnemonic to help you remember the purpose of each instance type.
Mnemonic: t is for tiny or turbo.
m instance type
The m5 instance type is similar, but for more consistent workloads. It has a nice balance of CPU, memory, and disk. It’s not hard to see why almost half of EC2 workloads are on “m” instances. In addition to m5, you also have the option of m6g, which are powered by Arm-based AWS Graviton2 processors, making them more cost-efficient. There’s also m5a, m5n, and m4 – most of which are safe to ignore unless you have a specific use case for one of the other processors besides m5’s Intel Xeon Platinum 8175 processors. If you aren’t sure what to choose, m5 is the most versatile of all the Amazon instance types.
Mnemonic: m is for main choice or happy medium.
a1 instance type
The a1 instance type was announced in late 2018 and can be a less expensive option than other EC2. They are suited for scale-out workloads such as web servers, containerized microservices, caching fleets, distributed data stores, and development environments. The instances are powered by Arm processors and suited for Arm-based workloads.
Mnemonic: a is for Arm processor
c instance type
The c5 instance type has a high ratio of compute/CPU versus memory. If you have a compute-intensive application – maybe scientific modelling, intensive machine learning, or multiplayer gaming – these instances are a good choice. There is also the c5d option, which is SSD-backed. See also the C5n which have up to 100 Gbps network bandwidth and increased memory compared to equivalent c5 instances. The c4 family is also still available.
Mnemonic: c is for compute (at least that one’s easy!)
r instance family
The r instance family is memory-optimized, which you might use for in-memory databases, real-time processing of unstructured big data, or Hadoop/Spark clusters. You can think of it as a kind of midpoint between the m5 and the x1e. In addition to r5, there are r5a which deliver lower cost per GiB memory and r5n which have higher bandwidth for applications that need improved network throughput and packet rate performance.
Mnemonic: r is for RAM.
x1 instance family
The x1 family has a much higher ratio of memory, so this is a good choice if you have a full in-memory application or a big data processing engine like Apache Spark or Presto. X1e are optimized for high-performance databases, in-memory databases, and other memory intensive enterprise applications.
Mnemonic: x is for xtreme, as in “xtreme RAM” seems to be generally accepted, but we think this is a bit weak. If you have any suggestions, comment below.
High Memory instance family
We’re not sure why these didn’t get an alphabet soup name like the rest of the AWS instances, but at least it’s easy to remember and understand. As you might guess, high memory instances run large in-memory databases, including production deployments of SAP HANA.
Mnemonic: we’ll leave this one up to you.
z1d instance family
The z1d instances combine high compute capacity with a high memory footprint. They have a sustained core frequency of up to 4.0 GHz, the fastest of AWS’s offerings. These are best for electronic design automation (EDA) and some relational database workloads with high per-core licensing costs.
Mnemonic: z is for zippy
p instance type
If you need GPUs on your instances, p3 instances are a good choice. They are useful for video editing, and AWS also lists use cases of “computational fluid dynamics, computational finance, seismic analysis, speech recognition, autonomous vehicles” – so it’s fairly specialized. p2 instances are also available.
Mnemonic: p is for pictures (graphics).
Inf1 instance type
The Inf1 instances are a specialized EC2 type for machine learning inference applications, such as recommendation engines, forecasting, image and video analysis, advanced text analytics, document analysis, voice, conversational agents, translation, transcription, and fraud detection.
Mnemonic: inf is for inference
g instance type
The g instance type uses Graphics Processing Units (GPUs) to accelerate graphics-intensive workloads, and also designed to accelerate machine learning inference. This could include adding metadata to an image, automated speech recognition, and language translation, as well as graphics workstations, video transcoding, and game streaming in the cloud.
g4 is the latest family, and g3 are available as well.
Mnemonic: g is for graphics or GPU
F1 instance type
f1 instances offer customizable hardware acceleration with field programmable gate arrays (FPGAs) – hence the “f”. Applications could include genomics research, financial analysis, and real-time video processing.
Mnemonic: f is for FPGA
i3 instance type
The i3 instance type is similar to h1, but it is SSD backed, so if you need an NVMe drive, choose this type. Use it for NoSQL databases, in-memory databases, Elasticsearch, and more. The i3en option has higher network bandwidth with Elastic Network Adapter (ENA)-based enhanced networking.
Mnemonic: i is for IOPS.
d2 instance type
d2 instances have an even higher ratio of disk to CPU and memory, which makes them a good fit for Massively Parallel Processing (MPP), MapReduce and Hadoop distributed computing, and similar applications.
Mnemonic: d is for dense.
h1 instance type
The h1 type is HDD backed, with a balance of compute and memory. You might use it for distributed file systems, network file systems, or data processing applications.
Mnemonic: h is for HDD.
What EC2 instance types should you use?
As AWS has continued to add options to EC2, there are now EC2 instance types for almost any application. If you have comparison questions around pricing, run them through the AWS monthly calculator. And if you don’t know, then generally starting with t3 or m5 is the way to go.
AWS CPU credits are unique to T-series instances – and they can be a bit tricky to figure out. Whether you’re using the AWS free tier or just trying to use the smallest EC2 compute instance you can, you’ll need to keep track of these credits. These credits are both generated and used by the T2 and T3 instance families to decide how much CPU power you can actually use on those EC2 instances. This can be confusing if you aren’t expecting your virtual machine to have it’s CPU power throttled, or are wondering why the cost is much higher than you thought it would be.
AWS first released a “burstable” instance type in the form of the t1.micro instance size in 2010, which was four years after the first EC2 instance size was released (m1.small in 2006, for you historians). Up until 2010, new instance sizes had always been bigger than the m1.small size, but there was demand for a VM size that could accommodate low-throughput or inconsistent workloads.
The t1.micro was the only burstable instance size for another four years, until the t2.medium was released in 2014. Soon, there was a whole range of t2 instances to cover the use case of servers that were low-powered while idle, but could have lots of potential compute resources available for the couple minutes each hour they were needed. In 2018, AWS introduced the t3 family that uses more modern CPUs and the AWS Nitro system for virtualization.
AWS CPU Credits 101
The key reason why T-series instances have a lower list price than corresponding M-series instances (in standard mode, more on that later) is the CPU credits that are tracked and used on each resource. The basic premise is that an idle instance earns credits, while a busy instance spends those credits. A “credit” corresponds to 1 minute’s worth of full 100% CPU usage, but this can be broken down in different ways if the usage is less than 100%. For instance, 10% of CPU usage for 10 minutes also uses 1 credit. Each T-series machine size not only has a number of CPUs available, but also earns credits at different rates.
Here’s where the math starts getting a little tricky. A t2.micro instance earns 6 credits per hour with 1 available CPU. If you run that instance at 10% utilization for a full hour, it’ll spend 6 credits per hour (or 1 credit every 10 minutes). This means that any time spent under 10% utilization is a net increase in CPU credits, while any time spent above 10% utilization is a net decrease in CPU credits. A t3.large instance has 2 CPUs and earns 36 credits per hour, which means the balancing point where the net credit use is zero will be at 30% utilization per CPU.
So what happens when you run out of credits or never use your credits?
Standard Mode vs. Unlimited Mode
One of the differences between the t2 family and the t3 family is the default way each handles running out of credits. The t2 family defaults to Standard Mode, which means that once the instance has run out of credits to use, the CPU is throttled to the baseline value we calculated above (so 10% for t2.micro) and will continue maxing out at that value until credits have built back up. In practice, this means that your process or application that has burst up to use a lot more CPU than normal will soon be slow and unusable if the load remains high.
In 2017, AWS introduced Unlimited Mode as an option for t2 instances – and later, in 2018, as the default for t3 instances when they were introduced. Unlimited mode means that instead of throttling down to the baseline CPU when your instance runs out of credits, you can continue to run at a high CPU load and just pay for the overages. This price is 5¢ per CPU hour for Linux and 9.6¢ per CPU hour for Windows. In practice, this means that a t2.micro that has run out of credits and is running at 35% CPU utilization for a full 24 hours would cost an additional 30¢ that day on top of the normal 27.84¢ for 24hr usage, meaning the price is more than doubled.
Using T-series Instead of M-series
These overage charges for Unlimited Mode of t2 and t3 instances means that while the list price of the instance is much cheaper than corresponding m4 and m5 instances, you need to figure out if the utilization pattern of your workload makes sense for a burstable instance family. For example, an m5.large in us-east-1 costs 9.6¢/hr and a t3.large with similar specs costs 8.32¢/hr with a 30% CPU baseline. If your t3.large server is going to be running higher than 55.6% CPU for the hour on a consistent basis, then the price of the m5.large is actually lower.
When to Stop T-series and When to Let Them Run
One perk of using the t2 instances in Standard mode is that each time you start the server, you receive 30 launch credits that allow a high level of CPU usage when you first start the instance from a stopped state. These launch credits are tracked separately from accrued credits and are used first, so servers that only need to run short-lived processes when first starting can take advantage of this fact. The downside of stopping t2 servers is that accrued credits are lost when you stop the instance.
On the other hand, t3 servers persist earned credits for 7 days after stopping the instance, but don’t earn launch credits when they are first started. This is useful to know for longer-running processes that don’t have huge spikes, as they can build up credits but you don’t need to worry about losing the credits if you stop the server.
At ParkMyCloud, we specialize in scheduling servers and databases to turn off on a schedule, which is perfect for non-production servers. We find that lots of users have t2 and t3 instances for these dev and test workloads, but want to know what happens to credits if you park those servers overnight. As we discussed, AWS CPU credits go away in T2 standard mode (but with additional launch credits) but persist in T3 Unlimited mode. Knowing this, you can pick the right instance size for the workload you’re running and confidently save money using ParkMyCloud.
Best for non-production instances that have a quick burst of usage when starting = T2 instance with ParkMyCloud parking schedule
Best for non-production instances with unpredictable, but sporadic spikes = T3 instance with ParkMyCloud parking schedule
Try it for free to see how we can make the cost of your t2 and t3 servers even lower.
When it comes to AWS training resources, there’s no shortage of information out there. Considering the wide range of videos, tutorials, blogs, and more, it’s hard knowing where to look or how to begin. Finding the best resource depends on your learning style, your needs for AWS, and getting the most updated information available. Whether you’re just getting started in AWS or consider yourself an expert, there’s an abundance of resources for every learning level. With this in mind, we came up with our 7 favorite AWS training resources, sure to give you the tools you need to learn AWS:
1. AWS Self-Paced Labs
What better way to learn that at your own pace? AWS self-paced labs give you hands-on learning in a live AWS environment, with AWS cloud services, and actual scenarios you would encounter in the cloud. There are two different ways to learn with these labs. You can either take an individual lab or follow a learning quest. Individual labs are intended to help users get familiar with an AWS service as quickly as 15 minutes. Learning quests guide you through a series of labs so you can master any AWS scenario at your own pace. Once completed, you will earn a badge that you can boast on your resume, LinkedIn, website, etc.
Sometimes the best way to learn something is by jumping right in. With the AWS Free Tier, you can try AWS services for free. This is a great way to test out AWS for your business, or for the developers out there, to try services like AWS CodePipeLine, AWS Data Pipeline, and more. While you are still getting a hands-on opportunity to learn a number of AWS services, the only downside is that there are certain usage limits. You can track your usage with a billing alarm to avoid unwanted charges, or you can try ParkMyCloud and park your instances when they’re not in use so you get the most out of your free tier experience. In fact, ParkMyCloud started its journey by using AWS’s free tier!
3. AWS Documentation and Whitepapers
AWS Documentation is like a virtual encyclopedia of tools, terms, training, and everything AWS. You’ll find case studies, tutorials, cloud computing basics, and so much more. This resource is a one-stop-shop for all of your AWS documentation needs, whether you’re a beginner or advanced user. No matter where you are in your AWS training journey, AWS documentation is always a useful reference and certainly deserves a spot in your bookmarks.
Additionally, you’ll findwhitepapers that give users access to technical AWS content that is written by AWS and individuals from the AWS community, to help further your knowledge of their cloud. These whitepapers include things from technical guides, reference material, and architecture diagrams.
So far, we’ve gone straight to the source for 3 out of 7 of our favorite AWS training resources. Amazon really does a great job of providing hands-on training, tutorials, and documentation for users with a range of experience. However, YouTube opens up a whole new world of video training that includes contributions from not only Amazon, but other great resources as well. Besides the obvious Amazon Web Services channel, there are also popular and highly rated videos by Edureka, Simplilearn, Eli the Computer Guy, and more.
As cloud technology usage continues to expand and evolve, blogs are a great way to stay up to speed with AWS and the world of cloud computing. Of course, in addition to aws labs, a free-trial, extensive documentation, and their own YouTube channel, AWS also has their own blog. Since AWS actually has a number of blogs that vary by region and technology, we recommend that you start by following Jeff Barr – Chief Evangelist at Amazon Web Services, and primary contributor. Edureka was mentioned in our recommended YouTube channels, they also have a blog that covers plenty of AWS topics. The CloudThat blog is an excellent resource for AWS and all things cloud, and was co-founded by Bhaves Goswami – a former member of the AWS product development team. Additionally, AWS Insider is a great source for all things AWS. Here you’ll find blogs, webcasts, how-to, tips, tricks, news articles and even more hands-on guidance for working with AWS. If you prefer newsletters straight to your inbox, check out Last Week in AWS and Inside Cloud.
6. Online Learning Platforms
As public cloud computing continues to grow – and AWS continues to dominate the market – people have become increasingly interested in this CSP and what it has to offer. In the last 8-10 years, two massive learning platforms were developed, Coursera and Udemy. These platforms offer online AWS courses, specializations, training, and degrees. The abundance of courses that these platforms provide can help you learn all things AWS and give you a wide array of resources to help you train for different AWS certifications and degrees.
GitHub is a developer platform where users work together to review and host code, build software and manage projects. This platform has access to a number of materials that can help further your AWS training. In fact, here’s a great list of AWS training resources that can help you prepare for an Amazon Cloud certification. The great thing about this site is the collaboration among the users. The large number of people in this community brings together people from all different backgrounds so they are able to provide knowledge about their own specialties and experiences. With access to everything from ebooks, video courses, free lectures, and sample tests, posts like these can help you get on the right certification track.
There’s plenty of information out there when it comes to AWS training resources. We picked our 7 favorite resources for their reliability, quality, and range of information. Whether you’re new to AWS or consider yourself an expert, these resources are sure to help you find what you’re looking for.
In December, AWS announced a new service called AWS Compute Optimizer that provides recommendations with the goal of properly sizing EC2 virtual machines. Rightsizing is one of AWS’s listed five pillars of cost optimization, and it’s good to see AWS following the trend of cloud providers making it easier for customers to optimize for cost and performance. Actually, this is not the first “rightsizing tool” they’ve promoted. Early last year they pushed what was essentially a collection of Python scripts in the AWS Solutions Portal called “AWS Right Sizing”.
As cloud cost optimizers here at ParkMyCloud, rightsizing is high on the list of optimization strategies we focus on. The ParkMyCloud platform offers rightsizing recommendations and actions, along with two other cost optimization pillars: “Increase Elasticity” through scheduled shutdown of idle resources, and “Measure, monitor, and improve” through cost and savings reports and an RBAC-enabled user portal. Let’s take a look at what the AWS Compute Optimizer offers, and how it compares to ParkMyCloud’s rightsizing.
AWS Compute Optimizer Overview
The AWS Compute Optimizer service generates size change recommendations based on your existing EC2 servers, including those that are in Auto Scaling groups. Each EC2 virtual machine can get up to 3 recommendations for different families and sizes that you could choose, along with the performance risk and costs associated with each option. While you are browsing the options, the interface will show you what the performance would have looked like over the past 2 weeks if you were running on the selected instance size instead of the current instance size, which is nice for analyzing the options against your organization’s risk profile. However, there is no direct way to take the Rightsizing action, so you must go and adjust the instance settings manually.
AWS Compute Optimizer is free of charge and available on all AWS accounts regardless of support level. You do have to choose to opt-in to use the service before recommendations will be made. A major limiting factor is the region availability: as of February 4, 2020, AWS Compute Optimizer is available in 16 regions, and supports the M, C, R, T and X instance families. It uses only the past 2 weeks’ worth of Cloudwatch data to generate recommendations, which is a small window that may result in odd recommendations if those two weeks include any anomalies.
If your EC2 instances line up with this subset of instance types and regions, then the AWS Compute Optimizer can provide some suggestions for cost savings. However, if your needs are a little more diverse or robust, read on.
ParkMyCloud Rightsizing Overview
ParkMyCloud has offered scheduling of idle cloud resources since 2015. Last year we announced a major advancement in the platform’s cost optimization capabilities with the release of Rightsizing.
Similarly to the AWS Compute Optimizer, ParkMyCloud’s Rightsizing capabilities offer up to 3 recommendations for different sizes that your instances could be based on Cloudwatch data. Additionally, ParkMyCloud’s Rightsizing can:
ParkMyCloud is multi-cloud, multi-account, and multi-region in a single pane of glass, so you can view recommendations across all of your cloud accounts in one place (including all AWS regions, not just the ones listed above and Azure and Google Clouds)
ParkMyCloud can take the Rightsizing action for you once you accept a recommendation, including scheduling that resize action for a future time (such as during a maintenance window).
ParkMyCloud’s recommendations are based on data from a period of up to 24 weeks, providing a much more robust recommendation compared to the 2-week data set imposed by Cloudwatch.
ParkMyCloud makes recommendations for and resizes RDS databases, including Aurora instances. RDS databases have an average cost of 75% higher than EC2 instances, which means this is a significant opportunity for cost savings.
All AWS instance sizes are supported, not just M/C/R/T/X
Users can reject a recommendation and give an explanation, so administrators know why actions weren’t taken.
Savings from Rightsizing (and parking) are tracked and reported in ParkMyCloud, so you can show management or the CFO just how much money you’re saving the company.
Optimize Your Rightsizing
The AWS Compute Optimizer is a great feature that AWS is offering for free to its cloud users, but the limitations and inability to take direct action from the recommendations makes it less useful for serious cost optimization. ParkMyCloud’s features make it the right choice for saving money on your cloud bill while optimizing performance, and the free trial makes it easy to get started today. Feel free to contact us if you have any questions.
As cloud users continue to use Alibaba Cloud, extending its global presence, we’ll review a comparison of AWS vs Alibaba Cloud pricing. Commonly recognized as the #4 cloud provider (from a revenue perspective anyway), Alibaba is one of the fastest-growing companies in the space today.
Alibaba has been getting a lot of attention lately, given its rapid growth, and making headlines after the release of their latest quarterly revenue and full fiscal year 2019 reports.Alibaba is at the top of the market in Asia, and dominating in China with cloud revenue up 66% year-over-year. While Alibaba is in the top 5 CSPs worldwide, they still have a lot of plans for the future to maintain this growth and continue to move up.
The company said it is focused on high-value security, analytics, and artificial intelligence tools and “rationalizing our offerings of commodity products and services.” With an annual revenue run rate of $4.5 billion, it is clear that Alibaba Cloud intends to compete globally with AWS and other major cloud providers.
However, on a global scale, AWS continues to dominate the market. In the latest quarter, Amazon reported Amazon Web Services (AWS) sales of $7.7 billion, compared to $5.44 billion at this time last year. AWS revenue grew 41% in the first quarter – at this time last year, that number was 49%.
ParkMyCloud supports Alibaba Cloud and AWS, and with that, let us focus on pricing and cost savings – our forte. In this blog, we dive a bit into the pricing of Alibaba Elastic Compute Service (ECS), compare it with that of the AWS EC2 service and whether Alibaba Cloud computing can offer better value than AWS.
Alibaba ECS vs AWS EC2
Elastic Compute Service (ECS) and Elastic Cloud Compute (EC2), respectively, are the standard compute services offered by Alibaba Cloud and AWS.
Both cloud computing services provide the same core features:
The ability to choose from dozens of instance types.
Support for virtual as well as bare-metal servers.
Compatibility with a variety of Windows and Linux-based operating systems.
The ability to create custom images.
The major differences between Alibaba Cloud ECS and AWS EC2 are that Alibaba Cloud provides a wider range of instance families and that AWS offers more regions globally.
Alibaba vs Aliyun
Finding actual pricing for comparison purposes can be a bit complicated, as the prices are listed in a couple of different places and do not quite exactly match up since pricing varies between different instance types, and no instances from the two companies are identical. If one searches for Alibaba pricing, one ends up here, which I am going to call the “Alibaba Cloud” site. However, when you actually get an account and want to purchase an instance, you can up here or here, both of which I will call the “Aliyun” site. [Note that you may not be able to see the Aliyun sites without signing up for an account and actually logging-in.]
Aliyun (literally translated “Ali Cloud”) was the original name of the company, and the name was changed to Alibaba Cloud in July 2017. Unsurprisingly, the Aliyun name has stuck around on the actual operational guts of the company, reflecting that it is probably hard-coded all over the place, both internally and externally with customers. (Supernor’s 3rd Conjecture: Engineering can never keep up with Marketing.)
Both sites show that like the other major cloud providers, Alibaba’s pricing model includes a Pay-As-You-Go (PAYG) offering, with per-second billing. Note, however, that in order to save money on stopped instances, one must specifically enable a “No fees for stopped instances” feature. Luckily, this is a global one-time setting for instances operating under all Pay-As-You-Go VPC instances, and you can set it and forget it. Unlike AWS, this feature is not available for any instances with local disks (this and other aspects of the description lead me to believe that Alibaba instances tend to be “sticky” to the underlying hardware instance). On AWS, local disks are described as ephemeral and are simply deallocated when they are not in use. Like AWS, Alibaba Cloud system/data disks continue to accrue costs even when an instance is stopped.
Both sites also show that Alibaba also has a one-month prepaid Subscription model. Based on a review of the pricing listed for the us-east-1 region on the Alibaba Cloud site, the monthly subscription discount reflects a substantial 30-60% discount compared to the cost of a PAYG instance that is left up for a full month. For a non-production environment that may only need to be up during normal business hours (say, 9 hours per day, weekdays only), one can easily see that it may be more cost-effective to go with the PAYG pricing, and use the ParkMyCloud service to shut the instances down during off-hours, saving 73%.
But this is where the similarities between the sites end. For actual pricing, instance availability, and even the actual instance types, one really needs to dive into a live Alibaba account. In particular, if PAYG is your preference, note that the Alibaba public site appears to have PAYG pricing listed for all of their available instance types, which is not consistent with what I found in the actual purchasing console.
Low-End Instance Types – “Entry Level” and “Basic”
The Alibaba Cloud site breaks down the instance types into “Entry Level” and “Enterprise”, listing numerous instance types under both categories. All of the Entry Level instance types are described as “Shared Performance”, which appears to mean the underlying hardware resources are shared amongst multiple instances in a potentially unpredictable way, or as described by Alibaba: “Their computing performance may be unstable, but the cost is relatively low” – an entertaining description to say the least. I did find these instance types on the internal purchasing site, but did not delve any further with them, as they do not offer a point of reference for our AWS vs. Alibaba Cloud pricing comparison. They may be an interesting path for additional investigation for non-production instance types where unstable computing performance may be OK in exchange for a lower price.
That said…after logging in to the Alibaba management console, reaching the Aliyun side of the website, there is no mention of Entry Level vs Enterprise. Instead, we see the top-level options of “Basic Purchase” vs “Advanced Purchase”. Under Basic Purchase, there are four “t5” instance types. The t5 types appear to directly correspond to the first four AWS t2 instance types, in terms of building up CPU credits.
These four instance types do not appear to support the PAYG pricing model. Pricing is only offered on a monthly subscription basis. A 1-year purchase plan is also offered, but the math shows this is just the monthly price x12. It is important to note that the Aliyun site itself has issues, as it lists the t5 instance types in all of the Alibaba regions, but I was unable to purchase any of them in the us-east-1 region – “The configuration for the instance you are creating is currently not supported in this zone.” (A purchase in us-west-1, slightly more expensive, was fine).
The following shows a price comparison for Alibaba vs AWS for “t” instance prices in a number of regions. The AWS prices reflect the hourly PAYG pricing, multiplied by an average 730 hour month. I was not able to get pricing for any AWS China region, so the Alibaba pricing is provided for reference.
While the AWS prices are higher, the AWS instances are PAYG, and thus could be stopped when not being used, common for t2 instances used in a dev-test environment, and potentially saving over 73%. One can easily see that this kind of savings is needed to compete with the comparatively low Alibaba prices. I do have to wonder what is up with that Windows pricing in China….does Microsoft know about this??
Aliyun “Advanced Purchase”
Looking at the “Advanced” side of the Aliyun purchasing site, we get a lot more options, including Pay-As-You-Go instances. To keep the comparison simple, I am going to limit the scope here to a couple of instance types, trying to compare a couple m5 and i3 instances with their Alibaba equivalents. I will list PAYG pricing where offered.
In this table, the listed monthly AWS prices reflect the hourly pay-as-you-go price, multiplied by an average 730 hour month.
The italicized/grey numbers under Alibaba indicate PAYG numbers that had to be pulled from the public-facing website, as the instance type was not available for PAYG purchase on the internal site. From a review of the various options on the internal Aliyun site, it appears the PAYG option is not actually offered for very many standalone instance types on Alibaba…
The main reason I pulled in the PAYG prices from the second source was for auto scaling, which is normally charged at PAYG prices. In Alibaba, “all ECS instances that Auto Scaling automatically creates, or manually adds to a scaling group will be charged according to their instance types. Note that you will still be charged for Pay-As-You-Go instances even after you stop them.” It is possible, however, to manually add subscription-based instances to an auto scaling group, and configure them to be not removed when the group scales-down.
In general, the full price of the AWS Linux instances over a month is 22-35% higher than of an Alibaba 1-month subscription. A full price AWS Windows instance over a month is 9-25% higher than that of an Alibaba subscription. (And once again, it appears Windows licensing fees are not a factor in China.)
When it comes to Alibaba Cloud pricing vs AWS, Alibaba Cloud is trying to attract business and expand their global footprint by offering special promotions typically consisting of free trials, specially priced starter packages, and time-limited discounts on premium services. In many cases, taking advantage of these promotions could be useful in order to save money, but so is AWS.
AWS Introduces Savings Plans for EC2
Amazon also has their fair share of money-saving offerings as well. AWS announced the release of AWS Savings Plans – a new system for getting a discount on committed usage for EC2.
There are two kinds of Savings Plan:
Compute Savings Plan – Apply to EC2 usage regardless of instance family, size, AZ, region, OS, or tenancy. For any given instance configuration, pricing is similar (if not identical) to an equivalent Convertible RI, giving up to a 66% discount.
EC2 Instance Savings Plan – Specific to EC2 instances within a family in a specific region, but regardless of size, OS, or tenancy. For any given instance configuration, pricing is similar to an equivalent Standard RI, giving up to a 72% discount in exchange for the reduced flexibility.
AWS Reserved Instance new queuing option
You can now purchase reserved instances that, rather than going into effect immediately, are scheduled for future purchase.
Now, when planned correctly, you can avoid lapsing on Reserved Instance coverage for your workloads by scheduling a new reservation purchase to go into effect as soon as the previous one expires. The furthest in advance you can schedule a purchase is three years, which is also the longest RI term available.
However, AWS RI purchases have few limitations, they can be queued for regional Reserved Instances, but not zonal Reserved Instances. Regional RIs are the broader option as they cover any availability zone in a region, while zonal RIs are for a specific availability zone and actually reserve capacity as well.
AWS vs Alibaba Cloud Pricing: Alibaba is cheaper, but…
Alibaba definitely comes out as less expensive in this AWS vs Alibaba cloud pricing comparison – the one-month subscription has a definite impact. However, for longer-lived instances, AWS Reserved Instances will certainly be less expensive, running about 40-75% less expensive than AWS PAYG, and thus less than some if not all of the Alibaba monthly subscriptions. AWS RI’s are also more easily applicable to auto scaling groups than a monthly subscription instance.
For non-production instances that can be shut down when not in use, PAYG is less expensive for both cloud providers, where ParkMyCloud can help you schedule the downtime. The difficulty with Alibaba will actually be finding instances types that can actually be purchased with the PAYG option.
Each of the ‘big three’ cloud providers (AWS, Azure, GCP) offer a number of cloud certification options that individuals can get to validate their cloud knowledge and skill set, while helping them advance in their careers and broaden the scope of their achievements.
Between the different PaaS specific, role-based (such as dev. or architect) or domain focused certifications, CSPs have numerous options available to help you bring more value to your organization as you keep up with the new business demands and continue to challenge yourself and grow with this world. With these certifications, you are more likely to achieve business goals thanks to your proficiency in specific areas – and benefit from an extra edge on your resume in your next job search.
Here’s an overview of the certifications offered by AWS, Azure, and GCP and what capabilities an individual validates by completing these certifications.
Amazon Web Services (AWS) Certifications
AWS offers certifications for different learning levels. The four different categories/levels of certifications include:
Foundational: individuals should have at least six months of basic/foundational industry and AWS knowledge.
Associate: expected to have one year of experience solving problems and implementing solutions with AWS.
Professional: aimed for individuals that have two years of comprehensive experience operating, designing and solving solutions using AWS.
Specialty: each of the certifications in this category are based on a technical AWS experience in the specialty domain. Requirements for these certifications can range from a minimum of 6 months to 5 years of required hands-on experience.
Individuals are expected to effectively demonstrate a comprehensive understanding of AWS fundamentals and best practices.
AWS Certified Solutions Architect – Associate
Individuals in an associate solutions architect role have 1+ years of experience designing available, fault-tolerant, scalable, and most importantly cost-efficient, distributed systems on AWS.
Can demonstrate how to build and deploy applications on AWS.
AWS Certified SysOps Administrator – Associate
This certification is meant for systems administrators that hold a systems operations role and have at least one year of hands-on experience in management, operations and deployments on AWS.
They must be able to migrate on-premises workloads to AWS
They can estimate usage costs and identify operational cost control methods.
Must prove knowledge of deploying, operating and managing highly available, scalable and fault-tolerant systems on AWS.
AWS Certified Developer – Associate
This is for individuals who hold a development role and have at least one or more years of experience developing and maintaining AWS-based applications.
Display a basic understanding of core AWS services, uses, and basic AWS architecture best practices.
Demonstrate that they are capable of developing, deploying, and debugging cloud-based applications using AWS
AWS Certified Solutions Architect – Professional
Individuals in a professional solutions architect role have two or more years of experience operating and managing systems on AWS.
They must be able to design and deploy scalable, highly available, and fault-tolerant applications on AWS.
Must demonstrate knowledge of migrating complex, multi-tier applications on AWS
They are responsible for implementing cost-control strategies.
AWS Certified DevOps Engineer – Professional
Intended for individuals who have a DevOps engineer role and two or more years of experience operating, provisioning and managing AWS environments.
They are able to implement and manage continuous delivery systems and methodologies on AWS.
Additionally, they must be able to implement and automate security controls, governance processes, and compliance validation.
Can deploy and define metrics, monitoring and logging systems on AWS.
Are responsible for designing, managing, and maintaining tools that automate operational processes.
AWS Certified Advanced Networking – Speciality
Intended for individuals who perform intricate networking tasks.
Design, develop, and deploy cloud-based solutions using AWS
Design and maintain network architecture for all AWS services
Leverage tools to automate AWS networking tasks
AWS Certified Big Data – Speciality
For individuals who perform complex Big Data analyses and have at least two years of experience using AWS.
Implement core AWS Big Data services according to basic architecture best practices
Design and maintain Big Data
Leverage tools to automate data analysis
AWS Certified Security – Speciality
Individuals who have a security role and at least two years of hands-on experience securing AWS workloads.
Exhibit an understanding of specialized data classifications and AWS data protection mechanisms as well as data encryption methods and secure Internet protocols and AWS mechanisms to implement them
Knowledge of AWS security services and features to provide a secure production environment
An understanding of security operations and risk
AWS Certified Machine Learning – Speciality
Intended for individuals in a development or data science role.
Ability to design, implement, deploy and maintain machine learning solutions for specific business problems.
AWS Certified Alexa Skill Builder – Speciality
Intended for individuals who have a role as an Alexa skill builder.
Individuals have demonstrated an ability to design, build, test, publish and manage Amazon Alexa skills.
Microsoft Azure Certifications
Following the Azure learning path under Microsoft, there are certifications available that allow you to demonstrate your expertise in Microsoft cloud-related technologies and advance your career by earning one of the new Azure role-based certifications or an Azure-related certification in platform, development, or data.
Intended for individuals that have an expertise in network, compute, security and storage so that they can design solutions that run on Azure
Individuals will prove their understanding of cloud concepts, Azure pricing and support, core Azure services, as well as the fundamentals of cloud privacy, security, trust and compliance.
Azure DevOps Engineer Expert
Individuals will demonstrate an ability to combine people, process, and technologies to continuously deliver valuable products and services that meet business objectives in addition to end user needs.
Azure Developer Associate
For individuals that can design, build, test and maintain cloud solutions – such as applications and services – and partner with cloud solutions architects, cloud administrators, cloud DBAs, and clients in order to implement these solutions.
Azure Data Scientist Associate
Intended for individuals that apply Azure’s machine learning techniques to train, evaluate, and deploy models that will ultimately help solve business problems.
Azure Data Engineer Associate
For individuals that design and implement the management, security, monitoring, and privacy of data – using the full stack of Azure data services – to satisfy business needs.
Azure AI Engineer Associate
Intended for individuals that use Machine Learning, Knowledge Mining, and Cognitive Services to architect and implement Microsoft AI solutions – this involves natural language processing, computer vision, speech, agents and bots.
Azure Administrator Associate
Individuals must demonstrate their ability to implement, monitor and maintain Azure solutions – this includes major services related to storage, compute, security and network.
Azure Security Engineer Associate
Individuals are expected to be able to implement security controls and threat protection, manage identity and access. Additionally, they must be able to protect data, applications, and networks in the cloud as well as hybrid environments as part of end-to-end infrastructure.
Azure for SAP Workloads Specialty
In this specialty, architects have extensive experience and knowledge of the SAP Landscape Certification process and industry standards that are specific and critical to the long-term operation of an SAP solution.
Azure IoT Developer Specialty
In this specialty, individuals must prove that they understand how to implement the Azure services that form an IoT solution – this includes data analysis, data processing, data storage options, and PaaS options.
Must be able to recognize Azure IoT service configuration settings within the code portion of an IoT solution.
Google offers three different levels of available certifications:
Associate certification – focused on the fundamental skills of deploying, monitoring, and maintaining projects on Google Cloud.
This certification is a good starting point for those new to cloud and can be used as a path to professional level certifications.
Recommended experience: 6+ months building on Google Cloud
Professional certification – span key technical job functions and assess advanced skills in design, implementation, and management.
These certifications are recommended for individuals with industry experience and familiarity with Google Cloud products and solutions.
Recommended experience: 3+ years of industry experience, including 1+ years on Google Cloud
User certification – intended for individuals with experience using G Suite and determines an individual’s ability to use core collaboration tools.
Recommended experience: Completion of Applied Digital Skills training course and G Suite Essentials quest, and 1+ months on G Suite.
Intended for individuals that can deploy applications, monitor operations, and manage enterprise solutions.
Individuals display an ability to use the Google Cloud Console and the command-line interface to perform common platform-based tasks to maintain one or more deployed solutions that leverage Google-managed or self-managed services on Google Cloud.
Individuals display an ability to set up a cloud solution environment, plan and configure a cloud solution, deploy and implement a cloud solution, ensure successful operation of a cloud solution, and configure access and security.
Professional Cloud Architect
For individuals that enable organizations to leverage Google Cloud technologies.
These individuals can design, develop, and manage secure, scalable, and highly available solutions that drive business objectives.
Individuals display an ability to design and plan a cloud solution architecture, manage and provision the cloud solution infrastructure, design for security and compliance, analyze and optimize technical and business processes, manage implementations of cloud architecture, and ensure solution and operations reliability.
Professional Cloud Developer
These individuals build scalable and highly available applications using Google recommended practices and tools that leverage fully managed services.
Have experience with next generation databases, runtime environments, and developer tools.
Have proficiency with at least one general purpose programming language and are skilled in using Stackdriver.
Individuals display an ability to design highly scalable, available, and reliable cloud-native applications, build and test applications, deploy applications, integrate Google Cloud Platform services, and manage application performance monitoring.
Professional Data Engineer
Intended for individuals that enable data-driven decision making by collecting, transforming, and publishing data.
Individuals should be able to design, build, operate, manage, and monitor secure data processing systems.
Individuals display an ability to design data processing systems, build and operationalize data processing systems, operationalize machine learning models, and ensure solution quality.
Professional Cloud DevOps Engineer
Individuals are responsible for efficient development operations that can balance service reliability and delivery speed.
Individuals are expected to be skilled in using Google Cloud Platform to build software delivery pipelines, deploy and monitor services, and manage and learn from incidents.
Individuals display an ability to apply site reliability engineering principles to a service, optimize service performance, implement service monitoring strategies, build and implement CI/CD pipelines for a service, and manage service incidents.
Professional Cloud Security Engineer
Intended for individuals that enable organizations to design and implement a secure infrastructure on Google Cloud Platform.
They are expected to have a thorough understanding of security best practices and industry security requirements.
These individuals design, develop, and manage a secure infrastructure leveraging Google security technologies and should be proficient in all aspects of Cloud Security.
Individuals display an ability to configure access within a cloud solution environment, configure network security, ensure data protection, manage operations within a cloud solution environment and ensure compliance.
Professional Cloud Network Engineer
Intended for individuals who implement and manage network architectures in Google Cloud Platform.
These individuals ensure successful cloud implementations using the command line interface or the Google Cloud Platform Console.
Individuals display an ability to design, plan, and prototype a GCP Network, implement a GCP Virtual Private Cloud (VPC), configure network services and implement hybrid interconnectivity.
Professional Collaboration Engineer
Intended for individuals that transform business objectives into tangible configurations, policies, and security practices as they relate to users, content, and integrations.
Individuals use tools, programming languages, and APIs to automate workflows.
Individuals display an ability to plan and implement G Suite authorization and access, manage user, resource, and Team Drive lifecycles, manage mail, control and configure G Suite services, configure and manage endpoint access, monitor organizational operations and advance G Suite adoption and collaboration.
G Suite User – User Certification
This certification lets employers know that you possess the digital skills to work collaboratively and productively in a professional environment, complete common workplace activities using cloud-based tools to create and share documents, spreadsheets, presentations, and files.
Where to Start
If you aren’t sure where to start, each cloud provider offers a certification that only requires a basic understanding of the platform and are a great way to help you get the ball rolling in your cloud certification journey. The three certifications for beginners are: AWS Certified Cloud Practitioner, Microsoft Certified Azure Fundamentals, and Google Associate Cloud Engineer. Good luck!