Introduction

I remember my first exposure to cloud computing quite clearly. I attended an enterprise architecture meetup, and someone from Amazon was talking about a new service from the company called Amazon Web Services (AWS) that offered what he referred to as “Infrastructure as a Service” (this was so early in AWS’s life that the term cloud computing had not even invented).

AWS streamlined access to industry-standard X86 computing resources. Instead of the interminable provisioning timelines typical of on-premises environments, AWS delivered virtual machines in less than ten minutes. To someone used to waiting weeks or months for computing resources, this seemed like sorcery.

Learn about the AWS architectural principles and services like IAM, VPC, EC2, EBS, and more with the AWS Solutions Architect Course. Register today

It was immediately clear to me that Infrastructure as a Service would revolutionize the technology industry. And so it has proved. Users flocked to AWS and its peers because they could design and run applications faster and cheaper than ever before.

Demand for cloud computing is so large today that the providers are building data centers at a furious pace. According to a Data Center Knowledge article, the big three of cloud—AWS, Microsoft, and Google (aka AMG) are spending about $30 billion per year on new infrastructure.

Each of them has moved far beyond industry-standard X86 servers. They employ in-house staff to create new hardware designs tuned to their environments. They work with Intel to develop new chips better suited for their use cases. One can think of them as implementing custom X86 computing environments designed to enable a massive scale while operating at the lowest possible cost point.

The rise of machine learning (ML) changes this formula. While X86-based computing is excellent for common application workloads, it’s not nearly as well-suited for ML workload execution. That is because parallel processing, which is a hallmark of ML execution, is a poor match for the X86 single thread processing approach.

GPU boards are often used for ML, as the parallel processing of graphics chips is a better match for these workloads. All of the big three currently or will soon offer virtual machines with GPU boards attached to allow customers to execute ML workloads.

However, while GPU parallel processing is better than X86 chips for machine learning, these graphics-oriented chips are not ideally suited for machine learning. Stated another way, GPUs perform ML more efficiently and less expensively than X86 environments, but there are additional capabilities available with ML-focused chip designs. This has caused Google and Microsoft to move beyond system design and into the realm of chip manufacture. Simply put, the demands of ML require specialized processors, and there isn’t an Intel equivalent for ML chips—so AMG has stepped forward to implement them.

Interestingly, AMG has diverged in their approach to ML processors, which reflects their overall approach to ML.

Do you wish to become a cloud expert? Gain the right skills with our Cloud Computing Certification Course and excel in your career, starting today!

TensorFlow

Google created an open-source ML framework called TensorFlow and is putting all of its ML efforts behind it. It focuses on a single framework and can implement the TensorFlow code directly in silicon—a term meaning that it has implemented the framework in an ASIC it refers to as a tensor processing unit. TPUs offer impressive performance gains compared to the CPU and GPU alternatives. To learn more on this topic, read the Google blog that describes their TPU initiative. Google launched its public TPU service in 2016 and has since followed up with an updated version a few months ago.

AWS Solutions Architect Certification

Brainwave

In contrast to Google, Microsoft is not wedded to a single ML framework. Its employees use several open-source frameworks (including TensorFlow), depending on the task at hand. This dictates against an ASIC approach, so the company recently announced an FPGA-based service called Brainwave.

Microsoft’s approach offers flexibility to its researchers and engineers. This gives them performance well beyond GPUs, but without restricting them to a single ML framework. Brainwave is currently only available to Microsoft employees, but the company plans on making it available to external customers in the future.

The quiet one in this flurry of hardware announcements? AWS. As noted, it does offer GPU capability, but to date has not announced custom hardware.

However, after a bit of a slow start, AWS is going headlong after AI, and the company is unlikely to be willing to take a back seat to the other members of AMG. I expect the upcoming Reinvent conference is likely to witness one or more ML hardware announcements from the company.

One might question the ML hardware investments these companies are making. Sure, the new hardware will run ML workloads better than X86 CPUs or even GPUs, but why go to the work and expense of designing and building entire new chip architectures?

In a phrase: customer demand. The use of machine learning is exploding with the technology applied across an enormous range of use cases, from clothing recommendation to railway maintenance to medical diagnosis.

Machine learning is now poised where cloud computing was a decade ago and will be accelerated by the same phenomenon: easy availability at a low price. Its growth is likely to be more rapid, though.

With cloud computing, we’ve spent a decade in debating whether on-premises or public clouds are better, with adoption lagging due to the ongoing controversy.

Machine learning will not undergo the same kind of bickering. ML’s natural home is a public offering because it improves with scale and data, which are both more available from a public provider than in a single user location.

It will be interesting to see if, in a few years we’ll see AMG doubling down with even more investment as they add ML facilities to their data center arms race. For sure, though, we’re only at the beginning of their custom hardware efforts.

Our Cloud Computing Courses Duration and Fees

Cloud Computing Courses typically range from a few weeks to several months, with fees varying based on program and institution.

Program NameDurationFees
Post Graduate Program in DevOps

Cohort Starts: 27 Nov, 2024

9 months$ 4,849
Post Graduate Program in Cloud Computing

Cohort Starts: 27 Nov, 2024

8 months$ 4,500
AWS Cloud Architect Masters Program3 months$ 1,299
Cloud Architect Masters Program4 months$ 1,449
Microsoft Azure Cloud Architect Masters Program3 months$ 1,499
Microsoft Azure DevOps Solutions Expert Program10 weeks$ 1,649
DevOps Engineer Masters Program6 months$ 2,000

Get Free Certifications with free video courses

  • Introduction to Cloud Computing

    Cloud Computing & DevOps

    Introduction to Cloud Computing

    2 hours4.683K learners
  • Introduction to Google Cloud Platform

    Cloud Computing & DevOps

    Introduction to Google Cloud Platform

    4 hours4.514.5K learners
prevNext

Learn from Industry Experts with free Masterclasses

  • Cloud Architect Career: Must-Know Trends and Technologies in 2024

    Cloud Computing

    Cloud Architect Career: Must-Know Trends and Technologies in 2024

    23rd Oct, Wednesday7:00 PM IST
  • AWS vs. Azure vs. GCP: Pros and Cons for Aspiring Cloud Professionals

    Cloud Computing

    AWS vs. Azure vs. GCP: Pros and Cons for Aspiring Cloud Professionals

    5th Sep, Thursday9:00 PM IST
  • Azure Deployment in Action: Setting Up VNets and Peering for OSS

    Cloud Computing

    Azure Deployment in Action: Setting Up VNets and Peering for OSS

    30th May, Thursday9:00 PM IST
prevNext