We recently published a case study, Fast and Reliable Ranking in Datastore, that describes how we helped one of our Google App Engine customers shorten their ranking latency from one hour to five seconds. They applied unique design patterns such as job aggregation to achieve over 300 updates per second with strong consistency on Cloud Datastore. The following are highlights from the article. ...
We recently published a case study, Fast and Reliable Ranking in Datastore, that describes how we helped one of our Google App Engine customers shorten their ranking latency from one hour to five seconds. They applied unique design patterns such as job aggregation to achieve over 300 updates per second with strong consistency on Cloud Datastore. The following are highlights from the article.



The problem of ranking

Tomoaki Suzuki, an App Engine lead engineer at Applibot, a major game studio in Japan, has been trying to solve the common, yet difficult problem faced by every large gaming service: ranking.




Tomoaki Suzuki, App Engine lead engineer at Applibot, Inc. and their game Legend of Criptids (#1 ranked game in the Apple App Store North America gaming category in October 2012)



The requirements are simple:




  • Your game has hundreds of thousands (or more!) players.

  • Whenever a player fights enemies (or performs other activities), their score changes.

  • You want to show the latest ranking for the player on a web portal page.


Getting a rank is easy, if it's not expected to also be scalable and fast. For example, you could execute the following query:

SELECT count(key) FROM Players WHERE Score > YourScore



This query counts all the players who have a higher score than yours. But do you want to execute this query for every request from the portal page? How long would it take when you have a million players?



Tomoaki initially implemented this approach, but it took a few seconds to get each response. This was too slow, too expensive, and performed progressively worse as scale increased.




The easiest way: scan all players

Next, Tomoaki tried to maintain ranking data in Memcache. This was fast, but not reliable, because Memcache entries are just caches and could be evicted at any time. With a ranking service that depended solely on in-memory-key-values, it was difficult to maintain consistency and availability.



Looking for an O(log n) Algorithm

I was assigned to Applibot under a platinum support contract. I knew that ranking was a classic and yet hard-to-solve problem for any scalable distributed service. The simple query solution requires scanning all players with a higher score to count the rank of one player. The time complexity of this algorithm is O(n); that is, the time required for query execution increases proportionally to the number of players. In practice, this means that the algorithm is not scalable. Instead, we need an O(log n) or faster algorithm, where the time will only increase logarithmically as the number of players grows.



If you ever took a computer science course, you may remember that tree algorithms, such as binary trees, red-black trees, or B-Trees, can perform at O(log n) time complexity for finding an element. Tree algorithms can also be used to calculate an aggregate value of a range of elements, such as count, max/min, and average by holding the aggregated values on each branch node. Using this technique, it is possible to implement a ranking algorithm with O(log n) performance.



I found an open source implementation of a tree-based ranking algorithm for Datastore, written by a Google engineer: the Google Code Jam Ranking Library.




Getting the rank of a score in a tertiary tree with google Code Jam Ranking Library

Concurrent Updates Limit Scalability

However, during load testing, I found a critical limitation with the Code Jam ranking library. Its scalability in terms of update throughput was quite low. When he increased the load to three updates per second, the library started to return transaction retry errors. It was obvious that the library could not satisfy Applibot's requirement for 300 updates per second. It could handle only about 1% of that throughput.



Why is that? The reason is the cost of maintaining the consistency of the tree. In Datastore, you must use an entity group to assure strong consistency when updating multiple entities in a transaction—see "Balancing Strong and Eventual Consistency with Google Cloud Datastore". The Code Jam ranking library uses a single entity group to hold the entire tree to ensure consistency of the counts in the tree elements.



However, an entity group in Datastore has a performance limitation. Datastore only supports about one transaction per second on an entity group. Furthermore, if the same entity group is modified in concurrent transactions, they are likely to fail and must be retried. The Code Jam ranking library is strongly consistent, transactional, and fairly fast, but it does not support a high volume of concurrent updates.



Datastore Team's Solution: Job Aggregation

I remembered that a software engineer on the Datastore team had mentioned a technique to obtain much higher throughput than one update per second on an entity group. This could be achieved by aggregating a batch of updates into one transaction, rather than executing each update as a separate transaction. So Kaz asked the Datastore team for a solution for this problem.



In response to my request, the Datastore team started discussing this issue and advised us to consider using Job Aggregation, one of the design patterns used with Megastore, the underlying storage layer of Datastore, that manages the consistency and transactionality of entity groups. The basic idea of Job Aggregation is to use a single thread to process a batch of updates. Because there is only one thread and only one transaction open on the entity group, there are no transaction failures due to concurrent updates. You can find similar ideas in other storage products such as VoltDb and Redis.



Proposed Solution Runs at 300 Updates per Second Sustained

Based on the advice from the Datastore team, I wrote Proof of Concept (PoC) code that combines the Job Aggregation pattern with the Code Jam ranking library. The PoC creates a pull queue, which is a kind of Task Queue in App Engine that allows developers to implement one or multiple workers that consume the tasks added to the queue. The backend instance has a single thread in an infinite loop that keeps pulling as many tasks as possible (up to 1000) from the queue. The thread passes each update request to the Code Jam ranking library, which executes them as a batch in a single transaction. The transaction may be open for a second or more, but because there is a single thread driving the library and Datastore, there is no contention and no concurrent modification problem.



The following figure shows the load testing result of the final PoC implementation. Kaz also incorporated another design pattern, Queue Sharding, to effectively minimize the performance fluctuations in each task queue. With the final proposed solution, it can sustain 300 updates per second over several hours. Under usual load, each update is applied to Datastore within a few seconds of receiving the request.




Performance graph of the solution

With the load testing results and the PoC code, I presented the solution to Tomoaki and other Applibot engineers. Tomoaki plans to incorporate the solution in their production system, expects to reduce the latency of updating the ranking info from one hour to five seconds, and hopes to dramatically improve the user experience.



-Posted by Kazunori Sato, Solutions Architect



Notes

Any performance figures described in this article are sampled values for reference and do not guarantee any absolute performance of App Engine, Datastore, or other services.

Today’s guest blog comes from Casey Wilms, product lead for cloud media services provider Brightcove, and Lee Chen, head of product at content delivery network Fastly. Fastly leverages Brightcove's Zencoder cloud-based encoding service, which runs on Google compute, to offer a powerful Live video streaming solution. ...
Today’s guest blog comes from Casey Wilms, product lead for cloud media services provider Brightcove, and Lee Chen, head of product at content delivery network Fastly. Fastly leverages Brightcove's Zencoder cloud-based encoding service, which runs on Google compute, to offer a powerful Live video streaming solution.



In the streaming video world, streaming live events without hiccups is a bit like the Holy Grail. Now, with a bundled transcoding and delivery package from Fastly, customers can leverage the power of Brightcove’s Zencoder cloud transcoding solution and Fastly's unparalleled content delivery network to offer events with an impeccable user experience.



Background on Brightcove and Fastly

Brightcove's Zencoder is a cloud-based transcoding solution for Live video and VOD. Customers like Funny or Die and SmugMug have built their video workflow around Zencoder to encode large volumes of video affordably and quickly. Performance is paramount for Zencoder, so Brightcove relies on the Google Cloud Platform for consistently great service around the globe.



Fastly is a next-generation content delivery network (CDN) that can cache any type of content and has zero-delay, instant purge (in ~150 milliseconds) across dynamic content, compared to two minutes (or even as long as six hours) on a traditional CDN. This allows Fastly to truly cache dynamic content for businesses like Twitter, Github, and Foursquare. Since Fastly’s goal is to deliver content that is as close to real time as possible, it made sense to package Zencoder as part of a broader live video delivery solution. Fastly’s event-based transcoding, powered by Zencoder, offers businesses a faster and more seamless video streaming service. For example, one customer test streamed video from Mexico to the central U.S. and had virtually no latency, around 50 milliseconds.



How the Brightcove/Fastly Integration Works

Let’s say you have a live event that you’re streaming online for your audience. Your production team provides you with a Program Out signal to your live encoders. You generate the source signal, which is sent to Fastly’s event-based transcoding, powered by Zencoder. Zencoder then transcodes that signal into different bitrates and resolutions (renditions), and Fastly picks it up and distributes it to the end user’s app or Web player. When the event is over, the transcoding instance spins down, and you won’t have to worry about paying for more time than you’re broadcasting for.

fastly_target_diagram-01.png

Why Brightcove Chose Google Cloud Platform

One of the benefits in using Fastly and Brightcove’s live online video delivery package is that Zencoder runs on Google Compute Engine, which means everything is hosted in the cloud. Brightcove tried other infrastructure hosting platforms, and found that only Google Compute Engine delivers the kind of high responsiveness, consistency, and performance needed to offer seamless live video streaming.



Because Google Cloud Platform is built on the same network Google uses, network latency is extremely low, resulting in best-in-class speed and reliability. Thus, Brightcove can spin up server instances as needed to meet spikes in demand. Google has achieved a level of consistency across their cloud in terms of compute and network performance that is simply impossible to find with other vendors. That means the 100th server that's spun up can be expected to exhibit the same compute profile as the first. Instances start up in about half the time of Amazon. Also, thanks to Google Compute Platform's per minute billing granularity and automatic discounts, customers using the live online video delivery package will get an unbeatable price point.



Another thing Brightcove doesn’t get from anyone else is the level of customer and engineering support that Google offers. If the team has an issue, they can email Google’s support and get a response right away. They can end up talking to a product manager, and if they file a support ticket, they get a call the next day, which is totally unheard of in the cloud platform world.



Faster Streaming, Happier Customers

Because Google Cloud Platform launches instances in less than half the time of the rest of the industry, Fastly is able to launch new customers through Brightcove in a turnkey way. If a customer has an on-demand broadcast, like a music broadcast or sporting event, that gets a choke in the CPU, it translates to latency, and viewers end up watching a stutter step or getting completely disconnected. If anything goes wrong with the live stream, it’s urgent to fix the problem immediately, since it’s impossible to redo an event that happens live. The amount of support required to satisfy customers for a live event is an order of magnitude higher than other types of video service. Thanks to Google Cloud Platform's consistency and speed, Fastly and Brightcove can rest comfortably knowing their customers will be happy.



-Contributed by Casey Wilms, product lead for media processing technology at Brightcove, and Lee Chen, Head of Product at Fastly

Cross-posted from the Google Analytics Blog



With 10.6 million cell phone customers and retail stores in 400+ markets, U.S. Cellular needs to reach a lot of people with marketing messages. That's why U.S. Cellular uses many marketing channels -- online, in-store and telesales -- to drive mobile phone activations.
Cross-posted from the Google Analytics Blog



With 10.6 million cell phone customers and retail stores in 400+ markets, U.S. Cellular needs to reach a lot of people with marketing messages. That's why U.S. Cellular uses many marketing channels -- online, in-store and telesales -- to drive mobile phone activations.



U.S. Cellular was challenged though. They didn’t know how many of their offline sales were driven by their digital marketing. This made it harder to adjust their media mix accordingly and also to forecast sales. To fix that situation, U.S. Cellular and its digital-analytics firm, Cardinal Path, turned to Google Analytics Premium and its integration with Google BigQuery.



Part of Google Cloud Platform, BigQuery allows for highly flexible analysis of large datasets. The U.S. Cellular team used it to integrate and analyze terabytes of data from Google Analytics Premium and other systems. Then they mapped consumer behavior across online and offline marketing channels. Each transaction was attributed to the consumer touchpoints that the buyer had made across various sales channels.



The result: U.S. Cellular got real insight into digital’s role in their sales. They were surprised to find that they could reclassify nearly half of all their offline activations to online marketing channels.



U.S. Cellular now uses this complete (and fully automatic) analytics framework to really see the consumer journey and forecast sales for each channel. Their team has the data they need to make better business decisions.



“We’re now in the enviable position of having an accurate view at each stage of our customer journey," says Katie Birmingham, a digital & e-commerce analyst for the company. "The Google Analytics Premium solution not only gives us a business advantage, but helps us shape a great customer experience, and ultimately ties in to our values of industry-leading innovation and world-class customer service.”



Be sure to read the full case study.



-Posted by: Suzanne Mumford, Google Analytics Premium Marketing

Editor’s note: Today’s guest blog comes from Ron Zalkind, co-founder and CTO of Waltham, Massachusetts-based CloudLock, a leading cloud security provider. The largest organizations in the world trust CloudLock to secure their data in the cloud, increase collaboration, and reduce their risk. ...
Editor’s note: Today’s guest blog comes from Ron Zalkind, co-founder and CTO of Waltham, Massachusetts-based CloudLock, a leading cloud security provider. The largest organizations in the world trust CloudLock to secure their data in the cloud, increase collaboration, and reduce their risk.



At a time when more and more organizations are moving their most sensitive data assets and applications to the cloud, security takes center stage, often at the price of user productivity. At CloudLock, we believe that each and every organization should work hard to protect their data and users in the cloud, not from it. With this philosophy in mind, CloudLock provides cloud security applications that help over 700 businesses using Google Apps and Salesforce to enforce regulatory, operational and security compliance.



We’ve been building enterprise products on top of Google Cloud Platform, specifically Google App Engine, for four years now. Collaborating with Google early on allowed us to leverage its best-of-class infrastructure security and scalability - both paramount for us as a security provider.



Our business must be as agile as our customers. Their accelerated SaaS platform adoption means that our business is data intensive, continuously processing changes in billions of objects and thousands of third-party applications connected to Google Apps user accounts. We’ve built a massive real-time data processing solution using App Engine, enabling us to focus on delivering core value for customers instead of focusing on infrastructure. App Engine’s auto-scaling capabilities provide our customers with a security solution that grows with their business, and its development features enable us to release code improvements frequently and seamlessly.



As a SaaS company, high-quality service is our bond. To keep delivering best-in-class service our team leverages central management features available through the admin panel such as Google BigQuery and Google Cloud Storage for advanced service analysis, monitoring and delivery. We have also been using Premiere Support since its launch, which boosts our ability to provide enterprise-level customer support.



As a premier Google Apps partner, we rely on Cloud Platform to provide enterprise-class cloud security to over five million users. It’s a beautiful synergy.



-Posted by Ron Zalkind, co-founder and CTO of CloudLock

A new, free Udacity online course, Developing Scalable Apps with Google App Engine, helps Java developers learn how to build scalable App Engine applications. As you work through the course, you'll build a conference management application that lets users create and query conferences.
A new, free Udacity online course, Developing Scalable Apps with Google App Engine, helps Java developers learn how to build scalable App Engine applications. As you work through the course, you'll build a conference management application that lets users create and query conferences.



magnus pizza boxes 5-cropped.png


The course starts with an entertaining introduction to Platform as a Service(PaaS). Magnus Hyttsten, Google Developer Advocate, discusses the evolution of server-side computing, from apps that could run on a computer under your desk, to applications that require the computing power of a data center. (This is not without disadvantages, as he points out. "It is no longer possible to warm your feet on the fan outlet.")



Urs Hölzle, Senior VP of infrastructure at Google, gives the background on Google Cloud Platform: "We built our internal cloud a long time ago, for our own needs. We had very large-scale applications, so we needed a very capable cloud, and now we're making our cloud available to everyone. In Cloud platform, App Engine is the one system that makes it really easy for you to start very small and then scale to a very large user base."



just the sandwich 3.png


After learning about the evolution of the data center and the need for scalability, you'll get right down to business and learn how to store data in the Datastore, use Memcache to speed up responses and cut down on Datastore quota usage, write queries, understand indexes, and use queues for tasks that execute outside front end requests.



Along the way, you'll build a backend App Engine application that uses Google Cloud Endpoints to expose its API to other applications.





conf-central-code.png


You'll learn how to implement Endpoints to make the API available externally, and how to use the Endpoints API from an Android application.



If you take this course, you'll not only learn about App Engine, but you'll use it behind the scenes too. Udacity uses App Engine to serve its online courses. Mike Sokolsky, Udacity co-founder and CTO, talks about Udacity's decision to use App Engine to host Udacity's MOOCs. He says, "It pushes you in the right direction. It pushes you to the best design practices for building a scalable application." And that's what this course aims to help you do, too.



You can take the course, Developing Scalable Applications with App Engine, at www.udacity.com/course/ud859.



The full course materials — all the videos, quizzes, and forums — are available for free for all students by selecting “View Courseware”. Personalized ongoing feedback and guidance from coaches is also available to anyone who chooses to enroll in Udacity’s guided program.



For more courses that Google and Udacity are developing together, see www.udacity.com/google.



"

-Posted by Jocelyn Becker, Developer Advocate

Starting today, Google Cloud Monitoring Read API is generally available, allowing you to programmatically access metric data from your running services, such as CPU usage or disk IO. For example, you can use Cloud Monitoring Read API with Nagios to plug in to your existing alerting/event framework, or use it with Graphite to combine the data with your existing graphs. Third party providers can also use the API to integrate Google Cloud Platform metrics into their own monitoring services.
Starting today, Google Cloud Monitoring Read API is generally available, allowing you to programmatically access metric data from your running services, such as CPU usage or disk IO. For example, you can use Cloud Monitoring Read API with Nagios to plug in to your existing alerting/event framework, or use it with Graphite to combine the data with your existing graphs. Third party providers can also use the API to integrate Google Cloud Platform metrics into their own monitoring services.



Cloud Monitoring Read API allows you to query current and historical metric data for up to the past 30 days. Also, you can use labels to filter data to more specific metrics (e.g. zones). Currently Cloud Monitoring Read API supports reading metric time series data from the following Cloud Platform services:


  • Google Compute Engine - 13 metrics

  • Google Cloud SQL - 12 metrics

  • Google Cloud Pub/Sub - 14 metrics




Our documentation provides a full list of supported metrics. Over time we will be adding support for more Cloud Platform services metrics and enhancing the metrics for existing services. You can see an example of usage and try these metrics for yourself on our getting started page. For samples and libraries, click here.



Example: getting CPU usage time series data

GET \
https://www.googleapis.com/cloudmonitoring/v2beta1/ \ # Access API
projects/YOUR_PROJECT_NAME/ \ # For YOUR_PROJECT_NAME
timeseries/ \ # get time series of points
compute.googleapis.com%2Finstance%2Fcpu%2Fusage_time?\ # of CPU usage
youngest=2014-07-11T10%3A29%3A53.108Z& \ # with this latest timestamp
key={YOUR_API_KEY} # using this API key

Your feedback is important!

We look forward to receiving feedback and suggestions at cloud-monitoring-feedback@googlegroups.com.



-Posted by Amir Hermelin, Product Manager

Today’s guest blog comes from Rafael Sanches, engineer at Allthecooks, a social media platform for people who love cooking. Allthecooks is available on Android, iPhone, Windows Phone, Google Glass and Android Wear, and is a top recipe app on Google Play. ...
Today’s guest blog comes from Rafael Sanches, engineer at Allthecooks, a social media platform for people who love cooking. Allthecooks is available on Android, iPhone, Windows Phone, Google Glass and Android Wear, and is a top recipe app on Google Play.



At Allthecooks, we’re connecting passionate chefs with casual and first-time cooks on every major mobile device including Android phones, tablets, watches and Google Glass. People use our app to find, rate and comment on dishes they can cook themselves, or post their own ideas complete with directions, ingredients, servings and nutrition information. We have chefs with tens of thousands of followers on Allthecooks, meaning that whenever they make an update we have to process hundreds of thousands of simultaneous API requests to feed that information to timelines of all their followers.



Creating a successful platform isn’t just about speed, it’s about scalability, too. Google Cloud Platform played a key role in helping us grow without worrying about our architecture. We launched in December 2012 with just three part-time engineers and have never taken funding, so building our own infrastructure was out of the question. Since launching, we’ve grown to over 12 million users with a million monthly active users. Our application now sees millions of interactions daily that run through Google App Engine and Google Cloud Datastore.



As our user base has grown, we’ve begun migrating the biggest pieces of our backend processing architecture from App Engine onto Google Compute Engine. This will allow us to operate at even higher performance levels by using the cheaper CPU and caching more data in the large RAM configurations supported by Compute Engine instances. We also plan to use Google BigQuery soon to make the process of finding the perfect recipe even easier.



Cost was a big concern when we were burning through our savings to build Allthecooks, but it wasn’t as important as making everything run as fast as possible. When a request hits, the response needs to be immediate. That’s why we built our recommendation engine on Compute Engine -- that lets us run entirely on RAM, so users get the results they need as soon as they need them. When you’ve got 60 seconds to sear a fish just right, you can’t be caught waiting on a laggy server request to see the next step in a recipe. We never want our app latency to stand between our users and a great meal.



There are plenty of cooking apps out there, but none with the same level of social interaction as Allthecooks, which is one of the biggest reasons our users spend so much time on the app. We let users ask questions and receive answers on every recipe, and upload pictures of their own dishes. Our recommendation engine in particular plays a pivotal role in making the Allthecooks experience so useful on mobile devices. Google Glass users love that they can find and follow a recipe instructions hands-free or tilt their heads up to see the ingredient list. They can even record their own recipes using voice, photos and video and send it to their Allthecooks account. All of this requires a reliable infrastructure.



We don’t have a system administrator, so we need a system that’s reliable, manageable and scalable while requiring minimal oversight. Cloud Platform gives us confidence that the app is stable, and it can even be left alone for a week without requiring maintenance. Google’s support team has also given us peace of mind. They’re always quick to respond, and they have provided great services the few times we needed help.



It gives us great pride that Allthecooks has helped helped millions of people live healthier, and discover new foods and products they love. We believe in launching early and listening to what our customers want. Cloud Platform is a crucial component to our success to date and strategic for our future growth.



-Posted by Rafael Sanches, co-founder and engineer of Allthecooks

If you saw our post about Cassandra hitting 1 million writes per second on Google Compute Engine, then you know we’re getting serious about open source NoSQL. We’re making it easier to run the software you love at the scale you need with the reliability of Google Compute Platform. With over a dozen different ...
If you saw our post about Cassandra hitting 1 million writes per second on Google Compute Engine, then you know we’re getting serious about open source NoSQL. We’re making it easier to run the software you love at the scale you need with the reliability of Google Compute Platform. With over a dozen different virtual machine types, and the great price for performance of persistent disks, we think Google Compute Engine is a fantastic place for Apache Cassandra.



Today, we’re making it even easier to launch a dedicated Apache Cassandra cluster on Google Compute Engine. All it takes is one click after some basic information such as the size of the cluster. In a matter of minutes, you get a complete Cassandra cluster deployed and configured.



Each node is automatically configured for the cloud including:


  • Configured with the GoogleCloudSnitch for Google Cloud Platform awareness

  • Writes tuned for Google Persistent Disk

  • JVM tuned to perform on Google Compute Engine instances




The complete set of tuning parameters can be found on the Click to Deploy help page.



So, get out and click to deploy your Cassandra cluster today!



Learn more about running Apache Cassandra on Google Compute Engine at https://developers.google.com/cloud/cassandra.



-Posted by Brian Lynch, Solutions Architect



Cassandra, is registered trademarks of Apache, Inc. All other trademarks cited here are the property of their respective owners.

We’ve had a great time giving you our predictions for the World Cup (check out our post before the quarter-finals and semi-finals). So far, we’ve gotten 13 of 14 games correct. But this isn't about us picking winners in World Cup soccer - it’s about what you can do with Google Cloud Platform. Now, we are open-sourcing our prediction model and packaging it up so you can do your own analysis and predictions.
We’ve had a great time giving you our predictions for the World Cup (check out our post before the quarter-finals and semi-finals). So far, we’ve gotten 13 of 14 games correct. But this isn't about us picking winners in World Cup soccer - it’s about what you can do with Google Cloud Platform. Now, we are open-sourcing our prediction model and packaging it up so you can do your own analysis and predictions.



We used Google Cloud Dataflow to ingest raw, touch-by-touch gameplay data from Opta for thousands of soccer matches. This data goes back to the 2006 World Cup, three years of English Barclays Premier League, two seasons of Spanish La Liga, and two seasons of U.S. MLS. We then polished the raw data into predictive statistics using Google BigQuery.



You can see BigQuery engineer Jordan Tigani (+JordanTigani) and developer advocate Felipe Hoffa (@felipehoffa) talk about how we did it in this video from Google I/O.



Our prediction for the final

It’s a narrow call, but Germany has the edge: our model gives them a 55% chance of defeating Argentina due to a number of factors. Thus far in the tournament, they’ve had better passing in the attacking half of their field, a higher number of shots (64 vs. 61) and a higher number of goals scored (17 vs. 8).



But, 55% is only a small edge. And, although we've been trumpeting our 13 of 14 record, picking winners isn't exactly the same as predicting outcomes. If you'd asked us which scenario was more likely, a 7 to 1 win for Germany against Brazil or a 0 to 1 defeat of Germany by Brazil, we wouldn't have gotten that one quite right.



(Oh, and we think Brazil has a tiny advantage in the third place game. They may have had a disappointing defeat on Tuesday, but the numbers still look good.)



But don’t take our word for it...

Now it’s your turn to take a stab at predicting. We have provided an IPython notebook that shows exactly how we built our model and used it to predict matches. We had to aggregate the data that we used, so you can't compute additional statistics from the raw data. However, for the real data geeks, you could try to see how well neural networks can predict the same data or try advanced techniques like principal components analysis. Alternatively, you can try adding your own features like player salaries or team travel distance. We've only scratched the surface, and there are lots of other approaches you can take.



You might also try simulating how the USA would have done if they had beat Belgium. Or how Germany in 2014 would fare against the unstoppable Spanish team of 2010. Or you could figure out whether the USA team is getting better by simulating the 2006 team against the 2010 and 2014 teams.



Here’s how you can do it

We’ve put everything on GitHub. You’ll find the IPython notebook containing all of the code (using pandas and statsmodels) to build the same machine learning models that we've used to predict the games so far. We've packaged it all up in a Docker container so that you can run your own Google Compute Engine instance to crunch the data. For the most up-to-date step-by-step instructions, check out the readme on GitHub.



-Posted by Benjamin Bechtolsheim, Product Marketing Manager

Today’s guest blog comes from Jim Totton, Vice President and General Manager, Platform Business Unit at Red Hat



Red Hat Enterprise Linux Atomic Host is now available as a technology preview on Google Compute Engine for customers participating in the ...
Today’s guest blog comes from Jim Totton, Vice President and General Manager, Platform Business Unit at Red Hat



Red Hat Enterprise Linux Atomic Host is now available as a technology preview on Google Compute Engine for customers participating in the Red Hat Enterprise Linux 7 Atomic special interest group (SIG). The inaugural SIG is focused on application containers and encompasses the technologies that are required to create, deploy and manage application containers.



Google and Red Hat actively collaborate on container technologies, approaches and best practices. Both companies are committed to standards for container management, interoperability and orchestration. As a gateway to the open hybrid cloud, application containers enable new possibilities for customers and software providers, including application portability, deployment choice and hyperscale and resilient architectures - whether on-premise or in the cloud.



At Red Hat Summit in April, Red Hat announced our vision for Linux Containers and expanded the Red Hat Enterprise Linux 7 High Touch Beta program to include Red Hat Enterprise Linux Atomic Host – a secure, lightweight and minimal footprint operating system optimized to run Linux Containers. Moving forward, Red Hat will work closely with these participants, with assistance from Google, to support them as they explore application containers. This will help us both gather important requirements and feedback on use cases for these technologies and enable the hybrid cloud for our joint customers.



We also announced today that Red Hat and Google are collaborating to tackle the challenge of how to manage application containers at scale, across hundreds or thousands of hosts. Red Hat will be joining the Kubernetes community and actively contributing code. Earlier today on our blog, we wrote:




Red Hat is embracing the Google Kubernetes project and plans to work to enable it with container management capabilities in our products and offerings. This will enable Red Hat customers to take advantage of cluster management capabilities in Kubernetes, to orchestrate Docker containers across multiple hosts, running on-premise, on Google Cloud Platform or in other public or private clouds. As part of this collaboration, Red Hat will become core committers to the Kubernetes project. This supports Red Hat’s open hybrid cloud strategy that uses open source to enable application portability across on-premise datacenters, private clouds and public cloud environments.



Both Google and Red Hat recognize the importance of delivering containerized applications that are secure, supported and exhibit a chain of trust. Red Hat's Container Certification program, launched in March 2014, supports this commitment and is designed to help deliver containerized applications that “work as intended” to trusted destinations within the hybrid cloud for software partners and end-customers.



Follow the Red Hat Enterprise Linux Blog to stay informed about Red Hat’s work on technologies required to create, deploy, and manage application containers.



-Contributed by Jim Totton, Vice President and General Manager, Platform Business Unit at Red Hat

Kubernetes is an open source manager for Docker containers, based on Google’s years of experience using containers at Internet scale. Today, Microsoft, RedHat, IBM, Docker, Mesosphere, CoreOS and SaltStack are joining the Kubernetes community and will actively contribute to the project. Each company brings unique strengths, and together we will ensure that Kubernetes is a strong and open container management framework for any application and in any environment - whether in a private, public or hybrid cloud.
Kubernetes is an open source manager for Docker containers, based on Google’s years of experience using containers at Internet scale. Today, Microsoft, RedHat, IBM, Docker, Mesosphere, CoreOS and SaltStack are joining the Kubernetes community and will actively contribute to the project. Each company brings unique strengths, and together we will ensure that Kubernetes is a strong and open container management framework for any application and in any environment - whether in a private, public or hybrid cloud.



Our shared goal is to allow a broad range of developers to take advantage of container technologies. Kubernetes was built from the ground up as a lean, extensible and portable framework for managing Docker workloads. It lets customers manage their applications the way that Google manages hyper-scale applications like Search and Gmail.



Containers offer tremendous advantages for developers. Predictable deployments and simple scalability are possible because Docker packages all of a workload’s dependencies with the application. This allows for ultimate portability; you can avoid vendor lock-in and run containers in the cloud of your choice. It is just as important that the management framework has the same properties of portability and scalability, and that is what the community will bring to Kubernetes.



We look forward to the contributions of the expanded Kubernetes community:




  • Microsoft is working to ensure that Kubernetes works great in Linux environments in Azure VMs. Scott Guthrie, Executive Vice President of the Cloud and Enterprise group at Microsoft told us, “Microsoft will help contribute code to Kubernetes to enable customers to easily manage containers that can run anywhere. This will make it easier to build multi-cloud solutions including targeting Microsoft Azure.”


  • Red Hat is working to bring Kubernetes to the open hybrid cloud. Paul Cormier, President, Products and Technologies at Red Hat, told us, “Red Hat has a rich history of contributing to and maturing innovative, open source projects. Through this collaboration with Google on Kubernetes, we are contributing to the evolution of cloud computing and helping deliver the promises that container technologies offer to the open hybrid cloud.”

  • IBM is contributing code to Kubernetes and the broader Docker ecosystem to ensure that containers are enterprise-grade, and is working with the community to create an open governance model around the project.

  • Docker is delivering the full container stack that Kubernetes schedules into, and is looking to move critical capabilities upstream and align the Kubernetes framework with Libswarm.

  • CoreOS is working to ensure that Kubernetes can work seamlessly with the suite of CoreOS technologies that support cloud-native application development on any cloud.

  • Mesosphere is actively integrating Kubernetes with Mesos, making the advanced scheduling and management capabilities available to Kubernetes customers.

  • SaltStack is working to make Kubernetes a portable container automation framework that is designed for the reality of the platform-agnostic, multi-cloud world.






You can view the Go source and documentation for Kubernetes on GitHub. We look forward to the contributions of these companies alongside the already vibrant open source community.



- Posted by Urs Hölzle, Senior Vice President

Today’s guest blog comes from David LaBine, director of education software for SMART Technologies. SMART Technologies creates interactive displays that encourage collaboration in business and education. The company recently launched SMART amp, a cloud-based collaborative learning tool for classrooms. ...
Today’s guest blog comes from David LaBine, director of education software for SMART Technologies. SMART Technologies creates interactive displays that encourage collaboration in business and education. The company recently launched SMART amp, a cloud-based collaborative learning tool for classrooms.



At SMART Technologies, we don’t want to just build a product and then close up the development shop. We want to keep making it better and add in new features without getting sidetracked by development chores like creating virtual machines. Google App Engine is giving us the freedom to build our SMART amp classroom software on a faster, more flexible timeline – which helps students to learn better and teachers to teach more effectively.



We’re pioneers in the interactive displays market, and we introduced the first interactive whiteboard way back in 1991. SMART amp, our newest product for education, is a logical progression in our line of collaborative products for workplaces and schools. It’s a cloud-based tool that allows teachers and students to connect using any mobile device and bring together all kinds of content in a collaborative workspace. To keep things organized, SMART amp lets teachers manage workflows, passing control of the workspace to different students or groups so that everyone gets involved. Teachers can even embed assessments in their workspaces to gauge student progress.



Before we began building SMART amp, we knew we needed a continuous deployment pipeline between our testing and production environments. App Engine allowed us to get started with continuous deployment quickly without having to build virtual machines or configuring infrastructure. Furthermore, we’ve noticed a tendency to over provision when using virtual machine setups, which means you end up developing more infrastructure than you need to get the job done – and paying more than you need to.



When we checked out various development platforms, we found that App Engine gave us the continuous development and cost savings we needed. Whenever a developer writes a piece of code, it automatically runs through a gauntlet of progressively more rigorous tests – from simple unit tests to expansive integration tests involving all of the services in the product. If the tests pass, the code goes directly to production. As a result, developers are more productive because they’re able to focus on writing new features rather than worrying about infrastructure, and they have a quick feedback loop when something unexpectedly breaks.



We’ve also saved time and have thrilled customers by integrating SMART amp with other important Google products like Google Drive and Google Apps for Education. These services make it so we can leverage millions of user identities and associated file storage already provisioned and in use by schools, greatly decreasing the time to get started with our product. The best part is that our GCP Technical Account Manager (TAM) has been a one-stop shop for all our support needs for other Google products. For example, when we had to work through some difficult Google Drive API use cases, our TAM was able to work internally to find a resolution even though the APIs weren’t explicitly covered by a support agreement or SLA.



Scaling as needed is also helping us manage costs and avoid the over provisioning problem. We can expand our use of App Engine when we have large amounts of traffic, but we can also manage our usage so that it matches the load on the applications, which cuts our costs by 23%.



When we see the results of our development projects in classrooms using SMART amp, it makes the hard work we’ve done to build this product all worthwhile. One of the first lessons our initial users created was on globalization. It was inspiring to watch students start talking to each other and get creative, pulling in news stories, pictures and maps into lessons about the diamond trade and oil production. App Engine plays a big role in making us enjoy coming to work every day.



-Posted by David LaBine, Director of education software for SMART Technologies

Last week, we told you about going 8 for 8 in predicting the Round of 16. We also gave you a sneak peek at our predictions for the quarterfinals.
Last week, we told you about going 8 for 8 in predicting the Round of 16. We also gave you a sneak peek at our predictions for the quarterfinals.



Perhaps in our excitement about using Cloud Dataflow, BigQuery and Compute Engine to arrive at our predictions, we may have been better served by heeding a more simple truth: Gary Lineker once said, “Football is a simple game. 22 men chase a ball for 90 minutes, and at the end, the Germans always win.”



And so it went in the quarterfinals. We gave France a 69% chance, but we were wrong. Germany defeated France 1 to 0. That was the only game that upset our predictions. After going 8 for 8 in the Round of 16, we're now at 11 for 12.



Why did we get Germany - France wrong?

World Cup teams are especially difficult to model because they play so few games together. USA coach Jurgen Klinsmann recently told the New York Times that he sees his players about as often as he sees his barber. If data is the lifeblood of a good model, we suffered for want of more information.



But, we know that in the same environment, others fared better in their predictions (h/t Cortana; their model relies more on what betting markets are saying, whereas ours is an inductive model derived from game-play data).



So, why did we get Germany - France wrong? In the first four games of the World Cup, France took more shots than Germany, had more shots on target, and their shots were from a more “dangerous location” (that is, closer to the goal). This information complements actual goals to form an ‘expected goals’ statistic in our model.



Moreover, in the first four games, Germany allowed their opponents to take more dangerous shots, and thus the expected goal statistic was higher for their opponents. And, it allowed their opponents to pass better in their third of the field. In the Germany-France game, France actually outshot Germany with 13 shots vs. 8 for Germany, and 9 vs. 6 on-target. With a little more luck on their side, they may have pulled ahead.



What about the semi-finals?

Enough commentary. Here’s our predictions for the next round:




  • Brazil vs. Germany: Germany (59%)

  • Netherlands vs. Argentina: Argentina (61%)




These predictions are based on the results of the quarterfinal games in addition to the previous World Cup games. Our model doesn’t take into account the fact that Brazilian striker Neymar is out of the tournament with a back injury, so the scales might actually tip a bit further in Germany’s favor.



- Posted by Benjamin Bechtolsheim, Product Marketing Manager

Webydo is a full B2B solution for professional code-free web design led by a rapidly growing global community of more than 92,000 designers, design studios and web design agencies.



To cope with its exponential growth, Webydo recently moved its infrastructure from a hosted VMWare private cloud to ...
Webydo is a full B2B solution for professional code-free web design led by a rapidly growing global community of more than 92,000 designers, design studios and web design agencies.



To cope with its exponential growth, Webydo recently moved its infrastructure from a hosted VMWare private cloud to Google Cloud Platform. Webydo now has access to our easily scalable and secure infrastructure, which can enable the company to grow exponentially. As a result of the migration, Webydo has also saved 37% on infrastructure costs. Learn more about Webydo's new architecture by reading the technical case study here.



-Posted by Ori Weinroth, Product Marketing Manager

At Google I/O, we talked about how you can use Google Cloud Platform to analyze large datasets, build statistical models and use machine learning to predict the outcome of future events. And we demonstrated the technology by analyzing the World Cup.
At Google I/O, we talked about how you can use Google Cloud Platform to analyze large datasets, build statistical models and use machine learning to predict the outcome of future events. And we demonstrated the technology by analyzing the World Cup.



As much as it broke the hearts of those of us rooting for Team USA, we got every game in the Round of 16 right. World Cup 2010 had Paul the Octopus. In 2014, we’re bringing Cloud Platform to the table.



If you don’t believe our record, you can watch Google Developer Advocate Felipe Hoffa (@felipehoffa) and BigQuery engineer Jordan Tigani (+JordanTigani) predict the outcomes at Google I/O on June 26. (Jump to minute 25 if you want to go right to the predictions).



Our model was built using touch-by-touch data from Opta covering multiple seasons of professional soccer leagues as well as the group rounds of the World Cup. Using data from the many leagues where World Cup players play, we were able to examine how behavior in previous games predicted performance in subsequent games. Our model also relied on two other data sources: (1) a power ranking that was built by BigQuery engineer Jordan Tigani and (2) a subjective measure of team support based on enthusiasm and the number of fans who had traveled to Brazil. This took the place of a simple “home team advantage” in the model.



We used Google Cloud Dataflow to ingest the data, Google BigQuery to build the derived features, iPython and Pandas to do the modeling, and Google Compute Engine to crunch the data.



Now, we know we’re not the only ones to get this round right. The Round of 16 didn’t feature any major upsets. However, we’re happy to show off how Cloud Platform can be used for doing machine learning and predictive analytics.



Other models use a poisson regression to calculate goal scores, but we found that a logistic regression to predict the winner rather than final score provided a more accurate prediction of future games.



And, now what you’re waiting for… We updated our model with the touch-by-touch data from the Round of 16 to predict the results of the quarterfinals. Here you go:




  • Brazil vs. Colombia: Brazil (71%)

  • France vs. Germany: France (69%)

  • Netherlands vs. Costa Rica: Netherlands (68%)

  • Argentina vs. Belgium: Argentina (81%)




We’ll check in after the games to let you know how we’re doing. And, we’ll give you a deeper look at our methodology as well.



-Posted by Benjamin Bechtolsheim, Product Marketing Manager

If you thought May was busy, June probably left your head spinning.
If you thought May was busy, June probably left your head spinning.



Docker Containers in App Engine and Kubernetes

We started the month at DockerCon, where we took the lid off our efforts to make Containers first class citizens in Google Cloud Platform. Now, you can deploy Docker images in Managed VMs or use our newly-released, open-source container manager Kubernetes to deploy containers across a fleet of VMs. The source and documentation for Kubernetes are hosted on GitHub - and we were excited to see 55,000 views, 917 stars, and 89 forks in the first day alone.



SSD Persistent Disks and global load balancing

The week after DockerCon, we released the SSD persistent disks as well as global HTTP load balancing, an example of how we can enable new networking capabilities through software-defined networking.



Discussing the evolution of computing at Google

That same week, Senior Vice President Urs Hölzle spoke at GigaOm’s Structure conference. You can hear him talk about the evolution of computing at Google and how we are externalizing our technology to offer a great public cloud.



One-Click Rabbit MQ and MongoDB

We also continued to make it easy for you to run open source solutions on Cloud Platform. You can now deploy Rabbit MQ and MongoDB with just one click. We have a new command-line tool in the Google Cloud SDK with great Windows support, improved scripting, and a much more.



Importing data to Google Cloud Storage is faster than ever

And, it’s faster than ever to import data into Google Cloud Platform thanks to Cloud Storage Online Cloud Import.



Cloud Platform documentation is better than ever

Our developer documentation page got a facelift. Check out developers.google.com/cloud to see for yourself.



Google I/O

And, to round up a month of big launches, we gave you a sneak peek at several new products & features at Google I/O. Urs led Cloud Platform’s section of the Google I/O keynote, which you can watch here. Also, we launched a number of new services & products:




  • Cloud Trace and Cloud Debugger for Diagnosing Systems in Production: Cloud Trace and Cloud Debugger allow you to understand, diagnose, and improve systems in production. If you want to see Cloud Debugger in action, you can check out our demo during the I/O Keynote beginning at the 2 hour mark, or read more about it on our blog.

  • Analyze Huge Datasets with Cloud Dataflow: We demonstrated Cloud Dataflow, a managed service for creating data pipelines that ingest, transform and analyze data in both batch and streaming modes. You can see us analyze World Cup results using Dataflow. We also released a preview of Cloud Pub/Sub, helping you connect apps and devices that send data into processing pipelines.

  • Mobile Developers: Save User Data to the Cloud without Backend Code: We introduced a number of new features that make it easier to build mobile applications, including a new version of Google Cloud Save, which gives you a simple API for saving, retrieving, and synchronizing user data to the cloud and across devices without the need for backend code. We also put new tooling in Android Studio that simplifies the process of adding an App Engine backend to your mobile app. You can read more about these new features here.




Rising Star on ABC, Secret, and Coca-Cola

But it wasn’t just product launches that kept us busy. We’re always happy to see people using Cloud Platform to build great applications. This month, Rising Star premiered on ABC, with real-time, interactive voting powered by Cloud Platform. We’ve also been excited by the success of Secret, which is built on Cloud Platform and has experienced 1000-fold growth in just two months - all with only one backend developer. We featured their story during the I/O keynote. And, finally, CI&T launched the Google Cloud Platform-powered Coca-Cola “Happiness Flag,” unveiled on the pitch at the opening match of the World Cup.



Connect with our team in a city near you

If you missed us at DockerCon, GigaOM Structure, or Google I/O there’s still a chance to connect with our engineering team at the North American Cloud Platform Developers roadshow.



-Posted by Benjamin Bechtolsheim, Product Marketing Manager