Feb 9, 2024
This writing series provides a systematic approach to productionalizing machine learning pipelines with Kubeflow on Kubernetes. Building machine learning models is just one piece of a more extensive system of tasks and processes that come together to deliver a Machine Learning product. Kubeflow makes it possible to leverage the microservices paradigm of containerization to separate modular components of an application orchestrated on Kubernetes.
Series: Kubeflow for Poets
Mar 16, 2020
The article provides an example for deploying applications on Cloud Run that makes API calls. It provides sample codes for using Google Cloud Text-to-Speech APIs and running a simple web application with Flask that is packaged as a container and deployed on Cloud Run. The entire application is open-source at https://github.com/dvdbisong/text-to-speech-cloud-run.
Sep 24, 2019
Online communities are a significant part of interactions over the internet. The rise of such communities comes with the need for community moderation to ensure that participants adhere to their prescribed guidelines and avoid “clear and obvious” and objective policy violations. Google Cloud AutoML for Natural Language provides the platform for designing and developing custom language models for language recognition use-cases. This article uses Google Cloud AutoML for Natural Language to develop an end-to-end language toxicity classification model to identify obscene text.
Jun 6, 2019
Google Cloud Run provides a managed platform for running stateless containers. As the Compute infrastructure is managed, it can auto-scale in response to increased user-traffic. Cloud Run Services receives web traffic via HTTP requests or via asynchronous events, like a Pub/Sub message. This project uses Cloud Run to run a stateless container that employs Pandas profiling to display the summary statistics of a structured CSV dataset.
May 10, 2019
Google Cloud AutoML Vision simplifies the creation of custom vision models for image recognition use-cases. This article uses Google Cloud AutoML Vision to develop an end-to-end medical image classification model for Pneumonia Detection using Chest X-Ray Images.
Apr 22, 2019
This workshop shows how to build Keras models on GCP Notebook Instances. The Keras Model is the core of a Keras programme. A Model is constructed, compiled, trained and evaluated using their respective training and evaluation datasets. Upon satisfactory evaluation, the model is used to make predictions on previously unseen data.
Mar 27, 2019
A Kubeflow pipeline component is an implementation of a pipeline task. A component is a step in the workflow. Each task takes one or more artifacts as input and may produce one or more artifacts as output.
Mar 27, 2019
Kubeflow Pipelines is a simple platform for building and deploying containerized machine learning workflows on Kubernetes. Kubeflow pipelines make it easy to implement production grade machine learning pipelines without bothering on the low-level details of managing a Kubernetes cluster.
Mar 27, 2019
Kubeflow is a platform that is created to enhance and simplify the process of deploying machine learning workflows on Kubernetes. Using Kubeflow, it becomes easier to manage a distributed machine learning deployment by placing components in the deployment pipeline such as the training, serving, monitoring and logging components into containers on the Kubernetes cluster.
Mar 27, 2019
When a microservice application is deployed in production, it usually has many running containers that need to be allocated the right amount of resources in response to user demands. Also, there is a need to ensure that the containers are online, running and are communicating with one another. The need to efficiently manage and coordinate clusters of containerized applications gave rise to Kubernetes.
Mar 27, 2019
Docker is a virtualization application that abstracts applications into isolated environments known as containers. The idea behind a container is to provide a unified platform that includes the software tools and dependencies for developing and deploying an application.
Mar 27, 2019
The Microservice architecture is an approach for developing and deploying enterprise cloud-native software applications that involve separating the core business capabilities of the application into decoupled components. Each business capability represents some functionality that the application provides as services to the end-user.
Mar 27, 2019
Machine learning is often and rightly viewed as the use of mathematical algorithms to teach the computer to learn tasks that are computationally infeasible to program as a set of specified instructions. However, the use of these algorithms happen to constitute only a tiny fraction of the overall learning pipeline from an Engineering perspective. Building high-performant and dynamic learning models includes a number of other critical components.
Mar 27, 2019
This writing series provides a systematic approach to productionalizing machine learning pipelines with Kubeflow on Kubernetes. Building machine learning models is just one piece of a more extensive system of tasks and processes that come together to deliver a Machine Learning product. Kubeflow makes it possible to leverage the microservices paradigm of containerization to separate modular components of an application orchestrated on Kubernetes.
Mar 7, 2019
Jul 25, 2018
Google Cloud partnership with Pythian could never be better at this time. As Diana Green put it, "customers really win when the tech companies partner". Pythian is firmly stationed as a key partner in the cloud ecosystem with Google, working hand-in-hand to bring enterprise data work-loads and products to the cloud and to enable companies to optimize their cloud strategy and develop their cloud competencies.
Jan 19, 2018
This post provides an introduction to probability theory and random variables for pattern recognition. The approach is not to provide formal, rigorous proofs - but outlines proofs by example. This part covers the fundamentals of discrete random variables.
Jan 17, 2018
Google Datalab and Amazon SageMaker have fully managed cloud Jupyter notebooks for designing and developing machine learning and deep learning models by leveraging serverless cloud engines. However, they exist key differences between the two offerings as much as they have a lot in common. This post carries out a comparative analysis to examine the subtle differences and similarities between the two cloud-based machine learning as-a-service platforms.
Jan 14, 2018
Amazon SageMaker is another cloud-based fully managed data analytics/ machine learning modeling platform for designing, building and deploying data models. The key selling point of Amazon SageMaker is "zero-setup". This post takes a tour through spinning up a SageMaker notebook instance for data analytics/ modeling learning models.
Aug 23, 2017
Aug 3, 2017
Learning is a non-trivial task. How we learn deep representations as humans are high up there as one of the great enigmas of the world. What we consider trivial and to some others natural is a complex web of fine-grained and intricate processes that indeed have set us apart as unique creations in the universe both seen and unseen. In this post, I explain in simple terms the origins and promise of deep learning.
Jul 28, 2017
Machine learning is a technology that has grown to prominence over the past ten years (as at this time of writing) and is fast paving the way for the “Age of Automation”. This post provides a holistic view of the vital constituents that characterizes machine learning. At the end of this piece, the reader can be able to grasp the major landmarks and foundation stones of the field. Also, this overview provides a structured framework to wade deeper into murkier waters without getting overly overwhelmed.
Jul 21, 2017
In a previous post titled <a href="/machine-learning-cloud-notes-layman/" target="_blank" rel="noopener">Machine Learning on the cloud</a>, we examined in plain language what is machine learning, what is the cloud, and the merits of leveraging cloud resources for machine learning practice. In this post, I introduce GCP as a simple, yet powerful, and cost effective cloud option for performing machine learning. Whats more, I provide a simple walkthrough on how to set up the environment for machine learning model development on GCP.
Jul 12, 2017
Computational expenses have always been the bane of large-scale machine learning. In this post, I explain the fundamentals of Machine Learning on the cloud and the opportunities of unbridled computational horsepower made available by leveraging cloud infrastructures.
Jan 16, 2017
YouTube as of yet remains the world’s biggest video content delivery network, and each day is responsible for recommending videos to billions of users worldwide. This paper by a team of engineers at Google provides an overview of the data representation methods applied in developing a video recommendation system for YouTube and the performance gains observed using deep learning.