Attention Is All You Need they said. Is it a more robust convolution? Is it just a hack to squeeze more learning capacity out of fewer parameters? Is it supposed to be sparse? How did the original authors come up with this architecture?
multiple
pipelineslocal
execution mode and a deployment
execution mode. This ensures the creation of 2 separate running configurations, with the first being used for local development and end-to-end testing and the second one used for running in the cloud.Reuse code
across pipeline variants if it makes sense to do soCLI interface
for executing pipelines with different configurations
and dataA correct implementation also ensures that tests are easy to incorporate in your workflow.
Named inputs and outputs are essentially dictionaries with string keys and tensor values.
Most machine learning pipelines read data from a structured source ( database, CSV files/ Pandas Dataframes , TF Records), perform feature selection, cleaning, (and possibly) preprocessing, passing a raw multidimensional array (tensor) to a model along with another tensor representing the correct prediction for each input sample.
Reorder or rename input features in production? → Useless results or the client — side breaks in production
Absent Features? Missing Data? Bad…
The following is a very, very brief and non-technical summary of a chapter in the author’s dissertation.
A very deep neural network that is designed to solve hard problems might take a few seconds to run on modern computers with hardware accelerators for linear algebra operations. Similarly, smaller neural networks that do not take that much time to run, still do not meet realtime constraints. Hardware resources and execution time constraints is what drives the research community to investigate different methods of neural network compression. In the next few sections, common compression methods are presented.
The tf.data.Dataset API is a very efficient pipeline builder. Time Series Tasks can be a bit tricky to implement properly. In this article, we are going to dive deep into common tasks:
Let’s begin!
With the dataset api this is simple to do. Assume the following configuration. input feature is a
and label is b
.
a, b
1, 0
2, 0
3, 1
4, 0
5, 0
6, 1
Each row can be described by a tensor shaped (2,)
…
You all know what this game is about. This is the best service-offline-sorry page in the world. People have made simple bots that time the dino’s jump to beat the game to reinforcement learning agents with CNN state encoders.
It’s a game and we’re supposed to have fun. Today, I’ll walk you through how to write some JavaScript code to play the game by jumping around in your room.
This thing is hard to play.
You can try the game here and view the full source code here.
Setting up a small webpage with basic javascript support to get…
If this production e2e ML pipelines thing seems new to you, please read the TFX guide first.
On the other hand, if you’ve used TFX before, or planning to deploy a machine learning model, you’re in the right place.
TensorFlow Extended (TFX) is an end-to-end platform for deploying production ML pipelines
The current version of ML Metadata by the time this article is being published is v0.22 (tfx is also v0.22). The API is mature enough to allow for mainstream usage and deployment on the public cloud. Tensorflow Extended uses this extensively for component — component communication, lineage tracking, and other tasks.
We are going to run a very simple pipeline that is just going to generate statistics and the…
The fully end to end example that tensorflow extended provides by running tfx template copy taxi $target-dir
produces 17 files scattered in 5 directories. If you are looking for a smaller, simpler and self contained example that actually runs on the cloud and not locally, this is what you are looking for. Cloud services setup is also mentioned here.
We are going to generate statistics and a schema for the Chicago taxi trips csv dataset that you can find by running the tfx template copy taxi
command under the data
directory.
Generated artifacts such as data statistics or the schema…
Hi there. I’m Theodoros, a Computer Engineering Student here in Greece and I love deep learning.
Welcome to the Understanding Machine Learning in Production. In this article we are going to go over what the main objective of this series is all about and a rough outline of what is going to be covered.
I’m creating these articles because I feel that although the tensorflow ecosystem and high level APIs like keras along with all these free (and non free) tools and services that big companies provide online, like the famous google colab, lower entry barriers to machine learning, the…