So I’ve been on a bender with this Python thing not having proper functional piping. I just can’t beleive it was not built in. Every language should have this. It cleans up your code so much its unbeleive-able. Fear not however, I have built one for Python :D. Now I just need to figure out how to create pip packages. Anyways, lets just do a quick walk through on what it is, how it works etc.
So here is a pretty raw blog article; not unlike most of my articles. The cognitive revolution. I’m going to coin this term today. What the heck is this thing? What does it mean for you? What does it mean for me? Where did it come from? What is it? These are questions I aim to answer in this blog article.
So this blog post is to get you operational with Docker, image and volume management with a pivot towards scientific computing and tensor flow. So I am working on building a Jupyter Notebook for the local mahcine learning meetup to learn the ins and outs of Tensor Flow and deploy this thing up to Azure. Part of getting this to work is not only managing the Docker Containers, but also the data on the volumes so when we deploy up to Azure and somebody opens up the notebook it comes pre-loaded with all the necessary tutorial data.
This is a short post. Basically I had a data set come in, where there were some funky characters involved. I was getting “Can’t read this; doesn’t appear to be UTF-8”. Looked around on stackoverflow for a while to little avail. I came up with this, which works.
So this is an interesting problem. You are collecting data from somewhere and you want to feed it into a neural network for classification. There is one main problem with this. The shape of the data! Neural networks and really just anything require specifically shaped data, you can’t just like give it something of ambiguous size. There are tons of papers out there on dimensionality reduction, but nothing on dimensionality reduction to a specified size. This article explains my approach.
This article is loosely based on a time series challenge from customer data. I have fabricated 3 data files such that they represent the same challenge and we will go through the process of discovering that data. The primary challenge in this data set is that it is from a sleep study and the researchers left the date portion of the time stamp off. What this means is that at midnight, the data plots at the beginning of the x-axis. The second challenge is lining up data to see if there is anything interesting with the time. So yes, you can simply plot using the index that python generates, however I’m also interested in the actual time itself as this is a study involving humans.
So today, I was asked to put some thought into what we should focus our entry level data scientists on in terms of tech skills. After I put a bunch of thought into it, I ended up coming up with this. I decided that the most important aspect of this was a few items fold
Don’t overload them
Can deliver to production where the target can be anything, including IoT.
They will not be concerned with building front ends.
This article is a high level discussion on where you might use various Microsoft Technologies in the field of robotics. I will begin with a side pet project I’m kicking off to get more familiar with some cool tools and tech I’ve lately discovered so I can hopefully get assigned to some really cool projects at work, including drones.
I’m writing this article because believe it or not, this process is a pain in the neck and not completely documented in any one place. Lets start with why in the world you would want to do this. For me, I want to use Tensor Flow and NVidia embedded robotics SDKs. Unfortunately the only supported dev environment for this is Ubuntu. Not anything against Ubuntu it just appears to be fairly unstable in comparison to Mac and Windows, but that is neither here nor there, if you want to build intelligent robots, you need these tools.
This article is meant to explain how the K-Means Clustering algorithm works while simultaneously learning a little Python.
What is K-Means?
K-Means Clustering is an unsupervised learning algorithm that tells you how similar observations are by putting them into groups or “clusters”. K-Means is often used as a discovery step on new data to discover what various categories might be and then apply something such as a k-nearest-neighbor as a classifier to it after understanding the centroid labels. Where a centroid is the center of a “cluster” or group.