Machine Learning for Product Managers
(gentle music) - Thanks for the opportunity to speak to you today about machine learning for product managers. The idea of the talk came because Matt and I have very different perspectives on machine learning. Matt comes from a very technical background, whereas my background is a long time in product. And what we found is that machine learning discussion that's out there comes from those two perspectives, but there's not much in the middle.
So there's very deep technical talks about machine learning, and then there's often very high-level strategic talks about how machine learning is going to change everything. What we wanted to do was a talk that looks at the overlap about what can we actually do today with machine learning with our current products. So what we're going to cover is first of all start with a definition and then we're going to take you through the classic product management lenses and look at machine learning from that perspective. So let's get started and I'll hand over to Matt to kick us off.
Okay, so, you know, what is machine learning? This is a conversation I have with a lot of different people because there's so many hype-related terms out there. There's artificial intelligence, there's data science. People google things and they get these crazy big formulas. But you know, really for me artificial intelligence is a very broad sort of philosophical sort of idea almost. Data science is like a collection of roles in a company, it feels like these days, but really this talk is about machine learning because there's, you know a subset of machine learning and that is really, really applicable to business. And I see machine learning as like a toolbox. We have a collection of problems that we can know how to solve, given certain datasets.
And so when we talk to people about a particular problem they might have, we sort of reach into this toolbox to try to find something that helps.
So we want to go through a bit of an example to start things off.
And so this is a bit of grounding.
I guess, you know, let's say we wanted to build a piece of machinery that predicts the price of a house.
So there's sort of a couple of different parts here we have. You know, we've got this robot in the middle that's somewhat representative of the software we want to write.
I'm going to call that a model.
And what that model does is tries to map between things that we might know about a house, like these sort of things: distance to the city, maybe what the house is made of, how old it is, and try to predict from those things what a sale price might be, say if it went to auction next week.
So the question becomes like, how do we do that? How do we build a bit of software that does this sort of thing? Well, in a lot of machine learning use cases, we use a technique that's called supervised machine learning.
It's called supervise because we sort of train this model. We supervise it by giving an examples of things that we sort of know the answer for.
So what we sort of do is we collect some sort of a priori info about a whole bunch of houses that we might know the price for.
And then the problem becomes not to directly learn or to specify this sort of mapping between these things we know and the price, but to more give examples to a piece of machinery that might then be able to learn for itself. So the really important part about this is that it's all conditioned on the idea that we have access to this data.
And so there's this interesting relationship that'll be a thread that we talk about during this talk, which is different data that we have access to results in a different model.
And that's a really important critical thing, because for a lot of software we write it's very deterministic, sort of have a really good control of things, but this dependency machine learning of a particular dataset can really be critical to success.
And also how sort of things work in general. So what we're going to talk about first is the idea of technical feasibility.
Let's explore some of those things I mentioned, the idea of datasets and models, sort of complexity. This slide we've got sort of these two axes. This is a common sort of way that I think about machine learning, this idea about what sort of data we have and how much we have, and how what I call complex the model is, like what sort of capacity that model has. You can think of it like how big is the piece of machinery is brain.
And there's this really interesting relationship which I think is quite intuitive that in general is a big rule of thumb that says the more data we have available on a problem, the better result we're going to get from this piece of machine learning.
So that's really critical because it means that we don't sort of ever build one piece of machine learning. It's not like we instal the machine learning model and it's sort of done.
It's very much an iterative process where as we collect more data or we make the data more complicated we have the opportunity to sort of rebuild this model in some way to sort of ask from the model more and more complicated questions.
So there's a couple of different dimensions that play in here.
This idea that more data means a better result in some way. What does more data mean? Well, if I think about this table of data here, this idea about things that represent a house that I'm trying to predict from, the simplest one, in terms of more data, it's just more rows, simply more examples of houses when we know the prices.
That's a really clear way of getting more data, and that feels intuitive because the more sort of examples you see of something the better the result might be in terms of how will we generalise to new things. Another interesting dimension of this data is you can think about the comps, like what sort of things do we know about each house? In this example, we've got things like how far it is from the city or what it's made of, but you could also imagine that there's a bunch of stuff which helps us predict the price, like how many bedrooms there are, or does this house have a pool.
So there's also this idea about what sort of information we have on each house. And that's on the data side, I think what's more interesting in relationship to the product is the idea of the complexity of the model. And that's often, not always, but often driven by the complexity of the answer we're trying to get. So in this case, we're talking about what's the predicted sale of a house next week or something. But it might be that the data we've collected or the type of question we're asking can't be necessarily answered from this dataset. So we have a lot of flexibility on this right-hand side, this output side to ask different questions of the data. So maybe this data is really good for saying how many days is it going to take to sell this house that went onto the market, or maybe even a more granular question, or less granular, like will this house sell in the next month? So this complexity of the question we have is also important because the less complex the question we have, the less data we have to have to build a model. So, when we're thinking about products and machine learning, it's important to think about, can we if we need to reframe our question? Is there a simpler version of the question we're trying to ask the machine learning that gives us a better result? And so it's a thing that we have to trade off a lot. As we get more data either by purely number of rows or more complicated data, we can start to ask more complicated questions of the machinery. And I think another thing that's critical that people don't realise is that it's not just one machine learning model in a company. It's not like we instal one piece of software that does this.
We end up with lots and lots and lots of these models that sort of interact in complicated ways.
And it's never just one input and one output. We might have some input which goes to a model, which becomes output, which becomes input for another model. And all of these things might be doing different things with different responsibilities.
And so it's quite complicated, but it's never the case that it's just one single model sitting off to the side. So, if we think about this idea of different types of models and different data requirements that happen, there's quite a continuum of where does this model come from? You know, a while ago, it used to be that you were sort of on your own and you'd have to build this thing from scratch, but more and more as common questions asked of common data, this becomes more and more a commodity in a sense. And nowadays there's cloud services that we can use that very easily represent things that 10 years ago were very difficult to build. So you've got a choice to go with.
Do I just use an existing cloud service, which is maybe trivial to use, and it has some very, very low cost, but doesn't give you any differentiation across the market? The other end of this is if you have very unique data in some way, or a very unique type of question of that data, you might have to build something a little more bold bespoke.
And, you know, there's a big continuum between. You don't have to necessarily just use the cloud service. You might use something that they provide, and then sort of do some tuning on top of it to make it a little bit more specific to your dataset. And these things are very rapid and they change very quickly.
One of the best examples I know of this is this idea of object detection in an image. So here's an image.
Nowadays with computer vision we can draw these bounding boxes around salient objects in the image in real time. Now 10 years ago, people didn't even know how to do this. It was inconceivable that we could have this technology. Five years ago people were having to build complicated pipelines to do this.
These days, it's a small fraction of a cent. You send this image up to the cloud and you get something back.
So everyone has access to this now in some ways. So you've got this interesting balance between what is really quite a new technology and how do you integrate it, but how do you integrate it in a way that maybe is a little bit different to your competitors? The next we're going to go to customer value. - Okay, thanks Matt.
So I'm now going to look at the next product management lens, which is customer value.
So we're going to go back to the matrix that Matt talked us through earlier, and add another element, which is the bubble, which represents customer value.
So the first thing I'm going to say is pretty obvious, and it's just basic good product management, is that we should always be thinking about the value we deliver to customers.
But I think often because of the excitement we have around the technology for machine learning it's easy to lose sight of this.
And so what we're looking for is, as we tick tack up that matrix and both increase the size and complexity of the data estate and develop the model capacity we're looking for an increase in customer value that matches that investment.
And really importantly, we need to set a baseline at the beginning.
So before we even establish a machine learning experience we know what it would look like if it was a non-machine learning one.
So if it's a recommender, what happens if we just have a simple rules-based system that just says products in the same category at around the same price point? And that gives us the baseline for measuring that customer value.
And when we do that, we find that often there's a kind of sweet spot in the middle going from bottom left to top right.
And the most valuable thing we can do is to stay in that sweet spot and increase value as we add to the data estate, improve the model complexity.
We'll know when we're moving out of that because you're not increasing the customer value. So if that's on the horizontal axis, it might be that it's redundant data and you see really kind of extreme versions of this where organisations just hoard a lot of data that doesn't actually add value.
On the vertical axis it might look like over-investment in the model where you're doing something that's exciting from a machine learning perspective, but maybe isn't delivering the customer value. So with that in mind, we'll look at three anti-patterns that we see in trying to develop that customer value in line with the investment in the technology. So the first of them is a fairly obvious one, which we've called the stalled value trap.
And that's really just where you're no longer getting a growth in value. So you continuing to invest, you're moving across and upwards, but there's no increase in value.
At this point, we should stop and solve another problem, because there's so many opportunities that can be addressed with machine learning. The next one is the hype trap.
And I think this is common where the organisation or the founder or the CEO just really wants to see some machine learning implemented.
And what this can look like is that the non-machine learning solution is actually a pretty substantial value bubble. And the investment might eventually get us to it and maybe even slightly increase it, but it hasn't justified the cost that's required to get us there.
And then the last one, the moonshot, is perhaps not really an anti-pattern in that it's a valid choice, but it's important for us to consider whether it's right for the organisation. And what that looks like is that there's no payoff until there's been really substantial investment, lots of data and a real complexity in the model. So this might be things like autonomous driving and despite all the investment that the industry has put into that, we're still not there, but clearly a lot of great companies believe we're going to get there.
And they're prepared to take that moonshot. What we would counsel is that we should understand whether it's right for our organisation before embarking on it, because you don't want to stop until you've got to that top right corner.
So what we do advocate is a portfolio of approaches. And so, like you would have a a product portfolio for a set of products, it makes sense to have different investment approaches.
So that might be an incremental one, which if we go back to the automated vehicles example, maybe this is developing lane changing technology. And then meanwhile, another part of the portfolio, we're investing in a more moonshot approach, which is actually developing level one autonomy. Okay, so hopefully that's given a view of how customer value can be considered with machine learning.
The next thing we look at is commercial viability. So I'm gonna let's stick with the two axes that Matt introduced.
And the first one I'll look at is the horizontal axis, so the data estate.
And there's two the big costs that we want to explore, and I should drop a caveat here, which is mileage may vary. It's very hard for us to really talk with great confidence about cost, but we do see these two patterns.
So one is around the hardware costs and data storage. And this isn't significant in the early stages of developing the data estate.
So it only grows as we get towards the end of that axis when you're getting real high volume data, and particularly data that's been drawn off the actual use of the product.
The bigger cost is the people cost and in data engineering. So it's often the case that the data of the organisation isn't in a great state and you actually have to put a fairly big upfront investment in order to get it into a position where you can do anything.
So that's the horizontal axis.
The vertical axis, in terms of developing the model capacity.
Again, we can look at it in those two pieces, the people and the hardware costs.
This is the probably one good news on the cost slide is it may be free to get to that very first step.
So as Matt said, machine learning as a service may be an important cost in order to get that basic experience.
You're then likely to incur some people costs in order to to implement a more generic model, but then you will see it really blow out into that custom space, so where you're actually trying to develop something that's original.
It's high risk and likely high cost.
On the hardware side, the compute costs grow as you move up that axis in a fairly linear kind of way. So, that's commercial viability.
Let's move on to the next classic product lens, which is usability.
So usability in is obviously really going to depend on what the application you're building is. But there's a pattern that I think is worth noting that we think applies to most machine learning based products.
And that's the idea of having an interface that's robust to failure.
So, there's always the challenge of developing the machine learning experience up those two axes.
And it's important that the UI you create reflects where you are.
So the example that we've got here is from the YouTube recommender, and what it does is it doesn't presume to completely solve the user's problem of what's the next video I should watch.
It kind of augments that experience.
So it says on the right, here's half a dozen videos that might be relevant.
And it mixes in different alternatives there. So that if the user is interested in just straight football, or broader sport, or humour, or a particular player, they've got options and they can choose from that augmented experience.
As we get more confident with the machine learning experience we can be more presumptuous in terms of the interface we build.
And so the auto-play features is example of that, where we see that as up on the top right of the matrix where we really automating, and rather than just augmenting that navigation, we're actually saying we think this is the one video that you should watch next.
So those are the four classic lenses that people like Maddy Cagan would talk about. The other two things we want to cover: what's the strategic context, and then what's the delivery context.
So I'll do the strategy piece.
One of the things that makes machine learning very exciting from a strategic point of view is this double flywheel effect that we see. So a lot of products will have the flywheel that you can see in the centre, which is that they create a good experience, this drives usage, which then generates revenue, which they can then invest in improving that experience and that attracts more usage.
So that kind of revenue based flywheel is common to a lot of products.
What's interesting about machine learning is that it then has this second flywheel, which comes from the unique data that is generated from that usage.
So by using that product, it creates unique data that only this product has access to, which can then be fed into the model innovation, which further develops the experience and hopefully generates more usage.
And the wheel continues to turn.
So that's why from a strategic point of view, getting there first with a machine learning product can create this unassailable competitive advantage. And with that in mind, it's interesting to look at what we think of the two main categories of machine learning opportunities.
So the first one is about value creation.
And the definition of this is where you're actually building a new product experience using the machine learning technology.
So examples might be voice or self-driving vehicles. And this is interesting because there's this real potential for the flywheel.
And given that potential, there's value in getting there first.
So there is some urgency, but there's also risky, and there's that moonshot risk that we talked about earlier. The alternative place to look for those value opportunities are around efficiency. So that's reducing costs and improving efficiency. That might be things like fraud detection.
It's more incremental.
You're more confident that you'll succeed, but it's less likely to have that flywheel effect that creates the urgency, but also the really big payoff. So the final piece we'd like to talk about is delivery. And I'll hand back to Matt for that.
- So there's a couple of different patterns and ways of working that we've seen in companies that are either small or large, and it sort of follows two axes.
The horizontal axes here, it talks about the complexity of the machine learning that's being used. And I really like this series of terms that comes from quite old statistics, this idea of descriptive, prescriptive, or predictive, different versions of machine learning.
So descriptive statistics is when we're building things like dashboards.
So if I have a large amount of data and I want to sort of summarise it in some way that I can aid humans in their decision-making. So, you know, big dashboards showing graphs, the distillation of complicated data into small data, that's the descriptive statistics.
And people will often move from that to this idea of predictive.
If I've got a whole bunch of data, then, it's really handy for a business to be able to predict the future.
So this is where a lot of the first dipping of the toes in machine learning comes in. Can I predict some of the things that are in these descriptive statistics, because if I know about them in the future, it's really handy.
And often as you get better and better at this, why do you want to predict the future? It's because you want to sort of change some leavers upstream in your system.
You want to be able to change things that are happening so that you can say that my predictive model says we'd make more money if I made more, say, widgets now than tomorrow, I should make widgets now.
And so the third round of this which is a little bit rarer is this idea of prescriptive, where we're actually handing over some aspect of the business to this machinery to act on our behalf. They're the three ways that we see the complexity growing. There's also this idea about who in the company is doing it. You know, it's very common that because this is quite technically challenging, there's maybe a single team, maybe some sort of centre of excellence that's sort of starting with this.
And eventually they might grow into some sort of cross team capability, or there might be some sort of mandate from a very high tech company to do this company-wide. And, you know, there's three arrows we've drawn on here that are sort of interesting ways of doing this. And two of them are in a way, failure modes. This arrow that's going straight up, this idea of descriptive statistics that are across the company.
This is very common.
We've got a lot of teams that are doing simple things, so to speak, and maybe failing is a bit harsh there. It's basically the first step that we see.
One of the other areas we've got here is this one that's completely horizontal.
That's this idea that there's a single team that are doing very complicated things, have deep expertise in an area, but they're the only people doing it.
And this is a big mode of failure.
Having a few people that are able to do amazing things, but no one else is really not a long-term way to do this. So the ideal state is this diagonal, this idea that maybe you grow single teams doing simple things, but you really want this capability across the world.
And one of the best ways of doing this is this idea of capturing that expertise, not necessarily always in people, but in tools and process. So it's a very important point to make.
And that is the end of our talk.
Machine Learning for Product Managers
Matt Travers & Matt Kelcey - Product Principal and Principal, Machine Learning, ThoughtWorks
What we'll cover today: A look at ML through six product management lenses. Seven pink text boxes with the following captions:
- What is Machine Learning
- Technical Feasibility
- Customer Value
- Commercial Viability
Slide titled:What is machine learning. Sub bullets read: Artificial intelligence, data science, machine learning. Beside the text, a graphic of complicated mathematical equations.
Slide titled: Breaking down the jargon: A house price example. Graphic of a spreadsheet with columns headed distance to central business district, material, age, and sale price with associated data in the cells below. Between the first three columns and the sale price column appears a graphic of a robot with a red arrow linking the distance, material, and age data to the house price data. One cell under the sale price column does not have a dollar amount but instead a row of question marks.
Replay of presentation overview slide with the text box: Technical Feasibility highlighted.
Slide titled: How models and data interact: Getting the balance right. Subtext reading: Start small and add more data and model complexity in increments. Beside the text, graphic of a line graph. X-Axis labelled: Bespoke data estate. Y-Axis labelled: Machine learning Model Capacity. In the data section, a growth arrow reading: investment.
Replay of Slide headed: Two dimensions of data:volume vs complexity which featured the graphic of the robot between the data sets related to house distance, material, age, and sale price.
Slide titled: Never just a single model. Illustration of a diagram featuring snippets of rows and columns of varying sizes interspersed with icons of robots. A number of arrows are interspersed throughout. Some arrows indicate a relationship between elements while others point to directions beyond the diagram.
Slide titled: Model options: Commodity versus bespoke models. Under the title, an image of a cloud icon captioned: Cloud services and existing models provide a commodity set of capability but don't differentiate. An image of a wrench icon captioned: Bespoke models give the most flexibility, but come and non-trivial engineering cost.
Replay of the prior slide with an image added demonstrating object recognition software. The image shows a crowded traffic lane with bikes, cars, traffic lights, people and trucks, over which the software has overlaid labelled and colour-coded boundary boxes.
Replay of presentation overview slide with the text box: Customer Value highlighted
Slide titled: Mapping value to investment: Getting the balance right. Sub bullets read: Customer value should increase to justify the investment in both data and models (just like all products). A clear measure of customer value and a non-Machine Learning baseline is essential. Beside this text, the same graph from an earlier slide with axes reading: Bespoke data estate and Machine Learning Model Capacity. Overlaid on the growth line graph in the earlier example is a scatter plot diagram charting customer value and how it increases relative to increases in data and models.
Replay of prior slide with another graphic overlaid onto graph consisting of a bubble higher up on the Machine learning model complexity axis labelled: Over-engineered model and a bubble farther along the Bespoke data estate axis labelled: Redundant data. Lines of best fit drawn from each bubble are outside the core zone charting positive customer value.
Slide titled: The stalled value trap: enough already. Subtext reading: Customer value plateaus as better models and more data provide no perceivable improvement. Beside, a replay of the The Machine learning model capacity to bespoke data estate graph, this time demonstrating customer value at the lower intersections of the axes and no growth in value at the higher and longer intersections.
Slide titled:The hype trap:We just need some Machine Learning! Subtext reading: The opportunity for Machine Learning is overestimated. Simple rules-based systems already deliver high value and a major Machine Learning investment is required to match the customer value. Replay of the model capacity to bespoke data graph, this one showing customer value as the anchor at baseline with small gains as complexity and volume increase, and a slightly larger gain much farther up and across the complexity and volume axes.
Slide titled: The moonshot: True faith. Subtext reading: Value relies on complex models and high volume and complexity of data. High risk, with no feedback on progressing customer value. Replay of the model capacity to bespoke data graph plotting customer value as steady throughout increases in complexity and volume and then growing exponentially at a point beyond the reach of the axes.
Slide titled: Complementary Approaches: The most reliable path. Subtext reading: Most successful Machine Learning products combine a portfolio of models and datasets, often at different stages of development, which maximize customer value over the life of the product. Replay of the model capacity to bespoke data graph. This time the graph charts steady customer value as volume and complexity increase. The prior moonset chart of little customer value increase until exponential growth at a point beyond the graph is inlaid as a faded graphic to demonstrate the difference between steady incremental growth and exponential growth models.
Slide titled: Data Cost Structure: How to manage costs. Subtext reads: Expect high people costs in data engineering when first accessing data. Storage is cheap but grows as complexity and volume increase. Zoom of the bespoke data estate axis with an icon of a person with large stacks of money beside them and an icon of s monitor with steadily large stacks of money beside it as you move across the axis.
Slide titled: Machine Learning cost structure: people and technology. Subtext reading: Compute costs can be significant but typically have gradual growth. Data scientists are expensive and can cause explosive increases during the transition from generic models to custom models. Zoom of the Machine learning model capacity axis with icons of a person captioned: data and a monitor captioned: compute. Three markers are placed ascending the axis next to icons of piles of money demonstrating comparative costs. The first marker is labelled: Machine learning as a service and shows a free cost correlation with data and a low cost correlation with compute. The second marker is labelled: Generic and shows low cost correlation with data and higher with compute. The final marker is labelled: Custom and shows high costs across both data and compute functions.
Replay of presentation overview slide with the text box: Usability highlighted.
Slide titled: Robust to failure: combine algorithms for flexibility and nuance. Under the title is a screenshot of a Youtube video of a football match with a list of recommended or 'up next' videos alongside. Next to each recommended video is a keyword associated with the video currently being watched. Some keywords are: Zidane (who is one of the players), humour, football, sport.
Slide titled: Augmenting versus Automating: Start with a humble User Interface. Under the title is a screenshot of a Youtube video of a football match. Beside this is a new iteration of the machine learning capacity and bespoke data estate graph. A straight diagonal arrow between the axis has the phrase: humble user interface and the word augmenting near the base, and the phrase: confident user interface and the word automating near the apex as volume and complexity increase. A marker near the intersection of confident user interface and automating is linked back to the youtube video.
Replay of presentation overview slide with the text box: Strategy highlighted.
Slide titled: Machine Learning's double flywheel: how usage can create unassailable competitive advantage. Underneath the title, a circular shaped flow chart. The word usage is at the apex. An arrow points from usage to a dollar symbol, which in turn points to model innovation, which in turn points to experience quality. A second arrow points from usage to the phrase: unique data or volume and complexity, which in turn points to model innovation.
Slide headed: Value creation vs efficiency: Where to look for opportunities. Under the title, two columns. One titled: Value creation. The other titled Efficiency. There are four rows in each column comparing the two across the following metrics: Definition; examples; competitive advantage; timing; and likelihood of success. Value creation is defined as: creates new customer value which can be monetized. Examples are voice services, self driving vehicles. The competitive advantage of value creation is the potential for flywheel. The Timing is to explore urgently. The likelihood of success is risky or moonshot. Alternatively, efficiency is defined as reducing or mitigating costs through efficiency. Examples of this are process prioritization, fraud detection. The competitive advantage of efficiency is incremental and replicable. The timing is to build gradually. The likelihood of success is more proven.
Replay of presentation overview slide with the text box: Delivery highlighted.
Slide titled: Machine Learning Maturity: Where are the skills? Subtext: There's usually a mix of skills and impacts within an organization. Beside text, a graph. Three different points are delineated on the Y axis starting in ascending order and are labelled: Single team, Cross team and company-wide. Three points on the x axis marked in linear order are labelled: Descriptive, predictive, and prescriptive. The graph plots three lines: Line one: many teams doing simple things, which reaches its apex at the intersection between company wide and descriptive. Line two: Deep expertise in few teams, which reaches its apex at the intersection between single team and prescriptive. Line three: Capture expertise in tooling, which reaches its apex at the intersection between the other two, between company wide and prescriptive.