Lenny's Podcast: Product | Growth | Career: How to measure and improve developer productivity | Nicole Forsgren (Microsoft Research, GitHub, Google)
Lenny Rachitsky 7/30/23 - Episode Page - 1h 17m - PDF Transcript
Starting with what is your problem or what is your goal, I would say this is a bigger challenge
than most people recognize or realize. 80% of the folks that I work with, this is their biggest
problem, even at executive levels. Teams will have gone off for several months and they're
tackling something and they'll come back with uncertainty and they'll say like,
will you told me to improve developer experience? I'm like, okay, what do you mean by this? Are
you talking about inner and outer loop? Are you talking about friction? Are you talking about
culture? But if you're talking about culture, this is totally different than if you're talking
about friction in tool chains. If you're on different pages, you're heading in completely
different directions. Welcome to Lenny's podcast where I interview world-class product leaders
and growth experts to learn from their hard-won experiences building and growing today's most
successful products. Today, my guest is Nicole Forsgren. This is actually my first recording
back since going on Pat Lee for the past couple months and what an awesome episode to get back
into the swing of things. Nicole is the developer productivity expert having written
the award-winning book Accelerate and she's been the co-author of the State of DevOps report
year after year. She's currently a partner at Microsoft Research leading developer productivity
research and strategy and she's helped some of the biggest companies in the world move faster,
improve product quality and transform their cultures. In her conversation, we get into the
weeds of how to go about measuring and improving your engineering team's productivity and experience.
We talk about the Dora framework and the space framework and how to actually implement them
to understand how your engineering team is doing. Nicole also shares benchmarks for what
elite companies are at. We talk about why moving faster turns out to be one of the best ways to
improve quality and stability plus pitfalls you want to avoid and also preview of a new book
that she's working on and so much more. Enjoy this episode with Nicole Forsgren after a short word
from our sponsors. Today's entire episode is brought to you by DX, a platform for measuring
and improving developer productivity. DX is designed by the researchers behind frameworks such as Dora,
Space and DevX including Nicole Forsgren who is my guest for this very episode. If you've tried
measuring developer productivity, you know that there are a lot of basic metrics out there and a
lot of ways to do this wrong and getting that full view of productivity is still really hard.
DX tackles this problem by combining qualitative and quantitative insights based on the very research
Nicole and her team have done giving you full clarity into how your developers are doing.
DX is used by both startups and Fortune 500 companies including companies like Twilio,
Amplitude, eBay, Brex, Toast, Pfizer and Procter & Gamble. To learn more about DX and get a demo
of their product, visit their website at getdx.com slash Lenny. That's getdx.com slash Lenny.
Nicole, welcome to the podcast. Thank you so much. I'm excited to be here. I'm excited to have you
here. I actually skip this question usually with guests, but I thought it'd be actually really
valuable to spend a little time on your background. You have such a unique role and unique set of
experiences. Could you just talk briefly about the things you've been up to in your career,
where you've worked and then what you're up to now and what you focus on these days?
Sure. I appreciate the question because you're right. I sort of had this choose your own
adventure background. I started as a software engineer at IBM. I was writing software for
large enterprise systems, which meant I ended up running them. I was also a sysadmin. I was
rocking and stacking. I was running these really, really large labs. Then I stumbled into this
seven-day march for several years and I was like, there has to be a better way. We're here in rumors
of it, but management was not buying in. I decided to win this battle with data. I was like, I should
go do a PhD. I ended up taking a slight pivot into PhD and management information systems,
which some people are less familiar with, but it's basically a cross between tech
and business. I ended up getting a fairly technical PhD. I went to a school. I went to
University of Arizona, which has a very, very technical degree, but I liked that it crossed
with business because then I had the ability to make these strong business case statements.
How is or is the way that we develop and deliver software tied to outcomes at the
individual level? Can I be more productive? Can I have better work-life balance? The team level
is the team more productive, is the team more efficient, and the organizational level. This
is what I was really interested in originally. Do I see better ROI? Do I see better efficiency?
Because then I could sell it to people. That was really what I originally went into.
I was a professor for a handful of years because if you're doing research, traditionally, that's
the job in academia. I also had a master's in accounting because that really helped me make
that kind of like financial tie and understanding financial statements. Then after a handful of
years, I walked away from tenure because academia was not convinced that DevOps was a thing.
The DevOps wasn't real. Stay at a DevOps report, who I was doing with Dora, DevOps Research and
Assessment in collaboration with Jez Humble and Gene Kim. We started that work with Puppet. Shout
out to Alana Brown for starting that and Nigel Kirsten and the team there. We kind of pivoted
away and Chef, the little configuration management startup at the time, hired me and they're like,
we'll give you half time to do research and half time to help our engineering practices improve.
That's cool. Yeah. I mean, they were incredible because what startup is going to be like, yeah,
do research. I was there for a year and a half and then left to do Dora full time. We actually
had a SaaS offering. We continued this data DevOps report just under the door of the inner.
We had a SaaS offering because so many large companies were like, I want my own customized
measurement reading and report. Then the joke there, when we met with Gartner, they were like,
your superpower here was that you tricked people into strategy, which was not only,
how do I benchmark? That was kind of our top of the funnel because everyone wants to know how
they compare. The important thing is what should you do next? What's the most important
next step? How do I measure? What do I do next? That gave me this incredible view into
advising large organizations into this transformation journey. Then we built out this
amazing partner network because we weren't actually consulting. We just had the SaaS piece,
but then how do you act on it? We were then acquired by Google. I was CEO and co-founder,
so I led that acquisition and then integration and building up these teams in Google. After that
point, I joined GitHub, which is the largest developer network. I had this amazing opportunity
to do more grounded and applied research again. I was VP of Research and Strategy.
Then I went over to MSR, where I kind of wear a couple hats. Right now, I have a research lab
there with an incredible team. It's the Developer Experience Lab, where we do a bunch of work across
productivity, community, and well-being. Then I also help with Microsoft's cross
company effort to improve their developer infrastructure. It's this round effort into
how do I really remain engaged in measuring, applying, thinking about this work, both in
very applied, concrete pieces, and incredibly forward-looking work with MSR.
Amazing. Just to clarify, MSR, is that Microsoft?
Yeah, thank you. MSR is Microsoft Research.
Okay, cool. You've shared a couple of these terms, DevOps, developer productivity. I'm
curious what the term you like to use for this area you focus on, developer productivity,
developer experience, DevOps, what's the best way to think about this?
I really love that you asked this question, because I think they're very related concepts
that people sometimes conflate, but I see them as being different, so related but different.
Productivity, I think, is basically how much we can get done and how much we can do over time.
I think that's why it's so important to have this holistic measure, because we can't just
brute force it. That's why when my team and I and a bunch of my peers study productivity,
we include this community effect, because software is a team sport, we joke, and also
why well-being is so important, because we see that when you do productivity the right way,
we see sustainability, we see well-being, we see reductions in burnout.
Developer experience is very related and very tied to this, and it contributes to productivity,
but developer experience is if you think about who your users are, developers really are your
users in this software engineering, in the software development piece, and so it's developer
experience is what is it like to write software? Is this a friction-free process? Is this a very
predictable and certain experience? Can we reduce this uncertainty and increase the predictability
here to contribute to productivity? Then how does DevOps fit into that just so that we have the
mental model of these terms? People have co-opted the terms, and some people named their tools,
DevOps. I'm maybe a little more old-school, so when I was doing a bunch of my DevOps research,
it was the capabilities and tools and processes that we can use to improve our software development
and delivery end-to-end, so that it's faster and it is more reliable. DevOps was this technical,
architectural, cultural practices that enable us to do this work better, so that it is, yes,
more productive. We have a better developer experience, so it was again this very holistic
picture. What I love about this topic is that I've never met a founder or a leader who is not
thinking, we need to move faster, we need our engineers to be more productive, we need to get
things out, the door quicker. We want engineers to be happier. Nobody doesn't want that,
and so that's why I'm excited to dig into a lot of these things. Is that roughly what you find as
well, that nobody's ever like, we're good, we don't need any of this, we don't need to focus on this
area? You know what? I'll say yes and, right? It kind of goes back to why I got into this, because
on the one hand, you won't say anyone who's saying, we don't really need to go faster,
everything's fine, but at the same time, very often, I will come into scenarios where I'll find
myself in scenarios where people are like, I mean, it would be nice if we were going faster,
but do we really need to? Show me the business case. What's the ROI? Or if we go too fast,
we'll have an instability, right? What are our safety measures? Are we going to lose reliability?
What is happening, right? And when I first started, like ITIL and ITSM, the old school
kind of change management processes, the common knowledge was that you had to have at least a
two-week wait for change approvals in order to get that stability. Turns out that's not, right?
It was just kind of an old lifestyle, right? And so we kind of have this weird balance of,
I want to move faster, but is it worth the investment? What am I going to get for it?
Are you sure this is the priority? Or I've been in meetings where it's like, oh, yes,
absolutely, right? Like, this is a priority, but it's the lowest priority. And I'm like,
right? So then what we want to do is we want to have these kind of pointed conversations
or these kind of like socratic type questions and conversations where it's like, help me understand
more what your concerns are. Are your concerns around reliability when you move faster? We're
not just trying to like all the guardrails down and sprint for no purpose of sprinting. And this
is where kind of the Dora and DevOps research program comes into play, where it's, we don't
just want to move fast and take all guardrails down. We want to implement good technical practices,
like automated testing, good architectural practices so that when you move fast, you are also
more stable, right? We want to be thinking about improving the developer experience so that when
we are faster, we are also saving time, right? And then we can highlight a handful of statistics like
what is your typical time for feature delivery? What is your typical time to first PR? What is
your typical time to steady state productivity? What is your typical time for code review and PR
process? And if we are to do like back of the napkin math, what sorts of time are you spending here?
And if we do a rough look at industry, what are your peers spending here? And are we losing
time, right? And if we can turn this into a value calculation, what does that look like so that we
can think about the priority and the strategy here? And I think that's where it becomes
a more focused conversation. This is a great segue to something I was going to get to a little bit
later, but let's just get into it, which is the Dora framework. And then there's also the space
framework. Can you just talk about what these two are when you use one versus the other, and then
how that essentially helps you measure and then improve productivity and developer experience?
Sure, sure, absolutely. And I'm so glad you brought this up. So Dora is an entire research
program. Now, many people, when they hear Dora now, they think of the four keys or the Dora four
or the four metrics. And I think that's what the research program and the company ended up
becoming most known for. And so that was the software delivery performance metrics. And
those are, there's two speed and two stability metrics. So the speed metrics are lead time.
So how long to take to get from code committed to code running and production, deployment frequency,
how often do you deploy code? And then the stability metrics are MTTR, meantime to restore.
So if something happens, how long does it take you to come back and then change fail rate for every
change that is pushed? What's the rough percentage of incidents that require human intervention?
Now, the thing that was really interesting is when we started measuring these,
we found that they move in tandem now with very strong significance from a statistical standpoint.
Now, what this means is now we say speed and stability move together. Most people only think
about this from the speed standpoint, which means when you move faster, you are more stable,
which means you're, you're pushing smaller changes more often, right? Because if you're
pushing all the time, it's going to be very, very small changes, which means you have a smaller
blast radius, which means when you push, you have an error in production. It's going to be easier
to debug, right? It's going to be much easier to figure all of that out. Your meantime to restore
and mitigate, it's going to be much faster. But that also means is the reverse, when you push
changes less frequently, you will have more unstable systems because when you push less
frequency, you will have very, very large batch changes, which means you will have a very high,
very large blast radius, which means when you do have a resulting bug error, you will have to
disentangle this big ball of mud, right? And figure out which piece actually caused the error.
Figure all of that out. That ended up being a big surprise, right? Because refer to my prior
comment about, you know, ITIL and ITSM, if you're forcing a two week pause for change approvals,
you're causing this batching up of changes. And sometimes people were waiting with two weeks is
good, a month must be better, or three months must be better, or six months must be better.
And I mean, just think about the merge conflicts you're causing, right? You're just causing so
many challenges and figuring out how to push this code into production.
So many people think of those four metrics, one, because we found that speed and stability
move together, and two, because we started publishing benchmarks on what this looks like for
low, medium, high, and then elite performers for many times. This, I believe, may have been
interesting. I'm not sure if it was useful or helpful, but I think it was interesting,
because it gave people at least something to shoot for, something to aim for. I will
definitely say, what's most important is knowing where you are and the progress that you're making,
right? It doesn't matter if, frankly, you're a high performer or you're an elite performer,
it matters that you know where you are and you're making progress, right? You know,
can you push daily or on demand? Or is your only technical capability that you can push twice a
year, right? Just know where you are, and is it a business decision or a technical capability?
That's basically what it comes down to. I'm going to jump in real quick, just to highlight what you
just said, which I think is extremely important and powerful, and people might kind of move on too
quickly. I also want to ask you what the actual benchmarks are, if you can share those, whatever
you want to share there, but before I ask that, essentially what you're sharing right now is just,
I feel like the $64,000 question of this episode is just, how do I move faster
as a team? And what I'm hearing is, essentially, it's ship smaller things is kind of at the core of
it. And also, if quality is low, you're also saying the answer is ship more often, ship smaller
things. Is that roughly the message? Yes, absolutely. It ends up being much, much safer.
Amazing. So I think that's an extremely important takeaway that I think people would, I don't know,
that's surprising to me to hear that it's quality comes from shipping faster, and then also to
ship faster and move, help your team move faster, it's ship smaller things and just deploy more
often. Yeah. Amazing. Okay, great. I know we'll talk more about this, but let me go back to the
question I was going to ask is, are there benchmarks you can share just right now that you think would
be useful to people? I know you said it was interesting and maybe not as useful to people as
you mentioned. Yeah. So I will admit, I only have the 2019 benchmarks top of mind. The team
of Google has continued that work since I left. It's been led by Dr. Dustin Smith. Nathan Harvey
continues the work. So huge shout out to that team. Many others participate. You can go to
dora.dev and find all of the continued reports. They've integrated all of this work, but I will
say they've remained fairly consistent. So really quickly, I'll share the elite performance. So
deployment frequency, you can deploy on demand, lead time for changes takes less than a day,
time to restore is less than an hour, and your change failure rate is between zero and 15%.
Amazing. Okay, I'm writing these down. These are extremely valuable.
And I will mention people will say, well, this is kind of like a chunk of time, right? It's not
super precise. Precision isn't really super important here, right? Like I don't, it doesn't
really matter if you can, like if your lead time is, if it's less than a day, it's less than a day,
right? Like that's fine from a business perspective. It doesn't matter if it's like
four hours or like four hours and two minutes, right?
Right. General categories are fine. Now, I will say like the next category for lead time for
changes by the day is if lead time is between a day and a week. And this is for, for good?
For high, yeah, between elite and high elite is less than a day and high is between a day and a
week. And then it goes between a week and a month and between a month and six months, right? So
so you can ask people and they can tell you, right? They can kind of hunch it.
And this is from committing code into the repo and it going out into production?
To it, to about like ring zero. So you don't like, don't worry if it's like, oh, well, now we need
to think about like the global deploy and like, which is the final endpoint? It's like, how long
does it take to get through your deployment pipeline? Because are you going to be surprised?
Do you have fast feedback loops? How does your deployment pipeline work? Does your
deployment pipeline work? Right? Or are you going to commit code? Are you going to wait for that
final review for about three months? Is something going to happen or break? And when it comes back
to the developer, because something happened or broke, because that kind of happens, right?
Are they going to have to insert themselves back in the code,
re-review all the things that happened three months ago? All so many other things happened.
That's incredibly difficult, which to your prior question, this is how it relates to the developer
experience. If something happened less than a day, and like it's a surprise and it's not great,
but like whatever, right? Something happened downstream and I got to fix it. I'm still sitting
in my code right in my head. I've got that mental model. I know what happened. Maybe it's not great,
but it's fine. If it happened three months ago and I get interrupted, first of all, interruptions
like suck. That's not fun. Second of all, now I've got to like, re-remember, re-read all of this
code, maybe reload an entire new workspace instead of libraries and everything. Because I,
maybe it's a whole quarter ago and like, we thought we were done.
And I got to do the whole thing all over. If a listener is working at a startup,
I imagine they're hearing this and they're like, it takes a day to ship that, we ship it all day,
a thousand times a day. I imagine these benchmarks are more valuable for larger companies. Is there
a kind of buckets you think about for like, here's the size of company this is meant for,
and then do you think about anything differently for a startup, say, I don't know, 10 people?
If anyone is only listening to this, I just got the biggest smile because we saw no statistical
significance between small companies and large companies. The only statistically significant
difference was with retail. I'll come back to that. It's so funny because large companies would say,
oh, but this isn't fair for us. We have more complex code bases. We have so many things to do.
Small companies just don't have to deal with this. Small companies would come to me and they would
say, oh, but this isn't fair. Large companies have so much money. They have so many resources.
They don't have to deal with all the things. This doesn't apply to me. So it's like, either way,
and on days when I was feeling real snarky, I'd be like, pick your excuse.
You've got your like, drop down. Now, when I say retail was a bit of an outlier,
they had a statistically significant difference. Their difference was that they were actually
better. Why? Now, I can't tell you why. I can, in a research paper, we'd have a discussion section,
and this is where you get to guess. I would surmise, and we do this in the report,
that it's probably because retail went through the retail apocalypse. If you didn't survive,
like if you weren't just killing it, you did not survive. So many retail firms just did not
make it through. You had to be at the top of your game. Black Friday, there's no such thing as not
having systems that are performing incredibly well. There's no such thing as not being in the
cloud, because if you cannot make it through bursting on demand, bursting like magic, sometimes I
joke, right, you're not going to make it. And so I suspect, if I were to guess, if you're not already
a high performer in the retail space, natural selection got rid of you.
That is really interesting. That makes a lot of sense. So I'm looking at these thresholds again,
and I'm thinking from the perspective of a founder who's just like, I wish my engineering team moved
faster. Essentially, you're saying if deploy times, if they deploy more than once a day,
if their deploy frequency is on demand, or I think it was hourly, was kind of the other bucket,
was that part of it? Yeah. And then their mean time, their fail rate is like less than 10%,
and their mean time to recovery is less than an hour. Basically, you're doing great. That's kind
of the message of this framework at least. And if you're not doing it through brute force and
killing yourself, now, can I jump in here? Because then people are like, but how do I do this? So
let's say that you're not in that category, and you're like, because this is the next,
this is the piece of criticism I'll get about Dora. People are like,
well, all you've done is make me feel bad. You gave me these metrics. You've judged me.
Now I feel bad. And then I'm like, so there's Dora there. I wrote a book called Accelerate,
which is like the first four years of the research compiled and put together and expanded
in a few things. And I joke, there's a whole rest of the book. Dora is best known for the four
metrics, but there's an entire research program supporting it. So it's not just these four metrics.
What we find is that if you improve a set of capabilities, I loved your question around,
what is DevOps? DevOps is not the tool chain you buy. Marketing teams label tool chains DevOps.
They wanted your money. DevOps is a set of capabilities. They're technical capabilities.
They're architectural capabilities. They're cultural capabilities. They are
lean management practices that predict speed and stability. And then speed and stability gives
you money, right? Because it's your ability to create these features that give you money.
So when you work backwards, if you want money, you get the features fast. If you want the features
fast and stable, you do the things. And the things are technical capabilities like automated
testing, CICD. And CICD is continuous integration, continuous deployment, is that right? Yes.
Trunk-based development, using a version control system. So do you have good technical practices?
Do you have good architectural practices? Do you have a loosely coupled system?
Are you using the cloud? Or if you're not in the cloud for whatever reason,
are you using the underlying architectural pieces that enable good cloud to do the cloud right?
Or if you're in the cloud and you're not realizing benefits because you're doing the
cloud wrong, right? Do you have a good culture? So you don't just magically go fast and have
stability, right? So working backwards, which pieces are you struggling at? Now,
you kind of noted down the benchmarks. If you go to dora.dev, the team at Google was
lovely. We worked really closely with the team. They're keeping this updated. You can take a quick
check. There's a button there that says quick check. You can plug in where you kind of think you are.
Like I said, you can hunch it. And it'll tell you where you are in the benchmarks today.
And what industry you're in. And then the cool part is it'll say, now, like you'll want to ask
yourself, like, where am I struggling? But it'll say, for your performance profile,
and for the industry that you're in, statistically over the last several years,
these are probably your constraints. Okay. These are probably the things that you're
struggling in right now, right? Like, for people in finance who are high performers,
they tend to struggle with these four things, right? Whether it's like culture or continuous
integration or whatever. I love that you're getting tactical with how to actually improve
these already, which is the bread and butter of this podcast. And so we'll link to this quick
check. It's just dora.dev slash quick check. And by the way, they do not collect your name.
They do not collect your info. There is no, there's no lead, any lead gen anything. Everything's just
there. And then there's deep dives into every single one of the capabilities.
Amazing. And also your book talks about all these things. So people should go check out the book,
obviously. It's on Amazon, Search Accelerate. Is that right? Yeah. Okay. So we were talking about
dora. This may be a good time to talk about space, which I think is a different framework you
recommend. What is that all about? Okay. So space is a way to measure, we say productivity,
developer productivity, but it's a little bit more than that. Space is a good way to measure any type
of complex creative work. Now, how do they relate? Let's say you go through the quick check.
It points out like four things and you decide you want to improve. Continues integration
and culture, right? Well, now you're like, cool, but how am I going to actually measure them?
This is where space comes in. Because space helps you figure out, space gives you a framework
to pick the right metrics. Now, some people are like, well, space, you didn't give me the
exact metrics. People love dora because it's like, here's the exact four you need. Well,
space is like when you want to measure something that's complex, creative work, maybe like developer
productivity, there's also an example at the bottom for incident management. When you have
something you want to measure, it says within your context, within the metrics you have available
to you, here's how to pick. That's what space is good for. Now, we called it space because
it stands for the five dimensions that you want to measure. So S is satisfaction and well-being.
So satisfaction well-being is kind of self-explanatory. Now, some people might jump in here and say,
oh, well, you're just, you know, touchy feeling. This actually matters because we find that
satisfaction well-being ends up being incredibly highly correlated with all of the other
dimensions of productivity and doing things well. And as soon as satisfaction and well-being,
things like sustainability, if you're satisfied, as soon as that starts falling off, other things
start to break. So this can be an incredibly strong and important signal. P is performance.
This is going to be the outcome of a process. So reliability within dora, the MTTM or change
fail rate, right? Those are both performance metrics. And so you pick one to kind of measure
as performance. Yep. A is activity. Anytime you have a count or a number of something,
these we see all the time because they're super easy to instrument and automate, right? Number of
pull requests, number of check-ins, number of something, that's A. C is communication collaboration.
This can be how people work and talk together. It can be meetings. It can be collaboration.
It can also be how our systems communicate together. It can be the searchability of a
code base. And then E is efficiency and flow. So this is going to be the flow through the system.
It can be the time through the system. If we think about SRE or incident management,
it can be the number of hops a ticket takes until it reaches the right person.
Now, to use space correctly, we want to use at least three dimensions at a time
because that helps us balance. Turns out Dora is actually an implementation of space.
So Dora would be space for mostly that outer loop. So again, once you found something that you want
to improve, find the metrics that make sense to you. Try to have them be in balance or intention
so you don't throw something at a whack. But pick three. So when you say Dora is an
implementation of space, one has five buckets, one has four. How do you actually think about that?
So space is there to help you think about how you want to pick metrics. So a lot of time I see
people, so we have to step back. I used to advise people on how to pick metrics. For years,
people would pull me in to advise on Dora or accelerate. They would ask me questions,
but it ended up being metrics questions a lot. How do I pick the right metrics
to improve what I'm doing? Like I said, they had the Dora numbers.
They would pick their constraints and they wanted to improve. But how do I improve?
How do I measure this? How do I show improvement? And so we would start thinking really critically
about which metrics were the right metrics to pick. And I would always say, make sure you pick
balanced metrics. Make sure you pick metrics that are intentioned.
And I could say it, but people have a hard time wrapping that around their heads because they
kept picking things like, number of lines of code, never picked number of lines of code,
number of, still every month I get an email like this, number of pull requests, number of commits.
And I was like, these are all activity metrics. And so finally I pulled a few of my friends
together and I was like, let's come up with a framework to help people think about it.
And so there are five broad categories. Pick three, because that will help force you through
the mental exercise of what could I possibly pick? You don't need all five, right? This isn't,
we're not playing bingo. We're not playing blackout bingo. You don't need all of them.
But try to have at least three across different dimensions. Now, one example here,
I was working with a group that wanted to improve their pull requests very generally.
They just said improve pull requests. So they were thinking about pinging someone every 15 minutes.
And I was like, oh, this is going to be bad. Because we know from other literature and research,
like nursing, you'll get alert fatigue where people will just start tuning out alerts.
Either they'll turn them off or they will just stop hearing them. So like number of alerts,
they're like, let's just think about number of alerts. And I said, well, but if we think about
efficiency and flow, how much time do you have to work on your coding? So those two are balanced.
So we need to protect time to work, as well as code review time, pull request time.
And so sometimes we can think about those. And then I think we added a satisfaction metric.
Are you satisfied with the pull request process and the selection of the reviewer?
How do you go about actually capturing and measuring the say satisfaction?
So for satisfaction, I would generally ask. Go ahead and ask. Now, the ones that you instrument,
you can instrument and pull out of systems all the time. Go ahead and grab that string.
For a satisfaction metric, I would only pull that periodically once every few months.
Like a survey to your engineering team.
Like a survey. Yep, absolutely. Awesome.
Don't discount what people say. Sometimes I hear, actually not sometimes,
a lot of times I hear people say, oh, but people lie. First of all, what is their incentive to lie?
Why would they lie about having a bad system? Because it's bad and they want it fixed.
Right? If it's absolutely a hostile work environment,
they might lie and then tell you it's good, then you have bigger problems.
Right? Also, do we ever see bad data from our systems or incomplete data from our systems?
That's a lie. But we find ways to deal with it. We see it. We acknowledge it.
We look for holes in our system data. We try to deal with it. Right? That's also a lie.
So I think there are better ways to think about and deal with that
and then try to work with it. Right? Because then I wrote a paper with Mick Kirsten on this
several years ago on how data from people and data from systems are really important
compliments because we can get certain insights from people that we'll never get from systems.
Right? Like let's look at lead time from changes, for example. Right? From commit to deploy.
The speed might be fine, but people might tell you it's taking absolute heroics.
Right? It's some ridiculous Rube Goldberg machine. The system will never tell you that.
Or you could get data on your version control system. I worked with a company several years ago
and we found out that there was a significant portion of code that was just
not going into any version control system. You're never going to find that out from your systems
because it's not in the systems and it was mission critical.
I can see why people come to you asking for advice on metrics because you have this framework of
here's the type of metrics you want and then I think and especially from an engineering team
there's going to be this like how do I optimize and make sure I'm doing the right thing and
measuring the right things. For someone that wants to do this and an hour-long podcast isn't
going to give them all the answers, would you recommend they go read or go do or look at to
help them figure that out? One, I hate to be this person, but I'll point to a few of my papers because
I will say I write things down because I get asked them so often and I want to make sure it
is broadly applicable or broadly available I guess. This space paper for sure. It's an ACM
and I think the year we published it it was like the most read paper at ACMQ. We tried to make it as
readable as possible so the space paper is nice because it outlines this framework and it gives
examples of metrics in every single category and so hopefully people can look at this and they can
say okay here's an example to use here. Here are some of the things that I could possibly use
and we're seeing that space is being used lots and lots of different places. Another good one
could be the paper that I mentioned with Mick Kirsten and it was about we talked about using
data from people and data from systems. We wrote it up in the DevOps context because I want to say
this is written in like 2016 or 2017 or something, but it helps you think through what types of data
are good in which situations, right? Because you will never find yourself in a situation when you
don't want both types of data. Even teams that I've worked with that are the most advanced,
they have absolute instrumentation in every possible scenario.
In the most detailed way, they will still survey their developers at least once a year
because you can get new insights, right? One book that I love, it's a little dense but it's
really interesting that I love is How to Measure Anything and it's by Hubbard and there are parts
of it that are like real stats heavy, but he has this portion in the front that's like
covering intangibles and so it's like what happens when you don't have data? You have no data,
you're starting from nothing, what are good ways to hunch data? I really love that because he covers
some really good ground there. Today's entire episode is brought to you by DX, a platform for
measuring and improving developer productivity. DX is designed by the researchers behind frameworks
such as Dora, Space and DevX, including Nicole Forsgren, who is my guest for this very episode.
If you've tried measuring developer productivity, you know that there are a lot of basic metrics
out there and a lot of ways to do this wrong and getting that full view of productivity is still
really hard. DX tackles this problem by combining qualitative and quantitative insights based on
the very research Nicole and her team have done, giving you full clarity into how your developers
are doing. DX is used by both startups and Fortune 500 companies, including companies like Twilio,
Amplitude, eBay, Brex, Toast, Pfizer and Procter & Gamble. To learn more about DX and get a demo
of their product, visit their website at getdx.com. That's getdx.com. You also mentioned offline that
you might be working on a book that will answer a lot of these questions. Is that something you're
up for chatting about? Yeah, absolutely. So as I mentioned, I tend to write things down when I get
asked questions on it a lot. And so this is one in particular. So we'll be covering, you know,
I'm starting to go through and I'm covering some of these. And I think some of the important
topics in particular are, you know, starting with what is your problem or what is your goal
and being super, super crisp on it, right? Like, what is it that we're trying to answer? And
I would say this is a bigger challenge than most people recognize or realize. Like,
I'm making this setup, right? Like 80% of the folks that I work with, this, this is their biggest
problem. Even at like executive levels, they'll, they'll ask their team or teams will come back
with uncertainty and they'll say like, will you told me to improve developer experience?
And like, okay, great. What do you mean by that? And then teams will have gone off for several months
and they're tackling something and they'll come back and they'll be like, oh, that wasn't what I
meant. And I'm like, okay, what do you mean by this? Are you talking about
inner and outer loop? Are you talking about friction? Are you talking about culture?
Because sometimes they're talking about culture. And if you're talking about culture,
this is an incredibly valid answer. But if you're talking about culture, this is totally different
than if you're talking about friction in tool chains, right? And if you're on different pages,
you're heading in completely different directions. So like that, that's one thing
we cover, which seems obvious, but trust me, it is not. And then even like, how do you,
we're going to do kind of a rough version of like, how do you start measuring from nothing?
And also the measurement journey, right? Like, how do you think about the trade-offs between
and the proportion of measurement between subjective data, right? Data from people. So you
interviews and you have surveys and objective data, stuff you get from systems. Because when
you first start off, you'll be relying much more on data from people because you can get it relatively
quickly. But as you kind of transition through this measurement journey, you'll get more and
more data from your systems because it's scalable, it can be engineered, you can be doing, you know,
much more with it. And also you should be thinking about, you know, don't let the perfect be the
enemy, the good, right? So like, how do we think about this very, very strategically? How do we
transition through this? How do we think about what each piece of data is for? And also like,
lots and lots of examples, right? So I have included like, example interview scripts. How do
you select people? How do you screen people? Example survey scripts? What are some of the
analyses we should do? And trying to make this incredibly accessible. So like, basically anyone
can do this. So you do not need to be a data scientist. But if you have one on staff,
like you can hand them some of this and just like, let them run.
I think this book is going to do extremely well. I definitely come back on when it is out. I think
you said maybe year-ish kind of timeframe? Yeah, probably about a year by the time we get all the
way through. If people want to be notified when it's out, can they sign up on your site for a
newsletter or anything like that? Or is there any way to kind of be in the loop as it approaches?
Oh, yeah, absolutely. Yeah, I'll add a link for that. Also, if anyone is doing some of this work
now, if they have major questions that they would love to have me to answer, if they have success
stories, if they have case studies, if they have anything that they would love to be included,
I remember when I wrote Accelerate Before, there were a couple of folks that reached out after
and they would like, oh, I wanted to have something included. Now, today I've learned, right? If there's
anything that folks would love to be in discussion with me about, I'm always eager to chat and nerd
out about DevX and especially measurement journeys. Awesome. I usually ask this at the end and I have
more questions, but while we're here, how would people reach out to you? What's the best way to
contact me? On my website, I've got info.nicolefi at Gmail. Awesome. Okay, a few more questions.
Awesome. Thanks. What are the most common pitfalls that companies run into when they're trying to
roll out any sort of developer experience, developer productivity, system measurements,
improvements? I think one, I just mentioned, right? Like not being clear or not understanding
what it is that they're looking for, because then you can have a thousand flowers bloom and
everyone's kind of running in a different direction. I think another one is not pursuing
this in both a top down and a bottom up structure. And I think that can really help
drive success and having good communication throughout is super, super important, right?
So getting your ICs bought in and helping them understand that this is for them,
we want to understand what they're doing, knowing what vocabulary they use, what terminology they
use is super important, and then chatting with leaders, right? And understanding what their
motivations are or helping them understand what the motivations could be. This kind of
harkens back to one of our earliest chats on why I even got into this and how I see two different
sides to the conversation on like, why is DevOps even a thing? Why should we even ship faster?
There are so many people that I talk to that are super passionate about DevX right now,
and they're like, how can I convince my executive team this is important because
their developers are just completely burning out or they use computers in anger every day.
And so it's like, how can we have the right tools to socialize this to our leaders as well,
right? Because this should be a priority. This needs to be a strategic piece. And how can we help
pull together the right value points to communicate this and to understand what their
priorities are so that we can see how this fits in, right? You've been working in this space for
a long time, probably longer than anyone that has ever worked on this area of developer experience
productivity. What have you seen change most from the time you started working in the space to
today? What kind of progress has been made? We have these increasingly large complex systems,
right? So like 10 or 15 years ago, like the internet was around, but like,
things were really different. Now, almost every company has a really large complex system, right?
We also have a shortage of developers, or at least a reported perceived shortage of developers.
More companies are technology driven, or at least they understand they're technology driven.
It's like, I remember a handful of years ago when I met with a financial institution
whose CTO insisted to me that he was not a tech company. That's not real anymore. That
doesn't happen anymore, at least like very, very rarely. So all of these things come together.
And suddenly, many more companies are like, we have to be better at this. And that was not
always the case, like five to 10 years ago. I used to have to really explain why this was
oppressing concern and why it would continue to be a pressing concern. And now, in the last
six to nine to 12 months, we have this AI moment happening. And it just poured gas on top of
everything. Because now what's important, like we've always said that like ideas aren't important,
execution is important. But now, this is absolutely true. Because it's not just about
what it is that you build. It's about creating absolutely novel, incredibly new experiences,
and doing them at a speed that no one has seen before. And the only way to do this
is to have this software pipeline that is fast, and is safe, and is stable, and is reliable.
And that's where we're seeing this really interesting convergence and pressure isn't
quite the right word, but it's really forcing the discussion, and strategy and prioritization,
right? I'm glad you touched on AI. That was actually exactly where I was going to go next.
Perfect. Yeah, obviously, productivity, AI, engineering, something that's top of mind for a
lot of people. There's a lot of layoffs that have been happening. There's a lot of talk of,
we don't need as many engineers. I actually had dinner not too long ago with a few, I'd say,
10x engineers. And those are folks that people sometimes say like, they don't need Copilot,
they're not going to use any of these tools, they're already amazing. And they were the opposite,
they're like, this is making me 100% more effective and efficient than I love it.
So clearly, good things are happening there. I don't know what the question is specifically,
but I guess, have you seen the impact of AI on engineering productivity? And has that shifted
how you think about developer experience and productivity beyond what you already just shared?
Absolutely. Yes, and, right, I think this is a super interesting open question. So can I answer
it just with a whole bunch of questions? Absolutely. We're absolutely seeing an impact,
and we continue to explore this. So I have an interesting question to see how it'll change
the space framework. What's open here? I think a few things will remain. Satisfaction is still
going to be there, performance is still going to be there, activity is still going to be there,
how you communicate with people and with the tool, efficiency and flow is still going to be there.
I believe it will change an added dimension like trust or reliability. Can I rely on it? Well,
I have an over-reliance on it, and what we're seeing is that, probably unsurprisingly,
people really fundamentally shift the way they
work when they work with an AI-enabled tool, like GitHub Copilot or Tab 9 or others,
because now instead of just writing code or having a short autocomplete,
you spend more time reviewing code than writing code. There's this wonderful paper
out that uses the CUPS model. I'll share it with you. A team at MSR did it. It finds that
about 50% of your time now is spent reviewing versus writing.
But it'll be interesting to see how that changes things longitudinally,
because some of my colleagues also did a paper that showed that
you can do certain tasks like build an HTTP server 50% faster. But I don't think that's what
productivity is about when you're using an AI tool, frankly. Anyone who's looking at that,
and dear CEOs or whoever who are like, now I can lay off at my workforce,
that's not what this is about. It's not about taking a task and cutting your time in half,
because now what we've enabled is your ability to do certain things faster,
and then free up some of your cognitive space so that you can do harder things with this new
co-pilot sidecar or something. But also, because now you're accepting text and then reviewing it,
we've changed what your mental model is. So we've changed the friction model that you expect,
we've changed the cognitive load of what you expect. We're changing reliance on code. So what
does this mean for reliance or over reliance? What does this mean for learning? What does this
mean for novices versus experts? How do we measure productivity? There are a handful of us
that are having these discussions on what does this mean, and how do we communicate it thoughtfully?
Again, we really need to have these holistic balanced metrics, because if it's an oversimplification,
we really risk losing the forest for the trees. But it's also super interesting and super compelling,
how can we think about learning or onboarding to new code bases or new languages for folks who
already know computational learning? I think it's also very different for folks who are just learning
programming languages and don't already know things like computational thinking.
If someone was excited to go down this road of we're going to focus on developer experience,
we're going to focus on helping your engineers be more productive, what are the next
step or two that they should take in your opinion just broadly knowing that you don't know any
specifics about the company that's thinking about this right now? I think if you're walking away
from this podcast and you're like, I'm already working on this or I think this is a thing that's
happening, I would say just go check your work basically. Has this been written down? Is there
a clearly defined challenge, problem, something? Start there. Absolutely. Because that is going
to be the thing that reduces confusion, the best. Absolutely. And then see if there's any data.
And data can be very loosely defined. Is there any signal that is related to the problem?
Like I'd start there. And you can do that. You can do that a week. You can hunt something down.
Sounds like something you could do in a day. Yeah. Well, depending on how scattered things are.
Are there any companies that you look at as good models of they do this really well?
I think Google does this incredibly well. And sometimes I hesitate to mention Google because
they're like, you know, some people are like, well, we can't be Google and we aren't really
advanced. But the thing I love about Google's approach is that they've really taken kind of
this measurement phase approach to things, right, even when they roll it out in new places.
They're very systematic in how they measure things. They have incredible telemetry and
tooling and instrumentation. And they continue to invest time in developer experience surveys,
and they triangulate them. And one thing that I also love being able to point out here is if there
is ever a disagreement between the surveys and the instrumentation, which is incredibly advanced
almost every time, every time that I've ever heard of the surveys are correct and not the
instrumentation. Amazing. I have just a couple more questions unrelated to this topic. Is there
anything else that you thought you think would be useful to share or leave people with around
this general space? I would say that like thinking about what it is you want to do
is always important, right? Like getting crisp, the ability to communicate clearly is always one
of the best things. One of, I think, one of my superpowers and one of the things that I've been
working with my teams on doing and kind of teaching them is, and one of the things that's
really leveled up our work in general is making your work incredibly accessible. And accessible,
not necessarily like the accessibility definition of the word, but making it very easy to understand
what you're doing for your key audiences. And so thinking about doing that for anything that,
you know, anyone who's listening for all of your work is super important, right? So who is it that
your audience is? What's their role? What words resonate with them? And then always being able
to translate your work into a few sentences or a paragraph or less. I love it. A lot of the listeners
of this podcast are product managers. And this is so core to the work of a PM. So I think this is
perfect. Speaking really directly to a lot of the listeners. Okay, so just a couple more questions.
Before this podcast, I asked you a few questions, including just like, what are people asking you
for advice often around? And are there any other frameworks that you find really useful? And so
there's a couple things I just wanted to touch on, see if there's something interesting there.
The first is you have this framework that you call the four box framework. I'm curious what that is
and what it's all about. Yes, I love this four box framework. I've used it for years. I actually
pulled it out first when I was a professor. And I still, to this day, get LinkedIn messages from
my students saying that it's like the most useful thing they've ever used. So here's what it is.
I literally pulled this out on napkins at bars at conferences to this day. So here we go. Draw
four boxes on a piece of paper, two on the top, two on the bottom. So they'll be kind of aligned.
The first two, to the left of them, write the word words.
And below them, write the word data. And then between the two on the top, draw an arrow between
them. So it'll say words, box, arrow, box, right? Does that make any sense? Then on the bottom,
it'll say data, box, arrow, box. Okay. So on the top half, this is where if you want to think about
measuring something or testing something, you have to start with words. So as an example,
let's just say, I think that customer satisfaction gets us more money.
Or customer satisfaction gets us return customers. Let's do customer satisfaction.
So the first box, you'll put customer satisfaction inside the box, and you'll put return customers in
the second box. Now, always start with words. Do not start with data. You always start with words.
And then you'll go around to a couple people, stakeholders, managers, others, and you'll say,
do you agree with this? Is this actually what we're doing? It can turn into a sentence.
And then in the boxes below it, this is your data. How are we going to measure customer satisfaction?
It could be a survey. And so like this is where you'll go and you'll say, like, what data points do we
have that could proxy for what could be our data points for customer satisfaction? And this is
where it gets tricky because you could say, well, customer satisfaction could be return customers,
but we think it leads to return customers. So we can't use that here. But return customers
or could be so like, that's where you kind of like roll this out. So how else would we measure
customer satisfaction? I made this hard on myself. Like a CSAT score or CSAT?
Yeah, CSAT, NPS, we could say the amount of money that they spent. It's a stretch.
Okay, now return customers. Let's go to the next box. How are we going to measure return customers?
Depending on our context, let's say that this is an online business, we could say that it's return
customers as measured through the website. We could say that it's return customers, we could just
ask them, right? Maybe we have a follow-up survey, return customers, maybe we're going to do a
stretch here, maybe we say it's a referral link. This helps us get super clear on what it is we're
going to measure. Now, the reason I like this is because if some of our data, now this data analysis,
we'll just do correlations here, right? If we have longitudinal over time, that's fine.
You can hand this to like a data scientist. You can hand this to someone and you can say,
what data do we have? Let's go run this. If something here falls apart, now you can point to the
data boxes and we can get mad about the things in the data boxes and we can say, what's wrong?
Is the data poor quality? Are we missing data? Was this a bad proxy, right? Proxy stands
for something else. Was this ridiculous, right? One of the things I made up, right? It was just
a bad idea. Instead of getting mad at Lenny for his really stupid idea or getting mad at Nicole
because this was a really bad idea, we can say, this was problematic. What's wrong here? Or we can
go back up to the words at the top and we can say, this is not actually something that is probably
going to hold or this is not something we want to test right now or this is something. It makes
things incredibly clear. It helps you communicate what it is you want to do fairly quickly.
I love it. Here's mine. Check it out. It's ugly. I'll zoom in right now. I will say advanced mode.
You can start with the same four box framework and you can say, what data do we have available?
What do we think the relationships are? But then you have to go back up to words and then say,
for these data points and we think that they represent something and we think this is the
relationship between them, what do they represent? If I turn this into a sentence, what do they
represent? Then you want to double check because Spurrius Correlation is one of my favorite websites
instead of charts. You'll want to go chat with someone, interview, make sure things are actually
right. But the challenge is I will see people run every correlation they could think of,
but they haven't turned it into a word or a sentence that you can communicate to someone else.
They don't do the check and they don't do that before, one, before running the correlations.
Two, if it's there, all of our data is so interrelated that we quite often will find Spurrius
Correlations. But it can be really helpful just to have that laid out, even if it's just not a
post it, to say, what are the things I expect to see? What is this actually testing? What
relationship do I suspect is there? Amazing. I have a newsletter guest post on how to do a
correlation analysis and regression analysis so folks can read that. Awesome. Oh, that's so great.
Plug and play, all kinds of it makes it easy for you. So what I'm going to take you away from this
is this is an awesome framework, especially for thinking about a hypothesis you may have.
In this case, it's like customer satisfaction is going to lead to more return customers. Here's
how we're going to measure it. And then you basically run the desk and see if it's true. And
if it's not, maybe you need to pick different metrics, maybe you need to pick a different
conclusion. And within the Dora framework, we would say if we want to improve our speed and
stability, we think improving build time would help. And then how would I measure build time,
right? These are the data points that I have available to us. Yep, to circle back.
I love it. It's all connected. Okay. And then last question. I asked you what advice people often
ask you for. And you said that it's around making decisions. And I'm curious, what advice do you
give people about making decisions? Yes. So this one comes up in business, but also comes up
personally. And among my mentees. So many times it starts with, you know, being very crisp about
your objectives and definitions. But then it comes down to, you know, really clearly defining
what your criteria is, what's important. And then among that criteria, what's most important.
Some of my friends know I have a decision-making spreadsheet that I have shared out with a handful
of friends on, you know, should you take a job? Where should you move? What are the different
things really useful? It is, it's well, it's funny though, because what's interesting is many times
I will like, I'll share it with someone and I've got a couple that are just funny, right?
But walking through the spreadsheet is often all you need to do in order to know what the
decision is. And by that, I mean, so we walked through the decision, I had one where I was like,
where should I move next? Or like, what job should I take? Right? So when I started Dora,
I did this. Starting Dora, I thought was my lowest. Once I walked through this spreadsheet,
it became my high. So what you do is you outline, like, of all of your options, what do you want
to do? And then you say, what are the criteria that are important to me? So if it's for a job,
is it something like total comp, cash money, prestige, team, job predictability,
work-life balance, identify the criteria that are most important to you. Now, it's really
interesting because sometimes I will only get that far when I'm working with someone I'm mentoring
or coaching, and they will say, I know what my answer is. We don't even get to the next step.
But just identifying the criteria that are important, is it? Now, when I was thinking about
where I wanted to move next, it was proximity to an airport, the relative tech scene, the food
scene, that was real high for me, a handful of things, that was important. Now, the next thing
I do is for each criteria, what's their relative weight? What's their importance? And I make it
add up 200%. And then I, like, this is the easy part, right? Like, you just put it in a little
spreadsheet and I then I give everything to score and I just multiply it out. Now, this is where
I'm data informed and I'm not data driven. There have been times I make a decision where, you know,
the whole, like, flip a coin and, like, whatever it's, wherever it lands on, what your reaction
is tells you what it should actually be. There have been times where, like, I multiply it out
and then I'll actually, like, fudge the numbers to get what I want, but it's still slightly off.
That's where your data informed. Same thing in business. There are many times where
you actually run the numbers and it'll give you a class or a category of things and then you choose.
Now, this is where, you know, one of my favorite quotes I heard somewhere about strategy comes
into play, right? And that's that the key to having a good strategy is knowing what not to do
and the key to executing a good strategy is actually not doing it.
So you can have many options, right? As a leader and as an executive, we have many
options and we only fund some of them. If you fund everything, things are going to fail.
So being able to think through and identify what your criteria are, identifying that criteria,
what's your selection criteria, what's your evaluative criteria, ranking them and then
deciding what the cutoff is, is important. You can't fund everything. You don't get to pick
everything. Amazing. I love the spreadsheet idea. I've made versions of it, but it's always,
I think like you said a lot of times, the exercise is just tell you what you already think
and just gives you like, yeah, all right, you're right. You probably should just do that thing
you already thought you should do. Yep. Have you thought about making a public template of
this spreadsheet even though it is simple? I bet it would be really helpful to people.
I have and this actually might be a good forcing function, maybe.
Okay. Awesome. So if you do it, I'll put in the show notes, it'll probably be near the bottom
at the end of the episode, but that'd be awesome. Perfect. Is there anything else that you want to
share before we get to our very exciting lightning round? No, I think that's it. Well,
welcome to our very exciting lightning round. I've got six questions for you. Are you ready?
Absolutely. All right. First question, what are two or three books that you've recommended most
to other people? We actually had the perfect segue because the book I've recommended absolutely
the most is called Good Strategy, Bad Strategy by Richard Rumelt. Another one is Designing Your Life
by Bill Burnett and Dave Evans. And the last one is probably Ender's Game. Orson Scott Card.
No comment right now on some of his political commentary, but I used to have extra copies
in my office when I was a professor and I would just hand it out to my students.
It's a fun, like, just easy nonsense read, but... I absolutely love it. Such a good pick.
Haven't read in a long time. And are they making a show of that at all? That'd be something.
They made a movie and I was afraid I wasn't going to like it, so I just didn't read it because I
didn't want it to ruin the book. But at least Harrison Ford was in it. Okay, I'm not going to
check it out. They're making a movie of three body problem. I don't know if you've read that,
but I'm really excited for that. It's on my list. Oh, man. Best sci-fi ever.
Next question, actually. Very correlated. What is a favorite recent movie or TV show?
I think going through some real, just easy fun watches lately. I'm re-watching Suits again,
but Ted Lasso is a favorite and I just tore through Never Have I Ever, which is fun because
John McEnroe narrates it, which is hilarious. John McEnroe, the tennis player?
Yeah, it's a riot. Yeah, it's so funny. I love it. Next question. What's
a favorite interview question that you like to ask people when you're interviewing them?
I love questions that I can kind of spin around hard decisions that people have had to make
and how they made them. I love hearing their thought process.
And I get a little nervous when people just like yolo and shoot from the hip constantly.
So what is it you look for there that gives you a sense that they're
someone you may want to hire or work with? I just like hearing if they have some sort of
process. If they have some kind of decision-making process, if they have criteria, if they have
how do they do evaluation. What is a favorite product that you've recently discovered that you
love? I have a big one, a little one. My big one is probably Sleep 8. So I live in Arizona.
It gets hot here sometimes. Oh, 8 Sleep. Yeah, 8 Sleep. Yeah, the other way around.
Yeah, so that one's fun because it makes the bed cold and also gives me some data,
which is probably like a little bit off, but in the approximation it's fun.
And then Korean face masks, they're just fun. Yeah, you can get some pretty good ones for like
just a couple dollars and that's always fun. Self-care. First mention of that one, Korean face
masks. Right. Listen, everyone get on board. I just did the TikTok. There's a filter now
where you could see how you look when you age and I'm not happy with how it turned out. And so
I might look into this. I had some basal cell cancer on my forehead a few years ago. And so I
am much more careful with my skin. And you can get like one of my favorites is Cosrx. You can get
10 for like $15. So it's fun to just like chill at the end of the day with a good face mask.
I was going to ask you for a specific pick and so we got one. Yep. Amazing. This next question,
I ask everyone and it's especially appropriate to you, but I don't know if you'll have an answer.
What's something relatively minor you've changed in your product development process
that has had a big impact on your team's ability to execute? And I feel like you have
a big perspective on this. So I'm curious what you have as an answer.
I think I alluded to this earlier. I would say that it's, you know, helping everyone. So I've
done this before, but I think it's helping everyone to ask who's our audience and how will
we share this now? And it's sort of interesting because right now I'm wearing two hats. One is
it MSR or Microsoft Research. We need very ambitious research, right? So like H2H3.
What is H3? The third cap? Oh, Horizon 3. And so it's supposed to be like 5 to 10 years out,
which right now is like who even knows, right? We're going to be in computers.
Yeah. AI has like completely upended how we kind of think of horizons. And so
when we're thinking really ambitiously and very, very, very forward looking,
what's our check in? How do we evaluate this? And then how can we easily communicate it to
our core audience? And so here, who's our audience and how do we bring the far near?
And then for the other hat I'm wearing, I'm working with Okto kind of across all of Microsoft
to take a data-informed approach to really improve and up-level our central developer
infrastructure. And so as we're thinking very, very tactically, what is our long-term vision
and how do we align with several of our broad stakeholders? And so there it's who's our audience
and how do we bring the near far? I love that. Final question. What is one tactical piece of
advice that listeners can do this week to help improve their developer productivity
or developer experience and move it in the right direction? You know, if you walk away from this
podcast right now, you could take a look at what's happening in your org today. Is it written down?
Is it clear? Do you have any existing data and efforts? And if not, go find a handful of developers
and ask them how they feel about their work tools and their work process and what the biggest barriers
to their productivity are. Also pick up a copy of Accelerate on Amazon or your local retail
establishment. Nicole, this was amazing. I think we're going to help a lot of companies move faster,
have better and happier engineers, which is going to create infinite value in the world.
Thank you so much for being here. Two final questions. Where can folks find you online if
they want to reach out? And how can listeners be useful to you? I'm on Twitter and Blue Sky,
NicoleFV. And my website is NicoleFV.com. And all my contact information is there.
And as we mentioned previously, I'm working on a new project and a new book digging into exactly
these ideas, right? How can we measure better? How can we improve? And what does that measurement
process look like? Both for one-time, really quick, unofficial measurement pieces and if we
want to do very formal, longer-term measurement pieces. So if anyone is interested in that or
has any success stories they'd love to share, I would love to hear more about it. So please
reach out and share. I'd love to hear more. Awesome. Nicole, thank you again for being here.
Thank you, Lenny. Bye, everyone.
Thank you so much for listening. If you found this valuable, you can subscribe to the show on
Apple Podcasts, Spotify, or your favorite podcast app. Also, please consider giving us a rating or
leaving a review, as that really helps other listeners find the podcast. You can find all
past episodes or learn more about the show at Lenny'sPodcast.com. See you in the next episode.
Machine-generated transcript that may contain inaccuracies.
This episode is brought to you by DX—a platform for measuring and improving developer productivity.
—
Dr. Nicole Forsgren is a developer productivity and DevOps expert who works with engineering organizations to make work better. Best known as co-author of the Shingo Publication Award-winning book Accelerate and the DevOps Handbook, 2nd edition and author of the State of DevOps Reports, she has helped some of the biggest companies in the world transform their culture, processes, tech, and architecture. Nicole is currently a Partner at Microsoft Research, leading developer productivity research and strategy, and a technical founder/CEO with a successful exit to Google. In a previous life, she was a software engineer, sysadmin, hardware performance engineer, and professor. She has published several peer-reviewed journal papers, has been awarded public and private research grants (funders include NASA and the NSF), and has been featured in the Wall Street Journal, Forbes, Computerworld, and InformationWeek. In today’s podcast, we discuss:
• Two frameworks for measuring developer productivity: DORA and SPACE
• Benchmarks for what good and great look like
• Common mistakes to avoid when measuring developer productivity
• Resources and tools for improving your metrics
• Signs your developer experience needs attention
• How to improve your developer experience
• Nicole’s Four-Box framework for thinking about data and relationships
—
Find the full transcript at: https://www.lennyspodcast.com/how-to-measure-and-improve-developer-productivity-nicole-forsgren-microsoft-research-github-goo/#transcript
—
Where to find Nicole Forsgren:
• Twitter: https://twitter.com/nicolefv
• LinkedIn: https://www.linkedin.com/in/nicolefv/
• Website: https://nicolefv.com/
—
Where to find Lenny:
• Newsletter: https://www.lennysnewsletter.com
• Twitter: https://twitter.com/lennysan
• LinkedIn: https://www.linkedin.com/in/lennyrachitsky/
—
In this episode, we cover:
(00:00) Nicole’s background
(07:55) Unpacking the terms “developer productivity,” “developer experience,” and “DevOps”
(10:06) How to move faster and improve practices across the board
(13:43) The DORA framework
(18:54) Benchmarks for success
(22:33) Why company size doesn’t matter
(24:54) How to improve DevOps capabilities by working backward
(29:23) The SPACE framework and choosing metrics
(32:51) How SPACE and DORA work together
(35:39) Measuring satisfaction
(37:52) Resources and tools for optimizing metrics
(41:29) Nicole’s current book project
(45:43) Common pitfalls companies run into when rolling out developer productivity/optimizations
(47:42) How the DevOps space has progressed
(50:07) The impact of AI on the developer experience and productivity
(54:04) First steps to take if you’re trying to improve the developer experience
(55:15) Why Google is an example of a company implementing DevOps solutions well
(56:11) The importance of clear communication
(57:32) Nicole’s Four-Box framework
(1:05:15) Advice on making decisions
(1:08:56) Lightning round
—
Referenced:
• Chef: https://www.chef.io/
• DORA: https://dora.dev/
• GitHub: https://github.com/
• Microsoft Research: https://www.microsoft.com/en-us/research/
• What is DORA?: https://devops.com/what-is-dora-and-why-you-should-care/
• Dustin Smith on LinkedIn: https://www.linkedin.com/in/dustin-smith-b0525458/
• Nathen Harvey on LinkedIn: https://www.linkedin.com/in/nathen/
• What is CI/CD?: https://about.gitlab.com/topics/ci-cd/
• Trunk-based development: https://cloud.google.com/architecture/devops/devops-tech-trunk-based-development
• DORA DevOps Quick Check: https://dora.dev/quickcheck/
• Accelerate: The Science of Lean Software and DevOps: Building and Scaling High Performing Technology Organizations: https://www.amazon.com/Accelerate-Software-Performing-Technology-Organizations/dp/1942788339
• The SPACE of Developer Productivity: https://queue.acm.org/detail.cfm?id=3454124
• DevOps Metrics: Nicole Forsgren and Mik Kersten: https://queue.acm.org/detail.cfm?id=3182626
• How to Measure Anything: Finding the Value of Intangibles in Business: https://www.amazon.com/How-Measure-Anything-Intangibles-Business/dp/1118539273/
• GitHub Copilot: https://github.com/features/copilot
• Tabnine: https://www.tabnine.com/the-leading-ai-assistant-for-software-development
• Nicole’s Decision-Making Spreadsheet: https://docs.google.com/spreadsheets/d/1wItAODkhZ-zKnnFbyDERCd8Hq2NQ03WPvCfigBQ5vpc/edit?usp=sharing
• How to do linear regression and correlation analysis: https://www.lennysnewsletter.com/p/linear-regression-and-correlation-analysis
• Good Strategy/Bad Strategy: The difference and why it matters: https://www.amazon.com/Good-Strategy-Bad-difference-matters/dp/1781256179/
• Designing Your Life: How to Build a Well-Lived, Joyful Life: https://www.amazon.com/Designing-Your-Life-Well-Lived-Joyful/dp/1101875321
• Ender’s Game: https://www.amazon.com/Enders-Game-Ender-Quintet-1/dp/1250773024/ref=tmm_pap_swatch_0
• Suits on Netflix: https://www.netflix.com/title/70195800
• Ted Lasso on AppleTV+: https://tv.apple.com/us/show/ted-lasso
• Never Have I Ever on Netflix: https://www.netflix.com/title/80179190
• Eight Sleep: https://www.eightsleep.com/
• COSRX face masks: https://www.amazon.com/COSRX-Advanced-Secretion-Hydrating-Moisturizing/dp/B08JSL9W6K/
—
Production and marketing by https://penname.co/. For inquiries about sponsoring the podcast, email podcast@lennyrachitsky.com.
—
Lenny may be an investor in the companies discussed.
Get full access to Lenny's Newsletter at www.lennysnewsletter.com/subscribe