Changelog & Friends — Episode 22

Developer (un)happiness

Abi Noda from DX discusses the Stack Overflow 2024 survey revealing 80% of developers are unhappy, why technical debt and complexity are the real culprits, and the DX Core 4 framework.

Transcript(85 segments)
  1. SPEAKER_00

    Welcome to Changelog and Friends, our weekly talk show about happy devs. Just a little happy dev, a little happy dev over there. Big thank you to our friends and our partners at fly .io. That is the public cloud that helps productive developers ship. Learn more at fly .io. Okay, let's get happy. Hey friends, I'm here with Dave Rosenthal, CTO of Sentry. So Dave, I know lots of developers know about Sentry, know about the platform, because hey, we use Sentry and we love Sentry. And I know that tracing is one of the next big frontiers for Sentry. Why add tracing to the platform? Why tracing and why now? When we first launched the ability to collect tracing data, we were really emphasizing the performance aspect of that, the kind of application performance monitoring aspect, because you have these things that are spans that measure how long something takes. And so the natural thing is to try to graph their durations and think about their durations and warn somebody if the durations are getting too long. But what we've realized is that the performance stuff ends up being just a bunch of gauges to look at. And it's not super actionable. Sentry is all about this notion of debuggability and actually making it easier to fix the problem, not just sort of giving you more gauges. A lot of what we're trying to do now is focus a little bit less on the sort of just the performance monitoring side of things and turn tracing into a tool that actually aids the debuggability of problems. I love it. Okay, so they mean it when they say code breaks, fix it faster with Sentry. More than 100 ,000 growing teams use Sentry to find problems fast. And you can too. Learn more at Sentry .io. That's S -E -N -T -R -Y dot I -O. And use our code CHANGELOG. Get $100 off the team plan. That's almost four months free for you to try out Sentry. Once again, Sentry .io. Well, developers are unhappy. That's the sentiment, right? That is the sentiment. Why? Are you happy, Jared? You're a developer, right? Are you in the 80 % rule or are you in the 20 % rule? It depends on the minute of the particular day. Okay. Whether or not I'm happy or unhappy. It's a fleeting thing, happiness. Am I satisfied in my work? Yes. Do I always think that? No. Am I a typical developer? Probably not anymore. We've been podcasters

  2. SPEAKER_01

    now

  3. SPEAKER_00

    for a long time. And so I don't hold a nine to five software job, which is probably the people mostly who are being interviewed or surveyed. I wasn't in that survey. So my sentiment was not in there. No, I did not take the Stack Overflow. We have Abhi noted here with us from DX. Abhi, did you take the Stack Overflow survey? I did not, but I've definitely been looking at the results. Interesting results. 80 % is a large number. I mean, that's an overwhelming number and it's not a small survey. Pareto's principle says

  4. SPEAKER_01

    80

  5. SPEAKER_00

    -20. That's the big principle. Apparently it's true in regards to developer's happiness. Or lack thereof, I guess. That would be the 20. The happy would be 20 and the unhappy would be 80. What's interesting about this. So this came out, as we said, from the 2024 Stack Overflow survey results. Synthesized by ShiftMag. So shout out to Anastasia Uspensky at ShiftMag for really highlighting this particular point and pulling together a few other data points to try to figure out. She was trying to figure out why. Why are they unhappy? And so you might think, well, it's the AI. The AI is taking away our joy. That doesn't seem like that's the case. At least that's what her conclusion is. It's not the AI. The AI is making us slightly more productive and maybe a little bit more apprehensive about the future, but the currently I think developers who are in their seats writing code know that at least today they aren't being replaced in large swaths by AI. So if you're a good software developer today, you're not too worried about that. At least not in the present. And it's not the stuff they're working on necessarily, but it is other things. Other things like tech debt and complexity. And so that kind of comes out in all kinds of different ways. But that was her finding. Abhi, do you have, you run a survey company, right? You guys create surveys for folks? He said it. He said it, Abhi. He called you a survey company. Sorry. Is that reductive? That's a jab. I'm just kidding. I don't know. I don't mean to reduce. I don't mean to reduce. I'm just throwing some jokes out there. It's friends. You got to do it, you know? No, it's fair. It's fair. I don't mean to reduce, but you all help people do surveys. Yep. Just curious your thoughts as we kick into this topic. We help people do surveys and collect other types of data on their developers. Just to clarify. Good job, Abhi.

  6. SPEAKER_00

    Yeah. The survey is really interesting. You know, one of the first things that came to mind when I read that headline that 80 % of developers being unhappy was something we see across organizations we work with. Something a little bit similar. We track something around that we call attrition risk. So what is the likelihood of a developer actually leaving a company in the next 12 months? And that number typically hovers around 10 % to 15%. OK. And so one of the first things that came to mind, what are the implications of 80 % of developers being happy, right? If only 15 % of them are actually going to leave the company, right? And, you know, that amounts to a lot of unhappy employees who are not doing their best work, who are probably not clocking in the 40, 50 hours that we're hoping for, who may be, you know, phoning it in a little bit. So that was interesting to me, just reading that headline. I always go bigger when I see something like a industry like software development and I start thinking, and we don't have answers necessarily, but I start wondering like, well, how many of workers are happy? You know, just in general, like is 80 % like ridiculously large? It is in absolute terms, right? It's four out of five. That's a large percentage. But if we compared it to some other industry, you know, medical workers, teachers, plumbers, pick your industry. Would they be at like maybe 75 % or 85 % or are they down there in the 40s and 50s? And we're way out of line. That's the question that I usually ask and I don't have the answer ever. So I kind of just twiddle my thumbs and move on. I was thinking that too, the macro versus the micro, what's the Des versus the world aspect of this? Because I would imagine that medical workers as an assumption are generally, especially since the pandemic are higher to be unhappy for obvious reasons. A lot of pressure put on them. A lot of change. I think a lot of bureaucracy, a lot of things in that system. Plumbers, I'm not so sure. Plumbers, if you're an indie plumber, you're probably pretty happy. Plumbers make pretty good money. And they generally call their own shots, kind of hard to replace. Call them in a pinch. It's like, hey, listen, I got water on my floor, man. You got to come help me out here. Right. And they jump on it. They're like eight, 500 bucks. Thank you very much. Right. All you did was turn the nut. Come on now. It's a relatively stable industry. I mean, you're always going to have people with plumbing, new plumbing, plumbing problems, et cetera. So it's not as much affected by perhaps the federal reserve like we are. The medical industry went through a huge swell, of course, during COVID where there was just so many needs for medical workers that their salaries went through the roof. They were in huge demand. Of course, they worked ridiculously long and trying hours. And so that was probably not producing happiness, but the pay was really, really good. And now coming down on the other side of it, it's similar to the software world where it's like demand is waning. Jobs are harder to find. You may go unemployed for a while. And so there's probably a similar chart, if you were to chart overall demand. Teachers, though, is a good one to cross examine on that. Teachers are never happy, are they? I mean, they're so under resourced. They're struggling. I just feel like most teachers probably would love to be happier. I don't know. What's funny, though, is I wonder how you can Venn diagram happiness with job happiness. Because I meet a lot of teachers that are very happy, very joyful, very purposeful. Serving loving people and happy in life. But then you say, are you happy with your job? And I wonder now if we zoom out to this happiness, unhappiness level with devs even, because some of the findings said that code was not what made developers unhappy, because most of them are doing things on the side, either through learning or for career development, things like that. I just wonder how much is it job unhappiness? Is it unhappiness generally? Because a lot of people, especially in the United States, that's where my lens is, that's where I live, is generally, generally unhappy, like with a lot of things. So does that like spill over the... Trickle over. Yeah. I think it does. This particular question was specific to like, are you happy with your job? And so that is the context that we're talking about. But of course, nobody just draws a wall up around their job. And like, as they walk through the door to work, all of a sudden they're like this different feeling person. These things do affect each other. It's interesting. Yeah, the question was how satisfied are you in your current professional developer role?

  7. SPEAKER_01

    And

  8. SPEAKER_00

    the options were not happy at work, complacent at work, and happy at work. So actually, of that 80 % who are reported to be unhappy, 47 % are complacent. So they didn't say they were unhappy. They said... Right. Meh. They said, meh. So like what this number is, is you take the happy people and it's 20%. And then there's two categories that make up the 80. And a large part of that is not like, I hate my job. They're just like, you know, it's a job. Which isn't all that bad. I mean, a job is a job because it's work. I mean, it's not... I know we have a culture and of course, the desire to like, follow your passion and do what you love and all of these things. But that's the few and the proud usually who can actually do that. You know, it's not very many of us who can do what we love all the time. And it feels like I would do this if I wasn't getting paid for it. Like that's not the normal. And so just being kind of meh with your job is... It could be worse, right? It could be worse. What I think is really interesting is the why, right? So why are developers satisfied or unsatisfied in their jobs? And I think the first images that pop into our minds might be pay or managers or layoffs or AI. But if I'm reading this correctly, the top contributors to satisfaction are actually the developer experience or technical debt, right? The tooling, the complexity of the systems and the code base. Am I reading that correctly? Is that how you guys read it? That's how I read it as well. Yes. Technical debt and complexity are the two driving factors to this unhappiness. Which effectively is developer experience. I mean, it's your work. And how did we get there? I think it's just like two decades of move fast and break things, isn't it? I mean, isn't that just

  9. SPEAKER_01

    kind

  10. SPEAKER_00

    of how we've gotten here? That's my best guess. Maybe. Yeah, two decades of move fast, break things, hire a lot of people, churn a lot of people, reorg many times. And now everything's a mess. So in this tracking that you do with regard to attrition, 15 to 25%, is that what you said? Yeah. In the next 12 months, likely to move somewhere else. Do you also get qualitative information about like, why? Like, why are they moving on? Is it similar things? Yeah. Yeah. So we're focused on measuring the developer experience. A lot of the things listed here, you know, difficulty of understanding code or developer environments, CICD, strategy on the team. A lot of these things are aspects of the developer experience we measure for lots of different companies. And then we correlate the two. So we correlate these different aspects or facets of developer experience against who's at risk of leaving and who's actually left. And our data actually aligns quite a bit with what I'm looking at here with the StackOverflow report. Yeah. The difficulty of doing work as a developer seems to be the preeminent cause of regrettable attrition for companies. Not pay, not liking your manager, not stock compensation. It often is just the difficulty of actually doing work. Which can manifest in technical issues, but also bureaucratic issues. Makes it harder to be productive. Right. You're feeling like you're not getting anything done or you're constantly

  11. SPEAKER_01

    like

  12. SPEAKER_00

    working JIRA tickets and you come in in the morning and you got 20 open tickets and you work eight hours and you sweat and you bleed and you leave and you got 22 open tickets and you're like, I'm never going to, I'm never going to get myself up from here into a place of progress. You just feel like maintenance, maintenance is all it is. And I can see how that would be demoralizing, especially over time. Yeah. When it's not getting better. I mean, it's demoralizing for developers. It's also demoralizing for leaders. I talked to who, you know, who run are getting this type of data at their companies and quarter after quarter, despite making efforts to, to make improvements around this stuff. The data keeps, keeps coming back that things are slow. People are frustrated. It's hard to get work done. So I guess the question is, is how do you solve that problem? Is it the organization's problem? Is it the leadership's problem? Is it the product's problem? Is it the market's problem? Because I think a lot of that complexity comes from the fact that solving software problems are hard generally. Being blocked is very common. Having to help others level up or answer questions is very common. And that's going to be pretty much a thing. I guess potentially if AI starts to solve some of this for us, that gets to be reduced some. This blockage, so to speak, this spending time looking for answers, spending time answering answers or repeating answers for people. This, the blockage that comes from the lack of awareness of where to go next to be productive. At GitHub, I think I told this story to you before, Adam. Tell it again. At GitHub, we had a lot of these problems. Developer tooling, getting releases out, the builds, developer environments, people were leaving and they were telling leaders that they were leaving because it was hard to get things done at GitHub. This is back in 2020, if I'm getting my ears right. Well, hindsight is 2020. And what we ended up doing, we froze features for a quarter. All of GitHub engineering, no features, whole quarter spent fixing these problems. It was dramatic, right? And things got a lot better as a result. And another example is actually Atlassian. Their CTO is very public about how they're focusing on developer productivity, developer experience. And at Atlassian, not only do they have a pretty substantial portion of their engineering organization that is devoted to this type of stuff, but they give all product engineering teams at Atlassian 10 % of their time to be spent fixing things, as they call it, fixing things that suck, right, that get in the way of productivity. So but to answer your question, Adam, what do we do about it and what's preventing us from doing things about it? I think it actually boils down to the fact that you see a survey like this from Stack Overflow, right? People are unhappy. You know, it's because of technical debt and the developer experience. But to actually do something about it as a business, you have to be able to calculate that the cost of doing something about it is outweighed by the return on investment you're going to get after you do something about it.

  13. SPEAKER_01

    And I think

  14. SPEAKER_00

    that's a really hard problem right now. No one knows how to actually quantify this

  15. SPEAKER_01

    set

  16. SPEAKER_00

    of things, the developer experience. It is something that you can take to the CFO or CEO and say, hey, like we were slow because of X, Y, and Z. And if we fix X, Y, and Z, we'll be this much faster and it'll be worth it. That's that's the hard problem. So no one can make the case for doing something about a lot of this stuff because you can't talk about it in terms of dollars. Yeah, I think this speaks to our technical debt metaphor and some of the argumentation we've had on this show with friends about is that a good metaphor or not? Because you can't really quantify it like you can actual debt. And you can take your debt and your debt service principle and interest, you can take your interest rate and you can put that on a chart and you can extrapolate it and say, look, if we don't pay this debt down now, maybe not if you're the United States government, but if you're like an actual business, you can say, we don't pay this debt down. We're going to go bankrupt in 90 days. And that convinces leadership to be like, okay, it's worth it. But when it comes to technical debt, we lack that quantitative ability to extrapolate forward and say, we're going this slow right now. If we don't dramatically change things, start paying this down, we're going to grind to a halt in 90 days. What has been the happiness level of past surveys? Can we just use Stack Overflow surveys as an example, because that's what we're lensing off of anyways, if that's a correct adjective or verb. Has the unhappiness changed dramatically from 40, 50 to now 80 percent? Has it always been 80 percent? Is that maybe a good baseline? Like mostly people are unhappy. That's a good thing because the reason why I ask that question is because innovation comes from angst. Unhappiness is a version of angst, right? And so you can only innovate and change if you have angst, I suppose, to some degree. I mean, it's one place. Like greed also drives innovation, right? Like I want to make money. I need to invent something to make more money. But the greed may be causing the angst of the developers there. So I mean, you know, they're in the same bucket, basically. Yeah, okay. The angst is there, therefore developers push, organizations push, products change, innovation happens, you know, the new Amazon occurs. Because I don't have past Stack Overflow survey data. Do you, Avi? Do you, Avi? I don't. No. Dang. Yeah. Someone in the audience is like, I've got it, but I can't talk to you. The way they ask that question, I wonder if that's how they've asked it before. So I'm not, I'd be curious if you could. Can you restate that question? Because I think that's a good point is the question and the only answers. It's multiple choice. This is not an open ended question of why. This is a scoped response. And so the 80 % is extrapolated from that scoped response. Can you restate the question and the options? So how satisfied are you in your current professional developer role? Not happy at work, complacent at work, happy at work. As someone who spends a lot of time on survey design, I do see a few potential issues with the, I mean, so they're asking about satisfaction. But then the responses are about happiness, which in satisfaction or happiness are really different constructs. That middle option is complacent. So it's not happy, happy or complacent. But complacency does, it's not really the perfect middle between not happy and happy. I think it captures the essence of in between not happy and happy, but it's not necessarily the perfect middle. So it's an interesting way that they've asked the question, because is it measuring job satisfaction or happiness? Happiness, I think, is really hard to actually measure. So I think that's why they worded it around satisfaction. You know, happiness, there's a lot of literature about how to actually measure happiness. There's entire fields where they've spent years trying to figure out how to measure happiness and what happiness is. And usually happiness is the sum of moments of feeling happy. Like if you took the day and divided it up into however many minutes, like in each minute or how many times throughout the day did you feel, have a feeling of happiness as opposed to non -happiness? And that's kind of how you get to how happy you are, right? Rather than a point in time reflection of happiness, which is pretty difficult. So anyways, I'm nerding out here a little bit on the survey design. I like that because you have experience that we don't have

  17. SPEAKER_01

    in

  18. SPEAKER_00

    trying to like craft those so that they are optimal because you only have so much time and opportunities to like pull somebody for their thoughts. And if you pull them out incorrectly on accident, then you're kind of wasting everybody's time. Let me rant for a split second here about stack overflow and URLs. Okay. So first of all, I appreciate you all doing the survey. No real hate here, but I was trying to answer Adam's question, which was, do we have past year's results? So I went to this year's results, survey .stackoverflow .co slash 2024 slash professional developers found the link to that survey question. And then I went to the URL bar and I changed the year from 2024 to 2023, 404 page not found. I mean, come on people. Respect the URL structure. This is like what address bar hacking is all about. Like, come on, help us get to things in a way that makes sense. I just appreciate good URLs and that's not a good one. Anyways, that was my mini rant because I was going to have answers for you, Adam. I was going to have last year's answers to this question. I wanted you to have answers so bad. I don't have it, man. I don't have it. Maybe chat GPT or something else might have a hallucinated version of an answer. The reason why I think you nerding out on that and camping out on the semantics of the question and the response is because it certainly, it corners the person. It forces the person in response time. At the same time, there are probably lots of questions in the survey, so they could be experiencing cognitive overload while at that particular question, while also being slightly unhappy for their day. They may have measured their happiness moments in that day and be like, you know what? I'm unhappy. Not saying it's cute, but it's important to scrutinize the question and the offered options as a response because that is what the sentiment is drawn from. And so if it's skewed or not so much poorly worded, I would, I would prefer you to say the Ivy versus me because you're the professional at crafting these in quotes, surveys. Just kidding. Poorly designed. Kidding. Yeah. You know, because that's really important, right? The way you ask a question, the options you offer is where the sentiment comes from. And if it is ambiguous or not super clear, it's clear why the answer is potentially skewed. And so to understand how the efficacy of the answer set based on the question, I think that's what's worth scrutinizing. All right. Real -time follow -up. I used their user interface to find last year's results. And as far as I can tell, they did not ask this question in 2023. Maybe it was just one year they didn't, but we do not have last year's answer to this particular question. That's not why at 404, okay, they still change their URL structure, but had it stayed the same, it still would have 404'd because they didn't ask that question. So unfortunately, we can't really go back and say, you know, which way is it trending or is this an anomaly or anything like that? I want to talk about this idea of how can people talk about these problems in terms of dollars, right? I would love to hear this. Yes. Yeah. So that's something we're working on at DX. It's funny that Adam posed the question earlier, why aren't people fixing this? Because that's the same question that we ask and our customers that we work with ask too when they surface all this data. So many years ago, I read this book called How to Measure Anything. Have you guys read that book? No. Okay. The topic is, as the title implies, how to measure anything. And in particular, how to measure things that are seemingly unmeasurable. So when we talk about what is the dollar ROI or interest rate or cost of technical debt and poor developer experience that just a few minutes ago, we were essentially calling that unmeasurable, right? In this book, they talk about how anything, if you want to measure something in terms of dollars or cost, that you can really do that with anything. So as long as you take something intangible, like technical debt or developer experience, and

  19. SPEAKER_00

    then you correlate it to something objective or something monetary. So an example of this would be the DORA metrics and the DORA report, right? Which I know you guys have followed. So what they essentially did is give us a primer for those who aren't caught up on that. What is DORA? DORA is the DevOps research and assessment. But since I want to say 2013, maybe they've been publishing an annual report on the state of DevOps. And right now we're talking about tech debt and developer experience. But eight years ago, people were talking about DevOps. And hey, what is the ROI of investing in DevOps? So it's the same problem. History repeats itself. And what they did was they said, here are some ways we can measure DevOps. So like metrics like MTTR and lead time, deployment frequency. So they said, here's DevOps. And what they did is correlate it to company's profitability, stock returns and increases, you know, EMPS scores. And by doing that, they were able to quote unquote prove and show the dollar ROI. Hey, companies, when they invest in DevOps and get X percentage better, their stock prices tend to be X percentage higher, right? They tend to be X percentage more profitable. And so it wasn't perfect. That seems a little bit brittle to me. A little bit brittle. Let me tell you what we're doing with developer experience. So we have this construct of what is developer experience. We have our version of what Stack Overflow has here, where we have it's called the Developer Experience Index. So it's 14 of the top drivers of developer experience. So we say, OK, that's that's how we measure developer experience. Then what we've been doing is correlating that measure to different outcomes. And one of them is actually self -reported time waste reported by developers. So how much time do you it's a series of different questions we ask about how much time do you lose each week, how much time is wasted each week due to inefficiencies. And when we correlate the two and we found that like a one point increase in the developer experience index score, which is the average of these 14 different areas of developer experience, a one point increase in that score. So one point increase in developer experience translates to almost a one percent decrease and time wasted. And so, again, this isn't perfect. You could call this brittle as well, but I think it's a little bit better because you're directly asking the people. Yeah. And it's more direct to dollars. It's not like stock price, which is a little bit of a leap. Right. A lot of things, so many things affect the stock price. So, you know, using this approach, we can folks can say, hey, if we improve developer experience by X points, that translates to X percentage reduction in waste, which translates to X amount of dollars. Right. So that's that's how we're approaching it right now. I like that approach. I think that time waste is reported by the actual people wasting the time. And so it's probably relatively reliable. Of course, there's always trolls and thoughtless respondents, but you can't get around that. Or estimators, apparently.

  20. SPEAKER_01

    Yeah.

  21. SPEAKER_00

    I'm totally just being inefficient because of all these other things. It's not me. It's you. There's that whole I mean, you can't really maybe you just account for that and your numbers. But yeah, if you are saying technical debt, complexity, bureaucracy, whatever it is, all these factors ultimately for the business are costing money, slowing things down, wasting time is really a decent measure for that, like how much time is actually being wasted. And so if you can track that against this DXIX, what's this thing called? The DX index? DXI. Yeah. DXI, developer experience index. At the same time, I don't know, seems like a pretty decent approach here. Is that bearing fruit? It is. Yeah. I mean, it's we can't think of any other way to do it. I think the feedback we get is this is great. If we can make this an industry standard, then my CEO is going to buy it. But there's still an education and marketing gap there where folks, what I just explained to you, it's hard to get that across in like a five minute executive summary, right? Right. Do you do an annual or semi -annual survey to the public like Stack Overflow does? No, we should. We already have the data because we are already we're surveying hundreds of thousands of people. Yeah. The nice thing about this particular measure or this combination of measures is that if it could be somewhat generalized and made public, it's now a tool and a resource for people who don't have those quantitative metrics inside their company to say, look, this stuff really matters. Look what it did for Walmart and these important companies. It's moving their bottom line. It's making them more productive and they proved it out over end years. And so if that's public information that I can take to my leadership and use that, then convince them that, hey, let's call a feature freeze or whatever it is that I'm trying to get done.

  22. SPEAKER_01

    Yeah.

  23. SPEAKER_00

    Hey friends, you know, we're big fans of Fly .io and I'm here with Curt Mackey, co -founder and CEO of Fly. Curt, we've had some conversations and I've heard you say that public clouds suck. What is your personal lens into public clouds sucking and how does Fly not suck? All right. So public clouds suck. I actually think most ways of hosting stuff on the internet sucks. And I have a lot of theories about why this is, but it almost doesn't matter. The reality is if like I've built a new app for like generating sandwich recipes because my family's just into specific types of sandwiches. They use Brown Schweiger as a component, for example. And then I want to like put that somewhere. You go to AWS and it's harder than just going and getting like a dedicated server from Hetzer. It's like, it's actually like more complicated to figure out how to deploy my dumb sandwich app on top of AWS because it's not built for me as a developer to be productive with. It's built for other people. It's built for platform teams to kind of build the infrastructure of their dreams and hopefully create a new UX that's useful for the developers that they work with. And again, I feel like every time I talk about this, it's like, I'm just too impatient. I don't particularly want to go figure so many things out purely to put my sandwich app in front of people. And I don't particularly want to have to go talk to a

  24. SPEAKER_01

    platform team once my sandwich app becomes a huge startup and IPOs and I have to like do a deploy. I kind of

  25. SPEAKER_00

    feel like all that stuff should just work for me without me having to go ask permission or talk to anyone else. And so this is a lot of, it's informed a lot of how we've built Fly. Like we're still a public cloud. We still have a lot of very similar low level primitives as the bigger guys, but in general, they're designed to be used directly by developers. They're not built for a platform team to kind of cobble together. They're designed to be useful quickly for developers. One of the ways we've thought about this is if you can turn a very difficult problem into a two hour problem, people will build much more interesting types of apps. And so this is why we've done things like made it easy to run an app multi -region. Most companies don't run multi -region apps on public clouds because it's, it's functionally impossible to do without a huge amount of upfront effort. It's why we've made things like the virtual machine primitives behind just a simple API. Most people don't do like code sandboxing or their own virtualization because it's just not really easy. It's not, there's just no path to that on top of the clouds. So in general, like I feel like, and it's not really fair of me to say public clouds suck because they were built for a different time. If you build one of these things starting in 2007, the world's very different than it is right now. And so a lot of what I'm saying, I think is that public clouds are kind of old and there's a new version of public clouds that we should all be building on top of that are definitely gonna make me as a developer much happier than I was like five or six years ago when I was kind of stuck in this quagmire. So AWS was built for a different era, a different cloud era and fly a public cloud. Yes, but a public cloud built for developers who ship that's a difference. We here at Change .io are developers who ship so you should trust us. Try out fly, fly .io over 3 million apps that includes us have launched on fly. They leverage the global anycast load balancing, the zero config, private networking, hardware isolation, instant wire guard VPN connections with push button deployments, scaling to thousands of instances. This is the cloud you want. Check it out. Fly .io again, fly .io. And I'm also here with Kyle Carberry, co -founder and CTO over at coder .com and they pair well with fly .io. Coder is an open source cloud development environment, a CDE. You can host this in your cloud or on premise. So Kyle, walk me through the process. A CDE lets developers put their development environment in the cloud. Walk me through the process. They get an invite from their platform team to join their Coder instance. They got to sign in, set up their keys, set up their code editor. How's it work? Step one for them. We try to make it remarkably easy for the dev. We never get any features ever for the developer. They'll click that link that their platform team sends out. They'll sign in with OIDC or Google and they'll really just press one button to create a development environment. Now that might provision like a Kubernetes pod or an AWS VM. You know, we'll show the user what's provisioned, but they don't really have to care. From that point, you'll see a couple of buttons appear to open the editors that you're used to, like VSCode desktop or, you know, VSCode through the web, or you can install our CLI. Through our CLI, you really just log into Coder and we take care of everything for you. When you SSH into a workspace, you don't have to worry about keys. It really just kind of like beautifully magically works in the background for you and connects you to your workspace. We actually connect peer to peer as well. You know, if the Coder server goes down for a second because of an upgrade, you don't have to worry about disconnects and we always get you the lowest latency possible. One of our core values is we'll never be slower than SSH, period, full stop. And so we connect you peer to peer directly to the workspace. So it feels just as native as it possibly could. Very cool. Thank you, Kyle. Well, friends, it might be time to consider a cloud development environment, a CDE and open source is awesome and Coder is fully open source. You can go to Coder .com right now, install Coder open source, start a premium trial or get a demo. For me, my first step, I installed it on my Proxmox box and played with it. It was so cool. I loved it. Again, Coder .com. That's C -O -D -E -R dot com. My idea for you, Bobby, is a growth hack. What's your way when you when you do this? It would make sense if I were you. This is probably how I would at least consider it. This is not a perfect one to one plan, but we're going to solve. OK, this Stack Overflow survey is obviously popular. We're talking about it. The results are shared and, you know, examined and analyzed by many. It's respected, right? It's it's got a core audience. If you have similar data, I would release whatever you're doing or whatever data you can do around the time of this survey's announcement. And to some degree, Venn diagram number one, you associate the brand for DX with a very beloved, mostly beloved brand Stack Overflow. Some love, some hate. I thought you were talking about WordPress. Oh, yeah, exactly. Not yet. Not yet.

  26. SPEAKER_01

    And

  27. SPEAKER_00

    then you can you can draw core layers between the questions that and the database siphoned from that question. Yeah. And then the question and or data set that you have that correlates and Venn diagrams across the two. One, to keep them honest and not so much that they're not honest, but to keep this survey, which all surveys have a

  28. SPEAKER_01

    an

  29. SPEAKER_00

    optimization opportunity, right? They're all flawed. Talked about. Right. There's no perfect survey. And I think you you almost better off the entire community because you give not one data set, but two data sets. So how true is this? Your findings and cross -examination and Venn diagram may say, well, this is actually pretty close to true because we have corollary data and we can corroborate this this findings. Second, you get to feature things like DXI and you get to have an opportunity for now way more people know about DX and now find the benefit and or interest in your beliefs, which is this DXI index being such a core thing to you all. One, that's the idea. How do you like it? I like it a lot. I'm writing it down. I like the cross -examination piece. And then I would say the second thing and maybe this gives a foundation to a foundation, which is in order to have an organization support or adopt this DXI, this developer experience index, what do they need to have in place to get to that point? Like, what is a mature data -driven organization like that has the ability to actually index this index and have that for themselves? Yeah, I love it. That's a question. Taking notes. That was a question. Oh, what do they actually need? I like the question, but that was also a question. Yeah. I mean, what's the foundation to get to this point to have it? 14 metrics, right? Isn't that 14 metrics? Yeah, 14 different metrics. So we have an open source that right now it's proprietary. Yeah, I mean, that's actually

  30. SPEAKER_01

    one

  31. SPEAKER_00

    of the biggest strategic questions I've been wrestling with for over a year and a half. Do we open it up or do we keep it proprietary? Well, when you have the opportunity to become the index, I think, I mean, obviously, you know way more about your business than I do and how important it is internally as a proprietary thing, but I can see a huge upside in the open sourcing of it. Yeah. Yeah, agreed. I think we can open source it while putting like a copyright on it. So you can't necessarily, you know, you're not technically supposed to use it for commercial, you know, within, like Walmart can't actually use, deploy it.

  32. SPEAKER_01

    Yeah,

  33. SPEAKER_00

    I mean, we can get into the licensing discussions and we're happy to. We do it all the time. Yeah. And depending on what it is, open source might not even be the right

  34. SPEAKER_01

    word,

  35. SPEAKER_00

    right? Like maybe it's creative commons, maybe it's right, but you could still hold trademark and copyright against it. Like DXI could be a trademark thing, but also how you go about doing it and how others can go about doing it. You can just let that stuff loose. You don't let go of the copyright, but you just let other people use it. So you can call it DXI. It is a product, right? You trademark the DXI, right? That's what you're saying. Right. Makes sense. They can be DXI compatible or, you know, whatever, but that's in the weeds. You were talking about the 14. Yeah. I mean, the deploy it, they, you know, we, we have the survey items, the, the measurements for, for those 14 and they deploy our platform, right? That's why it's proprietary because they can't do it without, without our product from you. Yeah. But in terms of like who, you know, we work with, we work with the Pinterest, the drop boxes, the Netflix's, you know, the, the, the, the bleeding edge tech companies. And we also work with, I mean, this isn't to, to be you know, diminish these organizations, but companies like Pfizer, P and G, Tesco bank in New York, BNY. So I think, you know, what we've seen is the DXI works in all kinds of environments, not just the bleeding edge tech companies, but, you know, you know, more legacy traditional enterprises as well. Sounds pretty cool. So you have found then if you have a DXI, which a lot of these companies do via deploying your guys's proprietary platform and you're tracking time waste, you found that a, there is an inverse correlation between the two that is measurable and repeatable and reliable. That's pretty powerful. I mean, we all know it's true, but like actually proving that it's true is a whole different thing. I mean, we, so we, we did a meta analysis, which we've published and I'll give you guys a link to that. We actually have a developer experience index .com. I have that vanity URL. So long. It is a little long. Gosh. Anyways, so we, you know, the, the R value was point, I think it was like 0 .75, but that's a really strong relationship between developer experience and time waste. So then on an individual company basis, we can look at their relationship, right? We just run that correlation for an individual company. And we, we always see a moderate to strong relationship at any given organization as well. And do you find that it follows Pareto's principle as well in terms of effort, like 20 % of your effort gets 80 % of your results or as you continue to improve your DX, are you, is it trailing off or not? That's a great question. Like, do we see a different relationship at different at the edges? Right. Yeah. We haven't studied that,

  36. SPEAKER_01

    but

  37. SPEAKER_00

    I will add that to my notes as well. That's, that's really interesting. Yeah. That would be worth knowing. Yeah. It's even, I mean, I think it's, it's logical that that would be the case in almost any effort at a certain point. You're, you're squeezing the radish, you know, but like, what's the sweet spot for, for companies where they can put in this much effort into their developer experience and get that much out. Yeah. I think that would be worth knowing. Absolutely. Can we dig into these 14 drivers? It is out there. Can we talk about them at least? Yeah. Did you go to developer experience index .com? No, I just googled get DX and then DXI and it landed me on this page that you can tell me that this is accurate. The number one. Yeah, that's the white paper. The one number you need to increase ROI per engineer. Yep. And about two scrolls down, you dig into figure two, which, which talks about the drivers and the outcomes.

  38. SPEAKER_01

    Yep.

  39. SPEAKER_00

    And so I'll do the work for you. If you don't mind, the drivers are deep work, dev environment, batch size, local iteration, production, debugging, ease of release, incident response, build and test, code review, documentation, code maintainability, change confidence, cross team collaboration, and planning. Those are the drivers. Those are the 14 dimensions. And those correlate to five outcomes. Speed, ease of delivery, quality, engagement, and efficiency. Dude, that's a good map. That's a really good map to maturity in an organization, a dev organization. Like all those things on the driver's side are really good. Like what is my maturity level and what is my, I don't know how you would describe it. I'm trying to think on the fly here, but how good am I? How good are we at these drivers? And then the correlations are obviously awesome with the outcomes, the speed, the ease of delivery, quality, engagement, and efficiency. Yeah. But that's a good map. I like that. That's taking years to arrive at those 14. You could almost map softwares as a service on top of that sucker. I mean, there's like an offering for each of these drives. I mean, there's a whole industry around. Oh yeah. Production, debugging, incident response, code review, et cetera, et cetera. I just find that interesting. Which one, which SaaS service correlates to change confidence most? That's the one that stands out to me a lot. Like change is hard anyways, and the confidence and change. You can be a senior engineer and feel good about it. You can also be a junior engineer and feel good about it. But what gives you the confidence? How do you measure that with a service, a tour, a SaaS? Yeah. I think change confidence, it's a lot about how easy it is to actually test a change, to get feedback on a change. So I think everything from cloud dev environments, all these things kind of interrelate. But so cloud dev environments that allow you to quickly spin up a staging environment for just your change to easily manually test stuff. Obviously, things like test coverage. Now, AI is coming into the picture there with helping write tests or even manually QA your work. So yeah, change confidence is about like when you make a code change, are you kind of like yellowing it when you ship it and just hoping it works? Or do you actually feel confident that when you make a change? It also just has to do with code quality, right? If you make a change in one area of the code, is it a house of cards or cascading effects and everything? So a lot of things go into it, but it's ultimately about the developers feel like they can actually make changes, or are they just duct taping things and hoping that it works when they deploy it. Another newish feature of a lot of cloud hosts are preview branches. That's another way where you can get change confidence. Netlify, Vercel, et cetera. They're providing a place where you can have your development branch, and it can be constantly publishing to a preview page on a subdomain on a website. And so now you can both look at it yourself in production ish and then also send it to your QA team or to your boss or whoever your customer. I think that definitely helps with change confidence because

  40. SPEAKER_01

    previews

  41. SPEAKER_00

    are nice. But yeah, there's so many tools that overlap in these things as well. Documentation, if you're modifying code, can you even understand how that code works so you're confident in changing that little bit of code? So yeah, a lot of things, all these factors interrelate. Even something like batch size, which is about incremental delivery. If you're working on huge changes, huge PRs, there's so much more risk. You just have a lot more surface area. So if you're delivering incrementally, your confidence is going to be higher each for each unit of change. Is this your next big thing, the DXI? It's been in the works for a while. DXI is one of the big things. The other is the core four, which is the other. If you go to that research tab. That one rhymes, so you know it's true. DX core four. Yeah. So if you go to that research tab on our website and there's the developer experience index and above it, the DX core four is something else we've been developing. And that is speed, effectiveness, quality, and impact, right? So that's the outcome of this. But the real problem we've been trying to solve is, I think last year I came on here at him and we talked about the DevEx framework I'd published with Nicole Forsgren and others. So ever since we published that, people have been coming to us and saying, Nicole created Dora, Nicole and Margaret created the space framework. And then you, Nicole and Margaret created the DevEx framework. We have three frameworks now for telling us what we're supposed to be measuring in our organizations. How do we actually, to sum it all up, what should we actually be measuring? And the funny, I was just talking. Add one more, add one more, you got the core four. This is the one to rule them all. This is the one to rule them, the one framework to rule them all. Because not replace, this encapsulates all of them. This combines them all into one framework. Because yeah, everyone would ask us that question. And the way we would always answer that question was, well, it depends. I was just talking to a CEO at a big tech company who said, I was talking to Nicole and I asked her like to sum it all up, what should we measure? And she told me, it sort of depends. And I get that it's situationally dependent. But it would be really valuable to have something like out of the box and standardized that we could benchmark across other companies and really have a way forward. So funny enough, I've had that same experience. I was talking to a CTO actually at Capital One who asked me, hey, I've been following your research for two years. So just tell me, what should we actually measure? And I said, it depends. We can do a consulting engagement on that until I figure it out. But having a starting point that is out of the box is really valuable. So that's what the DX core for is. I love that response. And as a person who has to deliver that response frequently, my next response is always, I have to ask you 20 questions to answer that one question. So you need to give me more time. If you want my true answer, the only way I can know what to respond with is to ask you several more questions. And those questions may lead to even more questions. And so if you trust me, you've been following my data, give me a little bit of your time. And I will answer those questions by asking tons of questions. The problem is no one has time. Well, they all want the silver bullet, right? Give me a yes or no answer. They want to go to the doctor and get the diagnosis, not go to the doctor and then have 16 follow -up appointments before you get the diagnosis. And I get that. I mean, if you were a time waster, then that's different. But like, you can only answer that CEOs of top tech companies answer a question well, if you understand more about their specifics of their business. What are their particular drivers? Not anymore, baby. Now he says core four. Core four, yeah. I think the core four provides a pretty good answer. I mean, we, you know, we want people to customize. This isn't pay, like do this and do nothing else. I like this actually. Core two, core three, maybe. But the DX core four does. And now we've started rolling it out to people and it's landing. It's landed well. You know, I actually asked the CEO, you know, after I showed him, I showed him the core four right after hearing about that, that experience that they had talking with Nicole said, Hey, we've been working on something for this. And he, I asked him to you as CEO, does this seem right? Does this seem correct? Does this seem like the right way to be thinking about a measuring developer productivity? And he said, yes. Wow. I was going to give you an idea, but that might actually be the answer because rather than say it depends, what if you said we have a survey that takes you five minutes to answer instead of saying it depends, it spits out your core customized core four. This is, this is your on -ramp like your specialized, personalized on -ramp. Yeah. Because I'm sure you can take that consulting session and to some degree distill it down into something, a CEO who has very limited time and answer in five to 10 minutes. Right. Hey, I don't have an answer for you in this moment, but we have a very fast 10 minute or less. It really could be 15 if you want it to be, but most people is 10. And if you answer these questions, I'll know exactly how to help you.

  42. SPEAKER_01

    That's

  43. SPEAKER_00

    like when, you know, some of those personal wealth front and betterment, some of those robo investment advisor platforms, right? You go through like a three minute wizard and then they tell you what your investment portfolio should be based on. So that I like that. More ideas for you, Abhi. Two, you're taken away from here. Yeah. Yeah. Lots of good ideas. Write that one down, Abhi. Write it down. Just write it down. Can we dig into these a little bit? So sure. The core four, speed, effectiveness, quality, impact. You say those are outcomes, not necessarily. Those are the dimensions. Those are the categories, right? But think of them as your stocks, bonds, cash, right? To use the stock portfolio analogy. You need that balance because if you only measure speed, anything else goes to crap, right? Like you're not doing it correctly. There's your move fast and break things right there. Like we're moving very fast,

  44. SPEAKER_01

    but

  45. SPEAKER_00

    we are breaking, not breaking things. Yeah. So a balanced portfolio. This is a nice metaphor. For each of these, you have a key metric. This is something that you're going to track. And then you have secondary metrics. So there's some balance there as well. But for speed, the key metric is diffs per engineer. And I don't know if I might take issue with that one. Yep. What? Tell me more. You know, I mean, I was probably last time I was on this show with Adam, I was probably dissing that metric, right? Like PR's for engineer. Finish change, man. Finish change. I'm a politician, right? You know, I flip flop on the issues. Yeah, it depends on who you're talking to. Yeah. But no. So it's been a journey for, I mean, just for me personally on this topic, because, you know, the whole reason I actually got into, you know, spending six, seven years on this whole problem space was because I felt like metrics like diffs per engineer were reductive and not correct and not helpful. All right. One of the things that the core four optimizes for is so we work with a lot of technical leaders, technical engineering leaders. And as we were talking about earlier, one of their big challenges is talking about rationalizing investments in developer productivity in a way that the CEO and CFO are going to agree to. And to do that, you need a shared definition of productivity that your CEO and CFO agree with. And to achieve that, I found that you do need some type of output measure. Right. Like like we're not at a point in human evolution yet where most CEOs and CFOs are down with this idea that like developer experience indexes is the one metric that matters for understand, you know, the maturity or effectiveness of the organization. A lot of CFOs and CEOs still think, I mean, there's Fortune 550 companies still measuring lines of code, benchmarking that. Right. So we're still at a point in human state of the art around software engineering where output measures need to be a part of that conversation. It needs to be part of the way you're framing developer experience and developer productivity. If you want the people you're pitching this to, to actually fund it and believe it and buy in. So there's a marketability optimization here like that. That's one of the reasons peers for engineers in here. But the other reason is, you know, we have come around to talking with a lot of companies like Uber, Microsoft, top tech companies where they use this metric as an input. It's not the sole metric, right? They're not performance valuing engineers based on this metric, but in aggregate, they are looking at this metric as an input to understanding how developer productivity is trending and compares to other organizations. And it's not useless, right? It is a useful indicator in aggregate. And that's why in the framework in the core four, we there's an asterisk and it says not to measure this at the individual level. So this is only to be looked at at a aggregate team group organization level and benchmark that way. And we've found it more useful than not. So it says difs per engineer, though, difs per engineer, then asterisks. Not at the individual level. Well, so, yeah, the metric is normalized. So you're looking at aggregate, aggregate divided by the population in terms of like visualizing or reporting this. You're not looking at a list of people, right? You're looking at teams and organizations. Yeah, right. I do see the contradiction there, though. Well, certainly at an individual level at face value, it seems contradictory, but it does make sense. Maybe you could reword that to say like averaged across whatever. But I'll write that down. Oh, man, so many good notes for you here. At an individual level, certainly it's a bad measure. Well, the problem is it becomes a bad measure, right? That's Goodhart's law. Like once everybody knows that that's what's being measured, well, we all know how to play the game.

  46. SPEAKER_01

    Same

  47. SPEAKER_00

    thing with I mean, it's lines of code moved to a slightly larger batch, you know? Yeah. And so I can criticize that one. I can also criticize lines of code. I can criticize features or tickets. They can all be criticized. But then you're at a certain point, you're like, well, what can we actually do that? Everything sucks. We're gonna have to pick one and go with it. And I guess if the industry is somewhat standardizing around that, then it's a decent compromise. And I think there's more we can do, right? I was just talking to a company actually working with a company, a Silicon Valley tech company, and all the other core four metrics were quite a bit below, like P 50, like an industry peers. But this per engineer was higher. And they this is bad for them because they're trying to show to their executives that they're behind peers so they can get funding to make improvements. Right. Sure. So we were just trying to dive into the data, like, why is your diffs per engineer inflated, even though clearly, like empirically and with the other core for data points, like you're not like a high performing organization. And so we couldn't really figure out an answer. I mean, there was a lot of speculate, like, you know, there's just there a higher number of like config changes, like small PRS that aren't real changes. But like every company has that, right? Like that should be kind of that fuzziness should already be kind of accounted for in our benchmarks. And so that led to this idea, like, you know, could there be a weighted metric? So so you're actually because not all depth peers are created equal, like we talked about, right? Some are one minute changes, some are some are one line changes that are actually eight hours, some are, you know, 800 line changes that are two minutes, like how do you so, you know, if we could apply some kind of waiting to like bucketing all these depths and PRS, almost the same way we do estimation, like T -shirt sizes or something like that. You know, I was thinking, could we use Gen AI like LLM to basically automatically try to categorize based on the title, the description of the task and the code changes? Like, you know, was this like a big change or was this actually a small change that could get kind of like a weighted number? That would be an improvement to the signal you're getting out of like an output measure like this. Yeah. Even with a confidence score. Yeah. Alongside that would be really interesting. There's still challenges with that because the amount of change does not always correlate to the amount of effort you can work an entire week on finding a bug and then you found it and it's a one character change and you're so exhausted by then that your commit message like says, I fixed it or something, you know? And so like the LLM just doesn't have much to work on.

  48. SPEAKER_01

    You

  49. SPEAKER_00

    know, I guess if you can just say, well, it's fuzzy, it's not going to be a percent. It's better than merely measuring. So did you come to my only guess would be like a culture of small diffs or a culture of you can't figure it out. Yeah. Why are they so much higher on that one metric? You haven't figured it out yet. We don't know, but they are definitely higher. And I mean, like I told them, look, like if I had a little bit more time here, I would take like a random sample of your 200 PRS and then like random sample of other companies and like try to do what an LLM would. I would like look at the titles and descriptions and like try to figure out like, are your PRS generally smaller, you know, lower effort or size tasks than other companies that I mean, that probably has to be the reason I can't. It's an interesting problem, though. My guess is you dig into that and you find there's some sort of scheduled pull request that's just like changing something that should be in the database, but it's not. And so here's the thing. Here's another twist. Okay. Okay. So the twist is we measure this two ways. We actually measure diffs per engineer, self -reported, meaning we just ask developers on average, how many PRS do you merge that you were the author of? Right. And we look at their actual GitHub data. And for this company, both numbers were like within point, like they were the same number. Which is remarkable in of itself. I mean, that sounds fishy right there. Like, what are the odds that they're like that close? Not exactly, but within like 0 .23. Yeah. I mean, why wouldn't it? Right. Well, I thought, okay, it all depends on how exact it is. You know, if you have a vote and you have 99 to one, you're like, okay, but if you have a vote and it's a hundred to zero, now you're like, there was some collusion here. Like something happened. Well, maybe people looked right. Maybe people looked at their own GitHub activity to answer the question. Because maybe they don't, they don't know. And so they're going to look at it. Yeah. I was saying if it was like exactly the same, then I'd be like, there's something wrong with our system here. Yeah, it wasn't exactly. So it was, it was, it was very close. And so that does exclude like bot authored pull requests, for example. And both measures, both the GitHub, the objective one and the self -reported explicitly exclude bot authored. I think you got someone in that org who just doesn't go to work and they just have a bot that uses their own SSH key and just does, does, you know, every day merges and stuff. And then you ask that person how much they emerged and they went and looked at their bot and they just guessed the right answer. Looking at outliers would be interesting though. The self -reported accounts for that because there's an upper bounds. Like the top option is actually, I think like 10, like 10 or more per week. You can't put in like, I merge a hundred per week. You can't be a self -reported 10 X dev. Yeah, exactly. Right. What an interesting problem though. Yeah, it was fascinating. What's up friends. I'm here in the breaks with Dennis Pilarinos, founder and CEO of Unblocked. Check them out at getunblocked .com. It's for all the hows, whys and WTFs. So Dennis, you know, we speak to developers who is Unblocked best for? Who needs to use it? Yeah. Unblocked helps large teams with old code bases understand why something has been done in the past. It helps them understand what happens if they make changes to it. Basically all the questions that you would typically ask a coworker, you no longer have to interrupt them. You don't have to wait for their response. You don't have to, if you're geographically distributed, you don't have to wait for that response. You don't have to wait for, you know, you don't have to dig through documentation. You don't have to try to find the answer in Confluence and Jira. What we basically do is give you the answer by you just asking a question. The way that we got to the problem was a consequence of our kind of lived experience. We're actually going to call the company Bothered, which is like, you don't bother me. I don't bother you, right? Instead of like being tapped on the shoulder or interruptive Slack messages, we could just use Bothered and get the answers that we wanted. We didn't go with that name because it's a little bit of a negative connotation, but helping developers get Unblocked by answering questions or surfacing data and discussions within the context of their IDE relative to the code that they're looking at is something that thousands of developers love so far. I think our listeners are really familiar with AI tooling. Very familiar with code generation, LLMs. How is Unblocked different from what else is out there? A lot of code generation tools help you write the code to solve a problem. We sit upstream of that. Our goal is to help provide the context that you need. If you think about where you spend your time when you're trying to solve a new problem, understanding how that system works, why it was built that way, what are the ramifications of changing something? That's the problem that Unblocked tries to solve for you. We take the data and discussions of all of the source code and all those systems to provide that answer for you so that you can get that context and then you can go and write that code. We have this great example of a company who hires very competent developers. It took them five days, that developer, five days to write 12 lines of code. And his feedback to us was, it's not that it takes you five days to write 12 lines of code, it took them five days to get the information that they needed to write those 12 lines of code. And then it takes probably about 30 minutes to write those 12 lines of code and rip off that PR. Okay, the next step to get Unblocked for you and your team is to go to getunblocked .com. Yourself, your team can now find the answer they need to get their jobs done and not have to bother anyone else on the team, take a meeting or waste any time whatsoever. Again, getunblocked .com, that's G -E -T -U -N -B -L -O -C -K -E -D .com and getunblocked. I was thinking though, as you guys were talking about this, that measuring speed is and it depends, right? Because like not every team can be measured speed -wise on the exact same metrics, which I think is why you have this key metric and then secondary metrics because you have the secondary metrics to sort of back up and correlate to what the key metric speaks of and the collection process is via systems. So collected data from a Git repo or other intelligence platforms and then self -reported. Yeah, that's a good suggestion. We didn't look at the secondary metrics. We got really trapped and like, why is this Dispar engineer inflated? Well, I almost wonder if the key metric is what swaps out because like on one team, the Dispar engineer may actually be the primary driver of the data you're trying to collect and a whole different team lead time or processes or deployment frequency is actually the better key metric and the others are the supporting. I don't know enough about your business how to do that. Yeah, this makes me want to go look at the perceived rate of delivery. Perceived speed is one of those secondary metrics. This is the analogy here would be for like aerobic athletes, right? Heart rate versus perceived rate of exertion, right? Those are the kind of two and like there's a lot of flaws in heart rate because I mean just the altitude, right? You could be training at different altitudes and the heart rate's different even though you're kind of doing the same load

  50. SPEAKER_01

    or

  51. SPEAKER_00

    you just wake up. You didn't get as much sleep. So your heart rate is more flexuring. So yeah, we really like that perceived rate of delivery. We literally just ask people to rate the speed at which their team delivers. Like one through 10, just rate it or? It's like a five -point scale. It's like extremely fast to slow. It's the actual speed rating. It's the words, not like a one through five thing. Yeah, very much inspired by perceived rate of exertion, which is on a 10, right? There's 10 options for perceived rate of exertion. Or perceived rate of pain. Have you ever seen that? For medical? I think they use that in healthcare. Yeah, there's a great Brian Reagan stand up where he talks about them asking him that question when he goes into the ER, you know, and him thinking through like what number should he say in order to get help as fast as possible. He's like, never pick seven, you know, like you're always an eight. And here's the full length unedited clip of Brian Reagan on this awesome bit. I was going to edit it, but I was thinking like, gosh, I would just edit this man's comedy and I just can't do that. So if you don't want to hear the whole thing, skip to the next chapter. Nurse finally comes in. How are you doing tonight? I'm

  52. SPEAKER_01

    on a journey. You have a painkiller or something? This is killing me. So she goes, how would you describe your pain? It's killing. She goes, how

  53. SPEAKER_00

    would you rate it on a scale of one to 10 with 10 being the worst? Well, you know, saying a low number isn't going to help you. Oh, I'm a two. Maybe the high ones. You could get me a baby aspirin and cut it in half. Maybe a Flintstone

  54. SPEAKER_01

    vitamin and I'll be out of your hair. You can go 10 to all the threes and fours and such. If anyone's saying such ridiculous numbers.

  55. SPEAKER_00

    I couldn't bring myself to say 10 though, because I had heard the worst pain a human can endure is getting the femur bone crack in half. And I don't know if that's true, but I thought if it is, they have exclusive rights to 10. And now I'm thinking, what was I worried about? Is there like a femur ward at the hospital? They would have heard about me and hobbled into my room.

  56. SPEAKER_01

    Nothing about 10. Give me a sledge hammer. Let me show you what 10 is all about. Mr. Tomie. Can I possibly say 10?

  57. SPEAKER_00

    I can't. So I thought I'll say nine. And then I thought, no childbirth. I better not try to compete with that. And then I'm thinking, you know, it must be hell giving childbirth when your femur bone is cracked. So I said, I guess I'm an eight. She goes, okay, I'll be back. I'm like, oh, I blew it. I ain't getting nothing with eight. But she surprised me. She comes in. She goes, the doctor told me to give you morphine immediately. I'm like, morphine? That's what they gave the guy in Saving Private Ryan right before he died. Okay, I'm a four. I'm a zero. I'm a negative 11 T. So they gave me morphine. Wow. All I know is about 15 minutes later, just for the hell of it, I was like, I'm an eight again.

  58. SPEAKER_01

    Guess who's an eight? And they finally checked me out. I'm walking out in the hall going, say eight, say eight, say eight, say eight. Happy eight day.

  59. SPEAKER_00

    And so it's very much Goodhart's law in a much funnier context. Yeah. Do you think, Abhi, that your North Star with DX as an organization, what you're trying to do is to define a path to happy developers? What do you think you're actually trying to accomplish? I mean, I know what you're doing as a result of giving survey results and this data, this formulaic and proprietary way to ask questions of an organization, how to disseminate this information and analyze it, that you're trying to help organizations be optimized. But like, do you think the true optimization factor is the path to have developers? Happy and productive, right? I mean, so that's the sack overflow survey once again confirms this, right? Because people aren't happy because they're unproductive is another way to characterize the findings, right? People are frustrated because it's hard to get work done because of their tools, systems, whatever. Therefore, they're unhappy and they're unproductive, right? There's a lot of time being wasted here. So no, I would say our North Star is helping every organization just become, every tech organization become the best version of themselves, right? I mean, that has different meaning to different people. But, you know, I mean, I'm the CEO of a company and we have engineers. And so the way I think about it is, you know, are we with the people we have, are we doing the best we absolutely could be? Are we as good as we could be? And, you know, we run the DXI and all these, the core four. And I'm looking at that like, how can we get better? You know, to another company that probably just translates to we spend a crap ton of money on engineers and we want to make sure we're maximizing that investment, right? Or it might mean, look, our competitors seem to be like creeping ahead of us. How do we go faster without just hiring more people? So lots of ways to tell the story in like a one liner. But yeah, it's about being good at building software. And as a result of that, people are also happier because all research repeatedly shows that happy developers are productive and productive developers are happy as the Stack Overflow survey also shows. I go back to the beginning of the conversation, which was 80 % are unhappy. And what we've failed to ask was why are the other 20 % happy? Because I feel like if your North Star is productivity, but that comes as a result, generally, in my opinion, and I don't know this qualitatively, is that you have productivity when you're happy and you can't create slash make happy developers unless you understand what makes them happy. So why is the 20 % of the 80 % that's not unhappy, happy? Like what is going on? Why are they happy? Yeah, I mean, whenever we look at this type of data, we're slicing and dicing. I mean, you do see some couple things I could share. We do see some differences across cross -culturally. So for example, we tend to see higher sentiment around this type of stuff, at least self -reported from populations in India. For example, we tend to see higher satisfaction with more junior developers, right? People who just don't have a frame of reference yet on what is good, right? They're new to the profession. So there's certain things that if we just looked at this data, it might be that the 20 % happy are coming from certain countries or certain levels of tenure and seniority. That could explain quite a bit of that 20. I mean, and some of them are probably legitimately in good situations with good developer experience and greenfield projects with no technical doubt where they're really happy. Yeah, I'm in full control. I have autonomy. No one yells at me. I'm getting paid. I'm not getting laid off. I'm not too old. I'm getting fired at 25 because I'm too old to code. That's the joke now. It's like you've aged out. I'm 25. No, come on. Well, there's another interesting data point on their survey, another interesting question, which is about coding outside of work. And if you want an indication of somebody doing something because it makes them happy, it's something that they would do outside of work. And so the same exact work of developing while there's 80 % unhappy at work, 68 % of respondents said that they write code outside of work as a hobby. That's like almost 70 out of 100 people. That's a large number and 40%, which there's some overlap there. These aren't mutually exclusive. Code outside of work for professional development or self -paced learning from online courses. So these are people investing in themselves, caring about getting better at what they do. And that's kind of amazing. So we have this dichotomy of people who love to write software, generally speaking, and yet unhappy writing software inside of their organization. And obviously you can look at your DXI and follow the 14. But the closer you can make your engineering teams feel like they're doing their hobby. Think about how a hobby works. It is self -directed, first of all. So autonomy is huge. Most likely, unless they have a bunch of kids running around, there's deep work involved. You can lose yourself in it. You can go into the, I was going to say the garage, but if we're coding, well, maybe the garage, wherever it is that you write software and just pound away at it for four hours without any interruptions and really lose yourself in it. A lot of these 14 metrics actually are manifest in hobbies. And so if you can, obviously a business is a business. And so you can't just be like, everybody do what they want. It worked for a little while for GitHub till they got to about a hundred, I think a hundred engineers. I was there for, I was there for the ride, not at GitHub, but here podcasting and paying attention and using it as a product and going to conferences where Zach Coleman was traveling around and talking about their engineering led development and everybody pretty much just works on what they want to, that worked for GitHub for a long time. Long meaning in years, not an employee count, like up to a hundred is not a large engineering team. They're way larger now, but at a certain point, that thing falls apart because there are, there's work that needs to be done that nobody would just naturally pick unless it was assigned to them and they're paid to do it. And so eventually that does, but if you can make your engineering team feel at least approximate, like they would be doing this as a hobby, then I think you're going to have a lot of happy programmers. That's how I've heard this described. You know, how can you kind of get the same feeling of joy and flow from that you do when you're working on a side project, right? How do you get that same experience while working in your job? How can you recreate that? If we can do that, we would unlock a lot more productivity. We would get a lot more out of engineers working at our companies. I think that's a good way to, good way to think about it. And a lot more happiness too, which everybody wins there. Like there's no losers. These drivers, these 14 drivers, have you ever done a survey where you've like asked developers to rank order? We do that. Drivers like they do. That's already, yeah, that's for every organization we work with that. So first we capture the data on it and then based on how they kind of stack up on each of the 14, we give them, hey, like based on here's how you answered these 14. Now out of these 14, what are the top one to three that would most benefit your productivity? Do you find that to be pretty subjective or are there certain ones that always float up to the top? There are definitely certain ones that tend to float more toward the top. Such as? Really? For sure. What's interesting is, you know, the Stack Overflow kind of technical debt is not one thing, right? Technical debt is actually like all 14 of these things, well, minus maybe two or three of them are actually types of technical debt. So, you know, documentation is actually a form of technical debt. Complex code is a form of technical debt. Slow CI CD is a form of technical debt. So all the technical factors do tend to float toward the top, actually. But some of the cultural, you know, cross -team collaboration, like delays due to different teams having to coordinate with one another is also, I'd say, a pretty common theme. Makes sense. Is

  60. SPEAKER_01

    that

  61. SPEAKER_00

    across engineering teams or product teams or is that like systems versus or ops versus devs? Yeah, good question. These are the people my response just now was deriving from how engineers report the friction. So from the perspective for developers, yeah, waiting on other teams, which could be cross functional or it could just be other engineering teams that they have, you know, different services or whatnot. So that tends to be a big area of friction. It's always about queues, right? CI CD is a queue, you know, being delayed or whatever is a queue. I can't work on this until you work on that. I can't work on that until you work on this. We can't deploy that because of this. It's all queues. Wait on code review. That's right. Then there's deep work. That's just meetings, right? Just like, hey, not just meetings, less meetings, please. No, not just meetings. What else? Yeah, it's people were asked. It can be people actually asking you questions. It can be code reviews. It could be, hey, can you do you know, can you fix this quick thing? Can you customer ask for this thing? Can you take care of it? It could be support. It could be incidents. So it's much more than meetings. Yeah, that's that's something we see a lot of companies just say, oh, yeah, we just need like no meetings Wednesday. And then somebody this problem solved. Right. And

  62. SPEAKER_01

    yeah,

  63. SPEAKER_00

    that's really the case. I was actually looking at a DX. What are top ranked areas where what matters most to your teams? Well, now, what does our yeah. What is our what do our developers say are the biggest areas that should actually be improved and press is actually that that code maintainability. So, you know, the ability, how easy it is to actually understand and modify code. It's actually also project clear project direction. So the projects they work on having clear goals and direction. And it's actually that batch size, which we've since renamed to incremental delivery. But, you know, are you working on kind of small, continuous changes as opposed to large ships? We're the top three for us. Well, those first two are driven by leadership, aren't they at me?

  64. SPEAKER_01

    I

  65. SPEAKER_00

    was thinking you're showing us cards here. I was like, dang. Yeah, yeah. Well, I'll say this. Our DXI score is three points above P90. So you're sitting pretty. Yeah, we're sitting really pretty. But yeah, even then, there's always room for improvement. Right. And actually, I just I'm looking at the data now and actually are clear direction. He's not upset. He's smiling. Yeah. Those top three I just mentioned actually are not above P90. Those three specifically. So you got some there's some room for improvement. Yeah. And same with code review. Turnaround actually is not above P90. Here's the better question, really. And I know you're poking fun, Jared. But I was. Yeah, and that's totally cool. And he likes it. You can't change what you don't measure, right? So now that you have this index and now that you have, you know, this awareness, even as a leader, you couldn't change it before if you didn't know it. But now you have awareness. Your team has awareness. Your team that is answering these questions feels heard, right? If you're going and making change and you say, hey, because of these results or because of these findings we're getting from our DXI score, we're improving these things in these ways. And the morale changes, the ability to speak to leadership and influence change changes, you know, all those things really come into play. Now, this gives me a lot of reassurance as a leader, actually, because I wasn't sure before we ran this last, we call them snapshots, right? This last kind of benchmarking survey and data collection exercise. I really wasn't sure. I was very pleasantly surprised by how good things are right now. I mean, that's what I would expect out of myself as a leader, right? But I wasn't sure. Am I just thinking we're good? And is it actually terrible working here or are we actually as efficient, you know, we actually kind of

  66. SPEAKER_01

    at

  67. SPEAKER_00

    that high level of efficiency that I would expect out of the way, you know, we do things here and, you know, we are pretty efficient. So it's good to see. Well, we all want to think that we're doing well in that which we set out to do well, but the worst place to be is to not be doing well and not know it, right? So at this point, of course, you are reassured because overall you're doing quite well, but even if you weren't, at least then you would know, okay, I thought I was doing well, but I obviously have some things to fix. Now, if we picked one of those three, let's not do a commit change size or whatever that one is. Let's go the other two and say, okay, these are room for improvement. So pick one of those two and just spit balling. Like, what could you, Abhinota, as a leader do to today, tomorrow in order to like meaningfully move that at your next snapshot? Do you have any ideas? Yeah. I mean, the project direction, you know, that's on me. And yeah, I mean, some of you were asking earlier like Pareto's principle, but like some of these are trade -offs, right? Because yeah, we can improve that, but that would involve a little more process, which would cost time and money. And given that it's already actually very good, you know, is that there's just like, it's more like something we want to keep in mind and be aware of so we can just lean in a little bit more there. The code maintainability, like that's already something we're really focused on. So that was like validating is, you know, that's something we need to continue focusing on is. And how do you focus on that? Like, what are your actual tactics? Good question. You know, having clear patterns, like, I mean, just really pretty strict code review and not just code review, but just making sure. I don't know if you've seen like Addy Osmani's post, like code is like a love letter to the next developer. It rings a bell. Yeah, it's so like that's in our like onboarding doc for engineering. It's like, look, the only thing that matters here when you write code

  68. SPEAKER_01

    is

  69. SPEAKER_00

    like, how easy is it for the other people on the team to understand that code? And we really try to make decisions on how we build things around that principle. So it's good to see that then reflected in the data, right? People are saying that it is easy for them to understand and modify other people's code here. So that's one of the ways. But yeah, it's a lot of hands on, like driving that principle forward, right? I've vetoed a lot of technical decisions, introducing new technology based on that principle that every new technology we add, every new pattern we add is something else someone else has to learn and is going to slow them down. I'm not sure where you scored on speed, but I assume it's pretty well considering these were the only three that weren't great. Have you ever considered compromising a little bit of speed? Like there's your trade off is like, let's slow down a little bit because a lot of times just time to breathe and refactor and maintain actually improves code maintainability. If you have maybe your snow leopard moment, for instance, I'm not saying do a feature freeze or anything, but like small bits. So get this. I'm looking at the data. So when I look at our core four, we are above industry P50 for speed, effectiveness and impact, but not quality.

  70. SPEAKER_01

    When

  71. SPEAKER_00

    I look at our second, some of the secondary metrics, so perceived speed and perceived quality, we are also above P90 on all of them except quality. So yes, I think we do sacrifice a little bit of quality for the sake of speed. And I mean, that's the data shows that very clearly. Now, whether that's a problem or not, that's an interesting question. Yeah, exactly. As a startup, I much rather be P90 plus on speed right now than because we have quality issues, but they're not they don't affect customers that much. That's the secret about our quality. Probably like we do have quality problems, but I actually have the principle on the team like, look, we're not building payroll software here. Like when we have a glitch in a report, it's not like it's not hugely disrupting or impacting our customers, businesses. And if we can quickly resolve it, then we're good. So that's a principle we have here. We have really fast recovery. But yeah, we do have not great quality, objectively speaking. Yeah, not abnormal for a company in your situation, I don't think. Do you have happy developers? It's a good question. So we don't measure for the reasons I kind of shared some concerns around measuring happy. We don't measure happiness. Do you have productive developers? Well, so yeah, we do based on I mean, our DXI score is through the charts, as I was saying, but we do I mentioned earlier, we do measure attrition risk. This is interesting. I haven't actually looked at this. Oh, this is a real time, real time demo. OK, I see. OK, I mean, this I can't share out loud, but OK, so for risk of attrition, we look at it very similar to like a blood test that you might get. So when you get a blood test, right, they tell you, like, here's the healthy range, right? Like, you know, if you're whatever blood pressure, cholesterol is within, you know, this nano milligrams to this, you're normal. So with attrition risk, the normal range and I may be having that coffee this morning, but I think it's seven to ten percent is the healthy range. So if your attrition risk is seven to ten percent of your organization is has signals of being at risk of attrition, that's normal. And I'll just say we're at the high end of the normal range, but I'm looking at the data. Our our reporting will break it, like tell you where that kind of risk of attrition. And so I'm looking at it now. It's I I'm aware of what

  72. SPEAKER_01

    is

  73. SPEAKER_00

    going on here. So it's good. He knows the inside story like he. Yeah. Well, how large is your team? I can't. It's around 20 people. Well, I mean, just roughly. Yeah, I'm not trying to drill down. My point is like smaller teams like these generalized aggregated numbers, you can see like, oh, well, there's a skew there because of this one person situation or whatever it is. Yeah. No, this is meaningful, though. No, this is actually this. I was actually worried about some flight risk

  74. SPEAKER_01

    in

  75. SPEAKER_00

    this area on this team. And this data is actually telling me feeling better. I'm feeling better about our product because today we've gotten to go through the all the data. And it's been really, you know, to be honest with you, when I originally got the data, I was so busy, I didn't like fully go through it. But like you guys asking these probing questions. I'm having fun. Thanks for going through that with us. So, yeah, 10 percent risk of attrition. Well, we'll we'll play with you every snapshot. We'll give you a overall score, the changelog score for your business. Yeah, this is great. I do enjoy this. I think that what I kind of find fascinating is you mentioned the year 2020. And you were at GitHub. And I don't know if you know what year it is, but I do. OK, it's 2024, just in case you weren't aware. This is four years later and you're this far with DX. Congratulations. I mean, like, thank you. You are we've asked you hard questions. You've shared some insider information that is in this report that only probable you and some others get to see these snapshots. Yeah.

  76. SPEAKER_00

    So kudos to you for being forthcoming on that. We've talked through DXI. We've talked to the core for we've asked you a lot of hard questions. And you're you're like only a few years into this. And you're this steeped in, I would say, Trajection and maturity like you've got a strong team operating at speed, not the highest quality, but you're you understand where the lack of quality is and why it's OK to have that. And you have a pretty good foundation and some assurance personally, it seems, on how to take action when you need to take action. That's a great place to be in. Yeah. Hopefully, you know, I mean, the way I look at this as a startup is can we try to not end up like GitHub? Right. That's the rule. Oh, well, well, same. Same on their edge and velocity specifically. Right. I mean, like they they were at a point where they weren't shipping and people were leaving because it was so hard to ship. So like, you know, honestly, a lot of quite a few companies we work with are kind of in that position, like these big tech companies that that went through that hyper growth and churn and, you know, tons of reorgs and are now confronting, OK, now we can't just keep hiring people, but we need to be shipping faster. Where do we get? What are the levers we can pull? And so, you know, it's kind of like health, right? Like if you can get ahead of this stuff and not have four decades of poor diet and exercise that hits you in your fifties, like if we can kind of stay ahead of it, you know, I would hope that we're able to scale the business while staying p90 on velocity. That that's the goal here right now. Would you be like GitHub in so far as they took a seven point five billion dollar pay out? I would take that. I would take that. Yeah, I would take a few billion dollars. I was going to say that in that way, I guess it'd be OK. Yeah. Well positioned, I would say. I don't know who would acquire you. Like who cares about what you care about to the point where you're an acquisition target? I mean, anyone in the developer business, I would say so. GitHub, GitHub, Microsoft, Amazon, Google. Right. I mean, all anyone in the cloud game cares a lot about like benchmarking, assessing maturity. Right. Yeah. I mean, even Salesforce is a little bit. Could you IPO? We could. I don't. What do you want to do? I don't know. I'll do whatever, you know, fate has in store for us. I mean, that's what I you know, we we've pretty much bootstrapped the whole thing. So we and we kind of control our own destiny. We don't have to

  77. SPEAKER_01

    have

  78. SPEAKER_00

    an IPO. We don't have to sell for eight billion dollars. We just want to keep you know, I'm just drawn to this problem. And I think I shared this the last time I was on the show. Yeah, this all started seven years ago when I first became the CTO at a startup and the CEO asked me, hey, I'll be all the other departments here are reporting metrics each month. Can you start reporting some metrics on the productivity of your engineering team? That was seven years ago. And so I joke with people. I'm still just trying to answer that question because I couldn't answer that question seven eight years ago. Right. I asked other CEOs, what are you reporting? And got 20 different answers. And it was a hard problem. Yeah, for sure. There's a Silicon Valley episode about that close to the end of the final season. Kind of funny. And you're a I think this is the other show we did together that deep dive. I think that you're the only owner of the business that you're the solo founder. I'm the majority. I have a co -founder, but yeah, I'm the majority owner. Yeah, I thought it was singular owner. My knowledge is the time between the last time I had this conversation is diffing on me. But no, it might have been my previous business, the pull panda. That was a singular owner of that. But now this one, I've had a business partner since the beginning. And you've taken the venture capital. You said bootstrap. Not venture capital. Just we took a little bit of angel investment at the very beginning from friends and family. Yeah, but we never spent that. So when I hear bootstrap, that means that your gains today cash flow positive or at least reinvested. If not break even or in the negative, you're in the negative potentially because you're reinvesting, not because you're losing money. Yeah. Now we're extremely cash flow positive. And to the point where my biggest concern each year is how to spend some of that so we don't pay taxes. Corporate 20 % tax rate because that just goes in the chest. So yeah, I try to reinvest it. It's better to reinvest it than lose 20%. Do you do much advertising? I was going to say, we can invest some of that for you. I mean, there's lots of opportunity. Let's just say lots of opportunity. Let's take that offline, Adam. Yeah,

  79. SPEAKER_01

    I'm

  80. SPEAKER_00

    only kidding with you. Our listeners want to hear about how we get new sponsors and new partners,

  81. SPEAKER_01

    but

  82. SPEAKER_00

    they kind of do. No, they want to hear that. Yeah. Yeah, they'll eventually hear. We won't let them though. No, this is fun. I've been digging into this. I think that we want developers to be more happy. Obviously. I think the question to me is how, what makes developers more happy? I think productivity is obviously one key metric and maybe some secondary metrics could be what? I don't know. Just a happiness in life potentially. Other things that influence happiness. Pay, perks. Pay, well, you know. Free beer and ping pong. Ping pong tables. Definitely for a certain demographic. RTO? No. Right. I do think that. That's my new tagline, Jared. Not rug pool, not cool. It's RTO. No. It's a good one. I was going to say, I just was going to say, I do think that like freedom to live your life in a way that suits you and still work is a huge driver for a lot of people. More, more than money, probably right up there with like productivity and enjoyment of my work is like, do I also get to live my life in a way that suits

  83. SPEAKER_01

    me?

  84. SPEAKER_00

    But that's maybe just me projecting. Cause it's always been my primary driver, even more so than money is freedom. And I very much enjoyed it for many years. So I'm appreciative that I have it. So maybe I overemphasize that, but I'm sure there's a survey out there that answers that somewhere. The stack overflow or next year's DX. What are you guys going to call this thing? Your public thing? Probably state of. State of DX. DX or developer productivity. We kind of use those terms interchangeably. Careful now, cause there is a state of organization run by a friend of ours who does state of JS, state of HTML, state of CSS, and they have this whole platform called state of. Yeah. We could just call it developer experience index. You can also team up with them and have them help you run it or something. And then you could, I'm sure they'd be happy to cause they, they did create, although you have all your own software, so maybe it would be a square peg round hole, but I know they did have opened it up and like Google runs the state of HTML survey with them. So like if you wanted to really use state of, I think there's definitely opportunity there. Anyways, in the weeds again, let's call it a show. What do you think, Adam? Yeah, I'm down. Obby. Thanks, man. It's been fun, Obby. Thanks for joining us, man. It's

  85. SPEAKER_00

    been cool. Thanks for the invite. All right. Bye friends. Bye friends. So RTO. No. Are you an RTO person? Are you being forced back to the office? Say no. I'm just kidding. Maybe you can't say no. It's a hard thing because now we're in this world where we were once given this, Hey, work anywhere. Hey, be remote. Hey, do whatever freedom. And the kind of jobs that we do in tech generally are jobs we can do remotely. We can do pretty much from anywhere. We can be nomadic and tour the world and have fun and enjoy our life or optimize for where we want to be in our life. And that makes us happy. But RTO is a thing. I say RTO, no. If I had to RTO and I couldn't say no, man, I'd be pretty sad. And if that's you, I feel for you. There's a place you can hang though. It's called Change Law Community. We have a full blown, fully open, no imposters. Everyone is welcome. Zulip instance or replacing Slack with Zulip. You can go to changelog .com slash community. Sign up. Everyone is welcome. Come in there, hang out with us and call it your home. Hang your hat and you are welcome. And I want to see you there. I also want to see you at all things open. So speaking of RTO, ATO, but that's a good one. All things open .org. We love this conference. We go every single year. We will be at booth 66 right by the ballroom. You will see us there podcasting with everyone we possibly can. Come by and say hi. Hang out with us. High fives, handshakes. And as you know, the occasional hug if necessary. And we can give you potentially a free ticket. Come hang out in Zulip or we can give you at least a 20 % discount that's available to everyone. Use the code media changelog 20 details are in the show notes. Camel case media and changelog and then add 20 at the end. No spaces. There you go. The link is in the show notes. Follow that. That's the best thing. And I want to see you there. Okay. Plus plus subscribers. We've got a bonus for you on this episode. If you are not a plus plus subscriber, go to changelog .com slash plus plus. It's better. OMG, it is better. I can't tell you why. You just have to find out for yourself. Go to changelog .com slash plus plus. Drop the ads. Get closer to that cool changelog metal. Get bonus content like today. Free stickers mailed directly to you. And the warm and fuzzies. Who doesn't want warm and fuzzies? I know I do. I like those things. Okay. Thanks to our sponsors. Sentry, Fly, Coder, Unblocked. Wow. A lineup of awesome sponsors. Sentry .io, Fly .io, Coder .com, and getunblocked .com. They love us. Go give them some love. And that supports us. And I appreciate that. Okay, BMC. Thanks for those beats. You are awesome. That is it. Friends is over. We're back again on Monday.