even have as he mentioned one of the copies out here which uh we have three copies they don't they're not this is the first time I saw it yesterday actually and I don't even get to keep one of the copies myself I'm giving them all out today so uh I won't actually get my copy until sometime later so it's exciting for me as well because it's the first time I've actually seen the thing in paper so good day and welcome my name is Brendan I'm here to talk about the systems performance book and and what is systems performance it is the analysis of everything application is down to metal and think lamp not amp so we're focusing focusing on the operating system but as well as the applications the basis is the the system but the target is everything as all software can cause performance issues and for many of them I need to go from applications go through libraries in the system call interface go into various parts of the kernel to really root cause what's going on and so that's systems performance it's a discipline that's been around for a long time there's been uh previous books on it a while ago and I'm really happy to launch system performance Enterprise in the cloud um this was written by myself and there were many many contributors who I'm very grateful for and I think some of them are in this room as well uh 635 Pages a big goal of mine was to make it as short as possible so that I don't waste your time as you as you read it and so a lot of hours was was spent condensing it down and condensing it down to figure out just enough what you need to to know to go and solve problems I've certainly I mean my publisher knows what I can what I can get through so my previous book was D training soured by prin Hall that's 1,152 pages so uh and funny enough it's about the same size so uh they printed dra on very thin paper so that you can carry it around printed systems performance on normal paper 635 Pages class peneses it covers background methodologies and examples and it does differ from PR previous books on systems performance I'll get into why the examples I lead with Linux and so I used a few different distributions Ubuntu Fedora and Centos and I also use Al lumos smos and Omni OS for examples it's actually great to cover two different operating systems because it gives you an additional perspective to understand the design choices of each OS and of each kernel and so when you see if I'm just focused on one operating system and you see various things are done um some sometimes you can forget to question why things were done and looking at two operating systems for comparison is great it really deepens your knowledge even if you only care about Linux or if you only care about Al Lumas you'll actually benefit from understanding how other operating systems do it because it will help you understand and your colel better uh the audience system administrators developers and everyone Enterprise and Cloud environments uh just briefly about myself because uh who actually wrote the book uh I'm currently at joint I was previously son and Oracle uh I do debugging of performance every day and so I'm going from any layer of the software stack down to to metal sometimes firmware which is which is really difficult and and annoying because I don't have Dynamic tracing down there uh previously I have worked as a kernel engineer I did the development work for the ZF Al AR um and I've also worked as a performance consultant and trainer I've written hundreds of published Performance Tools probably two many I I think to just I I feel a little guilty that I've contributed to the problem of having to go and learn all these tools uh I've created visualizations so heat maps for various uses and I I think think we're here in the office and and some of my visualizations I pinned on the wall as I was working on issues and some of them are here because they're pretty um you can see some of the visualizations I've created uh actually around me uh develop methodologies the use method and the TSA method which I was publishing this year and I've books per so the goals of the system performance book our primary goal for it was to be modern and so the last real system performance book I don't know if this is familiar at all but SL performance and TOS published in 2006 I was a co-author and we really went through systems performance in a lot of detail this was originally developed as part of sence internal second edition but second edition got so big that the publisher was worried about how to bind the book with so many pages and because at that point it was over 1,500 pages and so we split out the performance and tools section into a companion volume it worked really really well uh because this this is the volume you could carry around um for PR actually working on issues as a practitioner then you have the SCE and to's volume for reference but that was 2006 and we haven't had a it hasn't been a a a real systems performance book published for several years and because of that we've missed out on a lot of detail on cloud computing uh Dynamic tracing new visualizations that we we're using these days and also open source has really changed things and I'll get into more about that in a moment another goal was to be accessible to a wide audience because I want people to really read the book this is not just about if performance is your day job it's also for casual users where I I just want to go to reference because I have a problem I'm going to spend half a day on it and that's it help you maximize system and app PA performance obviously uh to quickly diagnose performance issues uh for example latency outliers and I and I've read this separately because with system performance it's not just about making systems go go 5% faster or 10% faster or 50% faster sometimes it's about the latency outlier that's taking 1 second or 10 seconds if I'm a big environment I care a lot about the 5 and 10% but I know there are many little environments who just don't care I'm A Cloud shop and we have three instances I don't care about 5% like I can just buy a fourth instance but I do care about latency outlines because that means my customers are unhappy and so this is about both and it's different audiences who really care about each another goal was to turn unknown unknowns into known unknowns because they're actionable and so with system performance there's it's there's a lot of detail there's a lot of complexity and when people are working on issues you work in the boundaries of your own knowledge and if you just don't know that areas of the colel exist or the various procedures are possible you're not going to use them so part of the goal of the book is to is to expose you to all sorts of areas just so that you know that they exist you may not use them straight away but sometime later when you have that really urgent customer critical issue you know I've seen that before there's I think it was in chapter six of that performance book let's go have a look and so remember as you're browsing this because does get into a lot of detail that a key goal is just so that you're aware of things that exist so you can look them up when you need to and a shelf life uh working with one of the one of the attributes of a paper medium this is also an ebook is that this is going to be on people's shows for a long time and I've been teaching system performance for a long time since 2003 and uh I used to used to be an instructor uh especially for some Microsystems during classes in system administration system performance uh students would expect the instructor to have read every single textbook that existed on the topic um and I was already a bit of a book nerd so I I ended up reading everything because the instructor can make recommendations to class it's like there's so many books on this topic what should I go and read and so there were various books I used to recommend people to read including this one the one you might be familiar with that's Asian cock's uh book is that a Ferrari or Porsche I keep Porche Porsche Porche okay um so obviously I'm not as much of a carard as as Adrian is but uh and I'd recommend this book to students even though it's getting older and older and older by the time I was teaching in 2004 and 2005 and uh especially if you open it up and it and it begins by giving you this particular Edition Begins by giving you some very specific recommendations and so one of my students tells me told me that the adage never buy a performance book that's more than 2 years old now please don't tell my publisher that because he would never publish a book again because they actually need a long shelf life for the publisher to actually take this on board but and I recommended to the student I said no I'm reading this book and this particular Edition at the time was more than 5 years old and it's very valuable because performance is a complicated thing and there there isn't that many times when an expert sits down and writes out all the methodologies that they do all the approaches that they do all the the nuances that they've learned their experiences it's really valuable even if if it's 5 years old even if it's 10 years old there's a book I should have brought I should have brought it in by my backpack would have been too heavy there's a performance book I I still refer to now it's Raj James the art of computer performance analysis printed in 1991 fantastic book I'm referring to it now 2013 and so that book has more than 20 years shelf life but certainly hearing my student tell me never by by performance book more than 2 years old it's got me thinking how can we do this how can we create a a book that's that's very useful for people in the longer term and so that was one of the goals of the system performance book and there is a way to do it which which I'll get into although I still think that as I told the student I don't think that was a fair fair thing to say because I think Adrian's old books uh there's also system performance shoting by Michael ktis from O'Reilly also another very useful book so I do branze a lot of these old books with the methodologies and the ideas from the performance experts but from the onset my goal was to make this book actually have a long shelf line my personal motivations I did did need a good reference for internal joint staff external customers and it at large uh frequently I'm explaining to people with their inside or outside of joint how to Benchmark things or how to diagnose things and it's been very handy for my previous books the dce book and last performance and tools to say well I've written this down it's documented here and I really needed a modern reference for everything we're doing in systems performance these days also as a reference for classes I I mentioned earlier I have been teaching for a long time and I learn a lot from teaching classes I really love it in fact this is something my publisher recommends to anyone trying to write a book is try teaching this to other people because it's harder than it sounds and when I first started teaching system administration for sun I would say things that I would understand of the colonel and as the words would emit out of my mouth I'd realize how ridiculous they were and how unresearched and and SE the pants these things where it's like ah what am I saying I need to look that up and understand it properly and of course I've just said it and the students are like hey you just said this that doesn't seem right so actually teaching and like announce like going through your knowledge and like saying it out loud really helps you paper over any gaps it makes you go and research things because they just sound silly when you say them and and gives you a better understanding of a lot of things so I learned a lot from that I also learned a lot from the from the students U the questions they ask and also a a type of performance course I like to teach and and I'm doing this for joyant now in fact I have one coming up November 18th on cloud performance and this is where I set up labs for the students to solve and I don't give you the answers and so I'll simulate a performance problem and say tell me when you fixed it and actually create a simulated performance ticket which says the customer reports that performance is bad for these reasons when you run some binary and then you have to go and debug it it's really interesting because I would think I would think that well I I definitely taught the students how to solve this particular issue and then everyone would be stuck like okay well I need to change how I'm teaching this and I need to explain various things and it was really important for me uh to this is really led into the methodologies that I've been documenting recently because I got to test them out in classes I'm explaining to new people and to see if they can then pick it up and actually apply it to solve problems so I'm I'm really lucky that I get to do a lot of real world things that joyant we I I didn't mention who who we are but in case joint you do you we are a cloud provider we have uh smart OS which is an OS virtual OS virtualize virtualization for cloud instances high performance we also have Kum for Linux so I get to do performance for a wide variety of things but I also get to do research and teaching helps me really um study systems performance also try it out on people and see if I can help other people learn so in the book the table of contents 13 chapters uh and I'll go through some of the highlights uh these sort chapters probably are recognizable application CPU memory file systems discs and network uh that's what a a formula that has appeared in previous uh systems performance textbooks as far as I know it may have come up the first time it may have been done was Mike marus and his original system performance tuning in a Ry book and since then everyone has has used that resource orientation oriented division for working through systems performance which works fine um I also have a chapter on cloud computing but I do mention it where appropriate everywhere a chapter on benchmarking and a detailed case study but there are others as well U appendices including some some uh my use method checklist for limits and solars and of course the uh this is 635 pages and then we've got 100 plus pages of extra content so some highlights chapter 2 methodologies uh this is what this is what I think people will find is really new and interesting from the book methodologies have been described as a as a as a black art that is passed from senior C admin to senior C admin and and it's I I remember being a junior uh system administrator and wondering how to get started and running tools like vmat and being lost it's like what do all these things mean what what what a page in and page out I don't even know what a don't really have a clear idea of what a page is littleone scan rates and deficit and all of these columns so um helping demystify performance and explaining because running these tools is often jumping the gun and you're going straight to the answers and helping you understand the scope of what we're trying to accomplish the concepts the background the methodologies and then you start using tools and you start solving problems operating systems something I found teaching classes is that uh some people Miss operating system class at University they've never had that sort of education and there are lots of really good books on operating system internals and so like here in the office if you look around on people's desks you'll see plenty of copies of Solus internals but the problem with Solus internals is it's a great book I love it I I helped edit the second edition it's big it's around th pages and if if you're a a CIS admin or doing devops or development it's hard to find the time to read through all of that um especially if there's a performance crisis and you need you kind of need that background right now and so I took took it on board to say how how much could I summarize in 30 pages this is a 30 page crash course into operating system internals and it's kind of something everyone in the industry should read and I complished it I'm really proud of it actually really nice 30 page summary U and of course I've worked on operating system Internal Documentation before with stus internals um this is written in a generic fashion although I begin by doing Unix in general and then talking about differences between l and the Solaris family of operating systems uh chapter 6 to 10 is the bulk of the the book which is the division of resources CPU memory call systems distance and network chapter 11 on cloud computing um and chapter 12 benchmarking benchmarking it's the chapter that had to had to happen I I wrote it for the good of the industry and I really need everyone to read it I've been doing benchmarking now for uh for about I really really got into it in 2008 and I've been doing some things that's that's 5 years and about I I I frequently get handed benchmarks in fact doing performance at joint since we're a cloud computing company when customers try us out often the first thing they do is they run a benchmark and if it doesn't look good they then say to whoever their rep is it doesn't look good we're not going to go with you and that often gets escalated to me or one of the other engineers in the company so I'm frequently spending my life looking at OS benchmarking and almost 100% of the time it's done incorrectly and so the numbers the numbers are either completely bogus or they have been misinterpreted because it is actually a really complicated thing to do and a really complicated thing to get right so I uh I think Adrian uh Adrian cockro made some suggestions I think he wanted me to put that chapter earlier but um it's actually quite a dark and depressing chapter because I have to drag you through all the all the problems that you may run into and I was a bit nervous about putting it earlier and and having you like shut the book and think I'm not going to read anymore this is just too depressing um but it is a chapter I want people to read this it's yes it is the most error prone thing you will do in Computing is when you try and evaluate performance and it's because if like just to give an example if I gave you a thermometer and said check your temperature you said radio you stick it in and pull it out and it says your temperature is 1,000° f you'd say well that's ridiculous that that can't be 1,000° fah that doesn't make any sense the thermometer is clearly broken however with performance people often grab a tool that've never heard of before it gives them some number and they just accept it and it's because they're new to it I I understand we're not born knowing how to do this stuff and so these numbers have you don't have any set of expectations for what for what fio or CIS bench or bunny Plus+ or ier or any of these tools should give you and you run them and just accept them so I'm frequently sent these Benchmark results where people are saying oh yeah my temperature is 1000° f it's like w something's really gone wrong here um but yeah that's that's just one of the the the things that makes it difficult for people is that you just m to it and and you don't have any set of expectations to refer uh so in a little bit more detail uh and I've got a I've got a PDF of the book I can browse quickly chapter 2 methodologies so this as I said it's documenting the black Artic system performance and where possible I do flowcharts and um give you ways of extrapolating and predicting performance the capacity planning or whatever lots and lots of methodologies this also summarizes Concepts that you need statistics and visualizations like line graphs and Scatter Plots and heat Maps um of which which there are many visualizations elsewhere in the book as well chapter 3 operating systems I mentioned that that was the the operating system crash course you missed at University and some really basic things like Fork creates processes usually um the process life cycle or the the it's really the thread State transition life cycle um understanding the operating system stack going through what's inside a process it's just the 30 pages of background you need need to um work on issues and and of course U part of the motivation of writing this is when I when I've been teaching systems performance classes and and students start to use tools like s trace or D trce or trust and get immediately lost because they don't have that background of understanding what system gos do and understanding what the colel does so I don't I don't to say You must read read a a Linux internals book or a slce internals book so that's why I came up with 30 page summary chapters 6 to 10 uh the structure for these is so this is the the bulk of the book where it's CPUs memory file systems discs and network I will begin with background I'll begin with terminology actually because part of the a big part of the problem is learning how to talk the talk let's go through the terms and terminology can change from from Context to context so explain how things sound how we use the terms then go through various Concepts and give you enough operating system and hardware and terns to work in that area then go into methodologies of how you apply that background knowledge so for beginners this these methodologies are forever beginners casual users and experts it's how to start and then the steps to proceed and like I said I think this is going to be the biggest surprise of the book is that I've really spent a lot of effort documenting how to start first time because when I when I teach people either I'm mentoring them at work or I'm teaching them in a classroom it's the biggest problem there's so many tools where do I start how do how do I begin on this particular perform issue and then I'll get into the example application of everything you've learned so how it works on Linux and lumos tools screenshots and case studies and then some tunables of the day for to to Really uh see the sort of ways that this works and this structure is very deliberately generic to specific and this is part of why the book is designed to have a long shelf life because this stuff won't change for 10 years 20 years 30 years it's the example application that will change and so tools will change the tunables which I put right at the end I mean there's not a lot of tunables in the book this is like putting a paper book full of tunables is just a bad idea because it it took 5 months to go from the draft and then all the procedures of uh copy edit proofs indexing um to the copy I have now and and chines can be out of date in 5 months so this is right at the end of the chapter to say this is kind of the areas you need to go and research just to like you might might not know that diss have controllers that you can log in and tune if you have the right tools etc etc the sort of things you should look for but don't copy the tunable parameters you're going to have to research your environment and then find out what the point in time recommendations of the day are so that's deliberate uh starting off generic and then getting more and more specific as the chapter goes on chapter 6 CPUs this is actually has been released today is a sample chapter uh I put the URL at the end and I'll just browse some of it I've actually got um a my author's review copy of the the book and so it's actually just to to browse some of some methodology which I think is going to be is going to surprise a lot of people talking about terminology talking about Concepts how L what ly actually is trade-offs and performance point in time recommendations and then getting into everything you need to know and then getting into this this is the methodologies so generics system performance methodologies and actually go through some anti methodologies for reference the street light anti- method the random change anti- method the blame someone else anti- method which a lot of people tend to do um adhoc checklist method these are there for comparison and then get into the the more productive ones the problem statement method the scientific method the diagnosis diagnosis cycle from Brian Cel who documented this a long time ago the tools method used method which I came up with and um have had that published for a while workload characterization drill down analysis lengthy analysis uh mentioned method AR is great for comparison event tracing Baseline statistics performance monitoring some queuing Theory there are fantastic books on queing Theory U but enough so that you understand that is turning one of the unknowns into known unknowns so that you know if the time comes um the sort of things it can help you with uh static Performance Tuning cash tuning micro benchmarking and capacity planning and that's just the generic ones each other chapter has more specific ones for their area like CPUs and memory and discs and network and file systems uh but I wanted to browse chapter six briefly because this is the chapter six sample chapter anyway and so as I as I just browse this quickly you'll see the pro progression from generic down into specific and so starting off the terminology and there was a nice glossery at the end but just explaining what what does a logical CPU actually mean or a virtual processor or Hardware thread um just so that we understand what we're talking about and that run um dispatch que is is what it's often called in the world of solarus so at that point you can then read the text more easily some some models of how things work going through various concepts for um for the C chapter it's actually nice because I for many of the chapters I'll go through all the basics to get to take you to the point where you have just a crash course on that topic and then you can apply performance uh analysis to it so starting off with so what is clock rate CPUs have clocks how do instructions work um they processed as an by a series of functional units a fetch decode execute memory access register right back um why do we have instruction pipelines increasing throughput instruction width so nowadays we're getting up to four wide processes over three wide we go even faster cycles per instruction or instructions per cycle really important metric which now that I've explained those Concepts you should be able to understand what that's what that's about and that's a really great performance metric which is it's just a number it's showing cycles per instruction is how many CPU Cycles it took to execute an instruction on average the higher it is the more memory IO or bus IO your computer is doing the lower it is the more quickly it's going through instructions and there's a dark area of computers that we're not very good at observing and that is memory main memory so what is main memory doing we know how full it is that's capacity what's the memory buses doing uh especially these days almost everything is a is a numer environment and memory locality can be a big difference and big performance difference and when that's not working so well and you're having to hop over CPU interconnects to access remote Banks of main memory the number of Cycles it takes to work on memory can go up and up and up that can really start to be a performance problem just I talk about all this in the book um CPA is like the first indicator that you may have that type of issue and it leads to actionable items like I should be changing my software to do zero copy or I should be looking at how I can configure memory locality and binding processes to CPUs so that they they don't walk around or change changing what Hardware I'm buying from time in time so more Concepts explain utilization uh use time incal time saturation preemption priority inversion and for some reason I I got excited and did the entire let's talk about priority inversion and exactly how it works I think I was that excited because we worked on a priority inversion issue recently which uh which Brin solved is is actually quite nasty it's and that's where threads can explain really nasty performance issue where the priorities can end up causing a foam tissue of the blocks that they're holding those are working so well multiprocess and multi- threading so lots of Concepts word size compiler op optimization then we've gotten past generic Concepts I'll get into architecture so um Hardware just a generic picture of a two core processor and then talking about all sorts of other stuff that you get like the preet cach and the right cache micro code ROM temperature sensors talking about caches I know these are old fashioned but they fit on the figure better the E doll I dollar D dollar with your level one level two cach and so on um look at examples this is memory access latency increasing the memory range um this is a really nice and I used LM bench this this is a really nice this is one of the most beautiful benchmarks I've ever done uh and that's where LM bench has looked at uh memory access latency and it is stepped up the size range so from 1 to about 64 kiloby my memory access latency was very small about 1 nond and then we step up up to whatever that is 256 KOB and then we step up to about a megabyte and then we Step Up further and can we guess what these steps are these steps are going to be working big bigger and bigger caches as you're going as pulling out of the your level one data cache and then you're going to the level two cache level three cache and then main memory so main memory will be up here and we can see the caches as we go back down to uh the level one cach this will be what's so beautiful about this these are by the way these are both log that's yeah these are both logarithmic axes what's beautiful is how even these steps are and that's the whole point of having these multiple stages of cash caches I think I was do is on an Intel machine and it's uh the the processor I think the processor designers have been looking at the same thing and like we need to make the size of the cache and the latency so that these steps are evenly spread so that it has the most benefit for workloads of varing sizes yeah I got really excited by this it's uh worked out so well um so yeah architecture talk about how that stuff works how The mmu Works interconnect um secret performance counters is a really fantastic but underutilized uh framework for understanding stall Cycles how CPU instructions are really working how they're being executed understanding level one two and three cache HS and misses floating Point operations memory IO and resource IO these are various instrumental counters on processors that um in Linux is the peral in um smart OS CPU stat from REM Space Systems give you an idea I mean this is here in the the architecture of the hardware section because they are Hardware talking about all of that giving some examples of performance counters then getting into the software background so how the kernel does scheduling run cues um how we're doing voluntary contact switches and involuntary contact switches of course in Linux they they aren't always exactly a queue you can have trees for doing uh thread scheduling but it's or task scheduling but similar Concepts talking about starting to talk about more specific differences so Linux and Solaris um thread schedule priorities different thread scheduling classes you'll encounter and why they exist and at that point I've I've introduced I've given you a cred course in the con the terminology the concepts the uh Hardware of CPUs and then the software of CPUs and have the kernel are scheduling decisions in different scheduling classes now I can talk about doing something with it and so now I'll go through methodologies and this is the methodologies for the CPU chapter tools method use method workload characterization profiling cycle analysis and so on uh and so tools method is way you just work through the you have a checklist of tools you work through them and then pick the metrics of use and it's kind of what people do for simple and then more effective methodologies like the use method utilization saturation and errors doing that for every uh resource but in here I'm talking about CPUs rep characterization um profiling some examples cycle analysis performance monitoring static Performance Tuning anyone know what static Performance Tuning is St Performance Tuning is great it was uh I think Richard Ellen popularized it a long time ago in a in a blueprint it is you it is where you're looking at not the dynamic performance of your system when a workload is applied but it's when you look at how is the system configured so if you turn off the workload and just look at okay how many CPUs do you have uh are they course Hardware threads are you single or multi processor what's the size of the CP caches what's the CP clock speed is it Dynamic do you have interal Turbo what CPU related features are enabled or disabled in the bias uh are there performance issues with this processor model yeah look at the Arata sheet of the Go download that from Intel are there performance issues bugs with the BIOS firmware version are there software imposed CPU usage limits resource controls present like you have in CL computer what other they so none of these are related to the performance grow that that's occurring on the system the workload this is just all about how it's going figured and how was set up and it solves so many performance issues because people just Overlook the configuration and they start getting in stuck into the how the workload is performing and its performance and then it turns out yeah you're running on the older system why are you running on that system why you running on the newest systems and faster CPUs time I forgot to check I forgot to check that I wasn't on the new system or I forgot to check that we still have those CPU resource controls turned on and that's throttling our performance that that method ology it's great I use it all the time static Performance Tuning uh priority tuning actually priority tuning is one of the that's existed almost forever Unix is uh for Unix it has existed forever it's always provided a nice system call for adjusting process priority the main page from Unix fourth edition says a value of 16 is recommended to users who wish to execute long running programs without flat from the administration so uh that's always existed and that's a way I can say these jobs should interfere with the high priority jobs so these can run at a higher nice level which means a lower priority resource controls which I'll talk more about in other chapters CPU binding um micro benchmarking CP binding did this um only this week where it was binding a particular customer issue where they had uh negative scalability and this is where the more CPUs you you give them the worse the performance is and it's because it's it's in lock contention hell and they're making such awful forward progress that even though this is actually a pona even though it's a a large multi-threaded application I thought why don't we just bind it to one CP and then take out this lock contention hell and see how it performs and I know that we're not going to use the seven other CPS that are there but let's just B it to one it ran 28 times faster when you took the application and just found it to Once in now of course they wanted to actually fix the lock in tension instead of this very bizarre work workaround but still I list it as a methodology even if you go with it or if you don't go with it that's extra information to help you understand isue micro benchmarking always useful another way to it's not a now I call these experimental methodologies in previous ones like static performance chaining is an observ AAL methodology so microbenchmarking is an experimental methodology where I can um run various things to determine how how fast the CPUs are how fast memory access is like I did with LM bench and so on then I get into analysis so at this point in in these big chapters I've gone through all the background you need I've talked about the methodologies so that you understand what in general you're trying to accomplish and now I'm going to talk about what tools can you use to accomplish that uh and this is again to to make this a a a long-term reference uh and even though the tools will change I'll have taught you how to find the tools and what you're trying to do with them so now it's here's a selection of example tools so that you can accomplish the previous tasks um and then a lot of these you'd expect to see and then some of them you may not up time up time out the load averages um here I've printed out the I plotted the 15 and 15 M load averages I've created a single hot thread and then at the there's the 1 minute Mark the 5 minute Mark and the 15 minute Mark and I can see so here's the 1 minute load average single hot thread what is the load average after 1 minute for a single hot thread this see one it's n it's like6 um and that's because the load average is this is why I included I included because you think well it's just one because it's the one minute load average and the term um average is very misleading and they shouldn't use the word average um they should call it a exponentially damped moving sum or an exponentially moving sum which is what it was called back in uh the early 70s where it originated so uh it's actually exponentially down it will eventually hit one eventually get that but um you can see the about 6 for the 15 and 15 minute load averages that's where they hit these 15 and 15 minutes it's not the average at all this is constants that are used in an equation to calculate the exponentially damped moving average so and the constants are based on5 15 minutes but the end result is not5 15 anyway maybe I'm getting too much Mia um I was really excited to find um this reference about load averages from RFC 546 so uh the 10x load average is a measure of CPU demand load average is an average number of rable processes over a given time period so an hourly load average of 10 with be there's one executing and nine ready to run fantastic and this is a very very old RFC from the early 70s uh on top about Linux load averages themm stat and tools like this in the CPU chapter I'm just going to talk about the CPU columns so user CS idle uh wait stolen in column is there as well MP stat Love MP stat the CPU the Linux version this version for Space Systems I'm always rating it on an Aluma system like SP um s PS and so all of this is probably what you'd expect to see at this point this is kind of the what you'd expect to see in the performance book going through the tools and what what the what the columns actually mean uh but but there is a lot of background and methodology before you get there and then getting into more advanced to and so pits that I really like pit on Linux uh because it lets me break down us system time for processes um the time command the time minus V on Linux has a lot of interesting breakdowns P time minus M on size bace systems gives you the micro cting which you can get from P at and then getting into D TR and when I started writing the book I had the idea of it being very short being like Josh the point oh my Lo averages picture thank you I actually I was excited because this this is actually from the RFC I mentioned earlier uh like stuck on the wall I so excited to find it this is a it's like a hand drawing of of uh allly load averages from July 1973 on 10x systems and that was part of the RFC like this handdrawn plot of load averages and so whenever you see a graph of line graph of load averages we've been doing it for a long time that's 1973 so uh I I should have put this in the book should included in there part of the RFC that would be great um first round of ver the first round of ver so when I started with the book I wanted to cover Dynamic tracing is really important for modern systems performance and it's made a huge difference Everything's changed we can now explore anything we want it's actually part of the reason that I've EXP methodologies because previously it would be an academic exercise to say let's come up with different methodologies for people to follow like answer all these questions and there's just no tools to do it but with d Trace you always have that you always have a way to answer anything you want so I can come up with a methodology that says prescribes you to answer these questions so you can then diagnose an issue and either they're easy to answer because they're in top or in dmad orad whatever and and if they're not use drace for so it has made a huge difference and when I started with the systems uh performance book I wanted to cover both dce because it's it's there for the Lum space system Solaris freeb mosx and I also want to do system tap throughout um and then some perf and some mention of of other ones and I did one chapter where I did both drace and system tab I did the discs chapter and it was huge the chapter was getting to 100 pages I was like uh and for a book that was that I I'm really trying to make it as short as possible I could see breaking the Thousand page Mark just like the dce book If I tried to do both um and at that point the that was about when we started to hear announcements from articles about foring D to Linux Paul Fox's separate independent project was making progress and so I made the decision of cutting out the system cap off coming up with eight left the crash course of system tap in there and I have an appendix for doing conversions from dce to system tap and so you can survive on system tab is fine but for the examples so that the book didn't have an extra 300 Pages the examples are drace if you end up using system tap ktap lttng per Dynamic probes uh how many other Dynamic tracing things are there on Linux there's a lot D Tracer does exist on Linux um if you end up using any of these ones the dce examples in the book will need to be ported which is actually fine the hardest part about using Dynamic tracing is knowing what to do with it like giving you this superpower you can x-ray anything and answer any question well I don't I don't have any questions I don't know what to do with it so I've taught a lot of drr classes and and that's been the biggest problem for students is yes I get it's amazing I get it's the the greatest thing for decades but what do we do with it actually that was part of the reason I I created the dce toolkit was to give people can scripts that they could go and run to answer questions so all the dynamic tracing examples not all of them but many of the most of the dynamic tracing examples in the book are dra trce it's okay if you want to use system type or something else you'll need to quot it um it may be very wise of me at 5 years from now drace has finished being ported to Linux and it's there by default in Linux and they'll be like well I'm glad I didn't make this book 300 Pages thicker by finishing the system type work so we don't know right now but you know that's the that's how I've uh done it right now is dra trce is the primary source of examples of dynamic tracing and I do run a bunch of the D trce examples on the Linux ports which are included in the book what's the status of D TR and Linux now D TR in Linux two ports in progress in progress there were two ports in progress one by arle one by another guy in the UK Paul fo I've used that one a lot to debug issues in the giant lab where customers have issues and I replicate it um that one will Panic systems and so I cannot wrun in production um the article one they seem to be on the right trajectory and they they're paying attention to the test Suite but they don't have that um completed yet either and so you can actually Brian can who's here in the room who's the the the father of D trce you can certainly ask for his opinions as well about how long will take before it's completely put it to Linux I would I would love to have it on Linux because and as you'll see in the book it's it's hugely useful but uh those two ports are still in progress you can help them out i' I've actually contributed things to P Fox not much but I've said I I need this to work and how they you fix this and it's been pretty helpful so uh it's coming along gu the question uh how how portable are G TR scripts do they work the same on solarus and lyrics um it depends on how the D script is written ideally they are very portable because they use stable providers and that's the point sometimes I'll need to go into like this script right here this will work on anything doesn't matter it uses stable providers that's profiling CPUs some scripts do do go and use colnel and tels to answer questions and you'll need to actually flip the whatever the function names and arguments are from one colel to the other again that's actually not the hard problem the hard problem is knowing what to look at and so if you got a file system issue all right what do what do I look at in EX4 well I don't know but if I gave gave you 10 ZFS scrips and said this is actually really useful a really useful methodology and approach for looking at CFS performance you just have to convert Port them to a different file system you have a pretty big head start so these are all examples of the things you can do with Dynamic tracing but yeah uh some of the scripts absolutely need to be puted from kernel to Kernel and even versions of the same kernel so so a newer version of FreeBSD you may have to go back and change some of the scrips to make them work so dce examples um and this is the CPU chapter so I've got an example of profiling CPUs some various one liners user level profiling and function tracing this is doing uh CFS C checks some generate I'm tracing some path of Z to see how long that runs CPU cross calls uh interrupt tracing schedu tracing looking at uh how long threads run CPU for and bringing that out is showing that as NCS system tap um introduction chapter 4 pendic seat for conversion uh perf I did do perf in the CP chapter because P's a really cool tool if you're doing Linux and per can be used persistent profiling profile all CPUs generate C 997 Herz it has its own display I can convert that into flame graphs it's easier to browse uh perf can also doedu latency for the CP scheduler perf can also do CP CPU performance counters so here it's actually testing sub giving me the instructions per cycle per can do lots of Hardware events so um depends on the the processor and colal version how many of these you get and so you can actually come up with custom things like let's do instruction Cycles level one data cach load misses uh last level cach load misses data TB load misses and PR really nicely so that's a that alone is the the Genesis of pu was created to do this that over time it's it's gained all this other functionality sub commands including software tracing and so there's software events and Trace Point events this is probes inside the Linux kernel and so there I'm using per to trace contact switches and then looking at the call stack to get to contact switches at this point your I closing over this is b i I do start with the the basic tools and go deeper and deeper and deeper um it was really satisfying to document per puf because it's a gayal tool and there's not a lot of great documentation on it yet so um it's fun to contribute to that and then CP St for the for the Solaris or Lumas bace systems um to do similar things where I'm getting into the CPU performance games other tools there's lots and lots and lots visualizations um there's my CP utilization heat map which is what you should be using instead of line graphs because uh line graphs of average in fact I me tweet tweeted this the other day line graphs of average CPU utilization will hide when you have a single CP at the to or if you have an imbalance of CP utilization and so I do CPU utilization as a heat map the line at the top is CPUs that of hit 100% And so we have software we can click on that and find out the host they on that's actually showing 60 seconds of CPU utilization on 5,000 CPUs and if I did that as an average line graph it would just be useless can you imagine un average of 5,000 CPUs are 5% utilized right do I have a problem like do do I have some CPUs that are maxed out or like you can't tell so heat map for utilization very very handy um subse offset heat map you may be you may be wondering when you're using these uh performance visualizations at what point so many of them will show you information at the 1 second boundary at what point do we say we need to get a half a second or a tenth of a second like when does when does that become a real problem and I've come up with this visualization which in some ways is analog it doesn't have a fixed interval and so I'm looking at CPU utilization but on the y axis I I show the offset within a second that and then colorize the pixel based on how many CPUs were online or active and so what this does is it paints strips for each second and I can see uh Behavior within fractions of a second the patterns within fractions of a second and going going along this at this point things are running normally and then all CPS go off line it actually wrapped there's the next column and then CPS came on back online again normally since this happened in um for second whatever that is second 24 25 and 26 normally when you're looking at an average line graph CP utilization will just dip a bit and go back up and you might think all right CP utilization went to half normal and then continued but by using a subsec offset heat map you can see that CPU utilization didn't go to half it actually went to zero CPUs all stopped for 900 milliseconds sequentially and that looks a bit odd here but just the way it's wrapped and and that is a serious problem and then this is a problem of of a lock contention where a lock a dress bace lock completely blocked an application for 900 milliseconds and so all was a database and all the queries were blocked behind it and you can identify it from the CP usage but you can't identify if you have if you're using that 1 second interval or you're doing uh utilization so sub an offset heat map was great and then flame graphs which I've talked about a lot flame graphs do uh let me look at profile data in in detail and I can understand uh where what caths are consuming the most C time so then I talk about some quick benchmarking and finish with some tuning just to give an idea of the tuning options that exist but I don't want to get into uh you're going to have to go back and relook at these things as the years go on to find out what the tuning of the day is so that's a quick look at what what's in one of the chapters uh I had some other things to mention um chapter 11 cloud computing was another highlight that was actually really difficult because did I wanted to talk about um comparison between Unix OS Spas fertilization Zen and KVM and how they all worked and um I don't think many people have actually done a lot of performance work on all of them and so fortunately at joint we do run obviously we have know normal systems where we have OS fertilization and KVM and we have runs in in the past and so had a lot lot of experience debugging performance on all of them and could summarize them for comparison talk about how you observe performance of each how they actually perform and how resource controls can be applied modern system performance which I talked about um that's what the book does have and at at this point this may be pretty obvious so I'll go through this quickly and one thing I've done is this is not there's a lot of documentation you'll see that kind of originates from the 1990s this is where have like text based um a lot of text based Tools limited metrics and documentation some perf issues can't be solved um but it's different nowadays and the book reflects that where um open source is the Norms so we can go and look look up how things work we can Dynamic Trace everything to see how they work visualizations like I showed how Computing is is often involved from that environment and now we get to do methodologies like we haven't before performance visualizations have come a long way since the '90s and things like xload and line graphs uh we're now able to do things like heat maps and Flame graphs so we can really get a lot more detail out uh Modern Performance analysis tools as well so nowadays there's a lot so these are actually slides from other presentations I've given I've just been updating them um that's the sort of tools you have these days on Linux and we have our spread of the Dynamic tracing tools to look at everything and then the uh the tools for lumos and again dra trce can see everything it's a bit simple because there's only one dce uh Dynamic tracing if you need it that's example of some of the scripts I've written for investigating different parts of the code and so uh and I do mention a lot of these scripts in the book just mention them because it's part of turning unknown unknowns into known unknowns so you know that this does exist it is possible you don't need to know how to do it now but if there's some crisis at work you know you can look it up and it's a possibility but those slides are actually not in the book people get excited I I've done these before and they really like them but they're not in the book because it's not about the tools it's about the methodologies that lead you to the tools and the tools are really the answers to that I do have two examples which I showed but it's more about teaching you how to apply the methodologies like the use method the T method drill down analysis workload characterization so that you can solve issues so that's a quick rundown about what's in the book um and that's I will if you find me on Twitter you can I will tweet this URL but I've linked to the sample chapter on my blog um you can find me BR and greg.com really hope you enjoy the book it's a fantastic seing page I think we'll be giving out three copies and I I hope it is is very valuable thank you Back To Top