hi everyone and welcome to this new mli meeting today river is going to present about a new front-end pattern language on top of the pdl pyth code hello i'm river i'm at google gonna talk about pdl which is uh a pattern language oh good if i can okay there we go uh so uh pdlo uh it's it's a front end dsl for for writing our pattern rewrites and the kind of the interesting thing about it or i guess the obvious thing from the name is that instead of tuplesplus the entire point is the cogent to pdl which is a an ir infrastructure for pattern rewrites uh written in mlir so there's no no c plus generated but uh it still allows for super plus or uh some other native code injection um and some nice caveat from that is that uh because pdl i mean pdl itself the single l uh because it's like a bytecode representation you can actually provide these patterns and compile them and stuff at runtime so the language itself uh as a c in a little bit um is effectively designed uh as a high level representation for nyr constructs uh and pattern rewriter constructs so this kind of comes out in the design of its language the language itself it takes a lot of a lot of things from my format um but also in the things that it's it's intended to represent which is effectively everything that we have in myr so it's like optional variatic regions successors uh type conversions all of that stuff is within bounds with what we want to do um but kind of aside from that is that uh don't really just want to build another language uh one of the main goals is to have a good experience so this is a kind of like we want to make sure that whatever we provide has the tools that people expect from a language so like code completion go to definition formatting all of the stuff that people are accustomed to um so yeah i guess you know the easy question is like why are we doing this so i mean if you think about it kind of pragmatically a lot of the complex rewrites in mlr kind of require uh c plus plus or or some kind of like the c plus plus api because they're necessarily complex but there are a class of use cases that can benefit from from things being done declaratively you can think of like instruction selection in llvm or you know if you think about ml programs a lot of those really benefit from from kind of a declarative specification for for rewrites um so kind of recognize that these users exist and we want to provide something that is a good experience for them um so i mean recently well recent is a relative term obviously but yeah when we were implementing pdl and providing pdl we started to think about like you know what's what is going to target pdo upstream really because it's intended for multiple front ends but uh like what are we going to provide is like the default thing that's oh that's always available um so that kind of spawned into pdl um i mean the the first thing to think about or i guess question is you know why why not just use the table didn't dr which is what is currently available but uh if anyone has ever used table gen you probably have mixed feelings about it especially because table jen's not really good for myr uh kind of sucks really um well sucks is relative a lot of like problems have come out of table gen with even like simplistic things like multi-result operations which are like basic myr things uh not even going into how you would represent like blocks or regions or anything remotely complex like that um so if you take even like a contrived example of like you know i have this like three result operation and i just want to like match some of the results uh you have to like bind them all to one name and then use like this magic like underscore underscore number to be oh that actually underscore underscore two actually means that it's like the second result group of this uh result which is kind of hacky ish or at least i don't feel good about it looking at it but if you were going to do this in like a more sane programming language and this is kind of spoilers first kind of a little bit of what pdl looks like you know you can have you know variables and then access results of those variables because operations are conceptually like structs or tuples or things like that in in other languages uh and because you know obviously this is something a bit more modern you get code completion and documentation and all of its stuff pulled directly from ods so all of that stuff's available so it's not as annoying or confusing to write at least ideally so if we're not going to use table gen you know what are we going to do there's obviously a lot of different options out there uh and a lot of times people ask you know why not x and x here is kind of factored into the the no and the the yes and no at the bottom which is that x means something different depending on who you ask because different people think of different things so x could be like python it could be rust it could be haskell if you're uh in that crowd um and when you're kind of picking a host language for dsl there's a lot of things that you have to kind of keep in mind uh especially when you're going to distribute this in an ecosystem like lvm like we are intending to so different host languages kind of have different dependencies they can have different environments performance that you know we might not want to constrain on all users or all of the the places that we want to deploy the language for example but not just that tooling is also a very important one because one of the intentions behind this is to provide a good development experience good being relative or subjective and if you're building tooling into a custom dsl that's embedded in a host language if your dsl doesn't really map directly to the the host languages concepts then your tooling is going to suffer like hugely and especially if you're in an established environment such as like a company or corporation or even a team that has like language servers or other development tools built around those languages already they don't usually take kindly to hey let me inject my custom tooling for this simple dsl into like the main python language handling uh usually they uh they say no to that stuff um so we kind of made the decision that uh to kind of provide the best experience and to build like be able to deploy in all the places we want to then having something custom is is not that bad uh and it's kind of easy it's easier to kind of achieve the goals we have in mind for it um but obviously the flip side is that of that is that uh you know pdl itself uh was designed around multiple front ends existing um so that's kind of a recognition that different users have different constraints um not necessarily the ones that we're placing on ourselves but i just want to make sure that you know we're still open to having multiple front ends and pdl is is kind of intended to serve as kind of like a baseline in terms of pushing pdl forward and also in terms of kind of establishing a baseline of features that you should provide if you're doing a front end so it's kind of on the fence about what to to show for the language itself whether to have like 80 slides uh or to just have code and have people's eyes glaze over so i chose the isoglazing glazing over one because that one's technically easier for me so i'm going to bore you with uh code of what it looks like because i feel like that's at least more interesting to look at than uh than slides and it has better formatting and syntax highlighting and all of that stuff so this part of the presentation is kind of free form mainly because i don't have any sleep but it could go fast it could go slow people can interrupt and ask questions or if they want to see something uh so i don't really care about being interrupted and that's fine uh so i'll just kind of go through this huge file of examples and kind of just talk about some of the different aspects to give an idea of kind of what the language looks like and at least what we have in mind initially for it so starting from like a baseline uh include files so one of the the key things here is that we want to be able to import stuff from ods easily because we want to avoid duplication of things obviously because it's a different language but we don't want to have to duplicate anything that users have already expressed um so here i have including like a few table done files and from there we kind of import like what operations you have to find what dialects you have to find interfaces traits uh constraints all of that stuff attributes types all of that so that it's immediately usable and that kind of helps the experience it's not technically required but it's extremely useful so from there i can kind of go into like patterns which are obviously the most important part of the language if it's intended for pattern rewriting uh and just like sql plus patterns have you know like technically like three main components uh they have like an optional name for like debugging or depending on how you're generating can be useful to have a name you can have metadata which is things like benefits or recursion so if you know you're writing writing a pattern you might have like benefit attached or uh rounded recursion so these are basically what you would see in uh uh spots if you're writing people supposed better than the types of metadata that you can attach and then you have the body and the body of a pattern is is conceptually similar to sql plus you have a match section and then like a rewrite section and the match section is important because it's describing your input ir um so you can see here don't pay attention to what these patterns are doing most of them are just things to illustrate concepts um but if we're kind of describing potentially a reshape of a reshape then you might have like an input operation that's your resa shape and it might take as an input a reshape that takes this uh this argument which is uh the input value so that is conceptually your match uh and then a rewrite you might want to replace this this root operation here with uh like a reshape of uh the argument they have bound um so the rewrite section uh traditionally is like the last statement of your pattern that kind of starts through right so uh in like a replace uh the roots is just like an expression for an operation and that is still technically like your match um but after that anything with uh that you're replacing it with is uh going to be created so if you have like an operation expression in the match section obviously that's an input operation but if you're like in the rewrite section you're creating ir so it's the difference between like checking if an operation is like a specific type versus uh creating one uh but then you know you can write that in a much simpler form if you want uh and even simple patterns can be like in lined into one line that's like a quick example of patterns it's not terribly complex hopefully just the things inside of it um so then metadata like i showed before benefit and recursion are the major ones that that meyer has right now but you know if more get added more can be supported that kind of thing um so maybe obvious uh from the uh the examples before but uh there are variables um and uh variables in pdl are are strongly typed uh and the types here correspond to ir constructs so things like values operation types uh all of that stuff and the types come from these constraints and there's a few kind of like built-in constraints that uh that give you the ir constructs from these things are the ones that imply the type so uh these are the gold version ones here are some examples of them so you know attributes operations types values etc and then from there you can kind of compose to make more complex constraints and i'll go over that a bit later but those can be things that are imported externally so if you define them in c plus plus and register them you can also define them in pdl itself and other things like that river how do the native callbacks work is that something that you um register like you kind of at run time past the c plus function pointer into the pdl infra uh yeah let me go down here so [Music] so uh right constraints uh like i said can either be defined in pdl which is just more of describing the input or like constraint and these constraints don't really turn into anything they all get in line they're like functions uh that kind of thing but if you have like an external constraint like i was showing before these are things that are registered with uh the c plus plus api by the user so this is something that the user has to like explicitly say you know here's a mapping it's like a jit yeah yeah it's like you pass in the string external constraint and a function pointer basically effectively yeah yeah um so you know these ones are the ones that the user is like providing as part of the library and you're kind of just importing that fact uh or you can kind of like drr define them directly so if you're doing kind of like an aot code generation where you you know how like table gen does today where you're like okay i'm generating something that's included in the cpp file uh if you're doing that you know in pdl you can like inline the sql blob if you want and we'll auto generate that and develop the wrappers for like oh you don't want like a generic interpreter value you want like an actual attribute or a value kind of like here for example uh so kind of just makes it simpler in the simple cases to just like i just want to call a c function i don't have to deal with like registering with the interpreter and stuff like that um but we still want to leave that open in case you know if you're we're calling from like a different language or if you already had like a library of things that you register things like that and that also applies with rewrites but i can go over that later right so what i'm showing with the the constraints the ones that you define anything that takes like a single argument uh so these are the the core constraints that they call them which is like value value range those types of things or you know if you've defined a constraint and it takes a single input those can be used to like directly on the variable definition itself so here you know you have value in one use but if we're going to do something like uh you know like capstop interface takes an operation conceptually because it checks if it's an interface um or you can use has one use directly because we know a value is an actual value because that's the type of the argument so this helps in the simple case if you're just like i want to constrain this variable directly and i know this has one argument then that makes it easy in those cases um and i kind of showed it before but you can define variables out of line but you can also use an inline syntax uh so if you ever use c sharp it makes it very useful if you're kind of trading like dag-like things where you don't want to have like in lines where it's just like okay i just want to have a value here and this is especially useful if you're kind of like constraining [Music] operands for an operation for example so river perhaps one question so like in the previous one like line 83. could you perhaps actually talk to it what is that pattern doing i mean the pattern itself it's not particularly useful but uh this one is kind of just matching uh this my dialect operation that has an argument and then replacing this operation with its value so it's the same as this one which is just you know i've matched an operation that has an input and i'm going to replace that operation with its input effectively so [Music] right in this example we define this variable inline so we have this argument that's already constrained uh and we want to check that you know the first and third arguments are the same thing uh but the second value and i'll go over why in a second we don't care about we just want to know it's there um so if you're kind of defining the uh the operands of an operation i'll go over a snippet but there's kind of two ways of doing it the first is to have like a value range which is like okay this is all the operands the second is to match what you've defined in ods for the operand description so this kind of maps directly to how you define the operation because pdl is generic so it needs to be able to map between these different constructs so if we go to the operation itself that's how it's been defined so i've gone over it quite a bit or at least shown a lot of examples of it but if we consider like an operation expression which is either in an input it's like is this operation and defining in the input ir like your root operation and in the rewrite it's modeling you know what you want to create so the syntax of of the this is expression if it isn't obvious directly is uh the exact same as the generic assembly format like down to the the letter um and that's kind of intentional so that when you're writing it's not confusing what constructs go where uh if you've ever written generic ir you at least have some idea of what constructs go where um so you know this isn't all of it but you know i have operands attributes and the result types there so you can have oops you can have a named operation which is you know dialect.blah or you can have it uh in your match you know i want to match any operation so i'm just not going to provide the type of it and the reason why i had something here before is you know if you're creating an operation expression uh the nice thing is because this is like a real language uh you can get code completion for all of the dialects that you have and all of the operations in those dialects uh and all of that stuff uh so if you want to match like uh constant you can see that go to its definition uh all of that stuff so hopefully makes it easier to write things as you're kind of going along because you don't have to guess what goes where uh yeah uh so i showed it before so operands you have two ways of writing them one with a single value range or one which matches the ods description um and if you have your ods thing imported through like an include file you obviously get like signature completion for what you're supposed to be putting there so it's you don't have to guess what goes where and you get the nice autocomplete or code completion for our signature help whatever it's called for what all of the different argument types mean and results are except like conceptually the same thing which is yeah the audience description uh you can get all the signature help but there's two ways of doing it one with the single type range which is all of my results or ones that matches the description and i showed it in kind of the introduction slides but uh you can access these uh results by name so if you're going to do that result but you can also do it by index if you're in like a template or something and want to be a bit more generic about what you're actually accessing okay and then obviously one of the the other things from the example that i showed is the the attributes themselves which again is exactly the same as uh the assembly format so if you're going to add uh the attributes here you can get autocomplete for the attributes that uh that are expected uh and what the audience description has for them and unit attributes are the same thing if you don't provide like an assigned value it's just well i expect this to be you uh okay so those are operation expressions hopefully not too confusing uh and and the constructs that they represent uh given that the the syntax for it is exactly the same as the generic format should be somewhat obvious or at least similar to what you've seen before um but kind of going on from there there's things to represent like literal mlr attributes and types because we're going to be generating pdl which is myr so it's trivial to just say i want to create this attribute in mir because we're already generating ir anyways so it's really trivial to kind of just inject these constant literals for different constructs that you want to have i won't go over tuples too much but it's basically a tuple if you've ever seen in another language which is like a kind of an ordered collection of blobs or objects and you can name the elements of the tuple directly or you can just kind of leave them as is and kind of get the uh the generic you know i just indexed it by number um this is kind of like a common like a glob of the way the different languages have tuples just make it easier uh right so i kind of showed this earlier when sean had the question about constraints but there's a few core ones for like attributes uh if you want to constrain that the type of an attribute all of those things operations so you can have like a generic oh any operation or one that's constrained to you know a specific dialect operation uh type type range value value range so all of these represent kind of like core ir constructs uh themselves and imply like the type of a variable because types in this language are directly corresponding to what you would have in the ir river what is the erase up here is it the same as eraser's roots uh so this is why i say don't pay attention to it because this is just this is just filler to make sure things compile and uh okay okay that system is basically if you were to really write this this is uh destroy everything which could be interesting uh but this is uh for this erase statement is basically erase all and or erase this operation uh which for this pattern would erase everything because the root has no name or anything constrained on it so it's kind of like rm started star i guess in a way but that's interesting though like that means that the what would the pattern match here what would be the roads when because there needs to be a match before the arrays so the erase this part of the erase is considered part of the match because you're erasing something that's that's being made right okay okay so that means that the match for the wall pattern would be everything basically that's what you mean conceptually again and because we match everything that we erase everything um so then we can kind of show our showing sean earlier uh which is uh defining constraints in pdll and these are basically functions uh if you've used any other kind of language which is uh you know if you want to have like a fragment of your match and share it across multiple different uh instances or have something imported externally things like that i mean they like utility functions in a way like shared functionality um and then the native concerns um sorry sorry quick question can you come back to the last constraint stuff you showed yeah what's the type of this multi-result thing is it a tuple or is it something else uh this one on the right hand side of the arrow right right value value range yeah that's a good question um yeah it's effectively a tuple so if we had you know a result here and you could call this uh directly yeah the result is effectively a tuple uh exactly as you were saying that's kind of why google's existing language anyways uh it's to kind of support this multi-result um so when you're defining the syntax you're basically describing the structure of a tuple so you're giving it a name here kind of like a swift uh does with with uh multi-results but you don't have to give it a name and in this case the return type of this inline constraint is inferred uh so if you're gonna have uh you know inline constraint instead um it takes whatever type was returned so if you have like a tuple here then it's going to use that tuple um so if we did that um now we have like the name from it so yeah in the case of multiple results it's effectively just the t-bone that's being created i see and the inference here happens because the definition in ods of multi-result has two things one result and one other results which is value range in table 10 or something so this uh inference right here so the the inference for the name comes from ods but for tuples themselves tuples are outside of ods because you're not like for here you know you could return like some like type adder or something and you have like typewriter here and then let me call that it's just um so you can access field2 so tuples can be constructed with any variables any values that you want because it's just a collection of uh kind of ir values the kind of named results so doing like op dot results those are the things that come from ods to be able to do that if that makes sense okay cool awesome thank you um right so kind of going into the rewrites um so there's there's kind of three main ones right now so like erase replace and rewrite a race is uh kind of what uh many asked about earlier which is you know if you have a pattern like this erase you know my dialect foo that will erase all instances of of that operation and if you have a replace you're replacing some operation with some set of values so in this case these are values constrained to the first and third inputs that are being reused down here because these are inline variable definitions uh if we go to the definitions here you can see where it's defined at right there yeah and then rewrite is kind of conceptually similar it's basically a way to have a block of uh some number of rewrite related things uh so all of all this root does is kind of anchor the pattern on a specific operation but inside of it is is how you're actually going to transform the ir so in here is all treated as like a rewrite so you could have replacements you could create you know other operations in here anything like that so this is just a giant rewrite scope effectively and like constraints uh rewrites can be defined as well and rewrites are called from the rewrite section so if you had a simple rewrite that was going to create like uh my dialect up two operation then you could call that with a specific value and this is obviously a very simple example but the rewrites could be as arbitrarily complex as you want and do effectively anything you want straight anything call into us whatever and the native right can you make can you make line 392 fail to compile i guess or something if you return there create fu up return an extra after and you try to replace it what do we see so if you're going to do something like kind of like that right um it's expecting a kind of an operation expression or something else um okay so it's going to complain right there that there's an adder he doesn't know about it doesn't expect it okay awesome yeah so uh i think that's something that's uh interesting to show because if you had like a value here that you constrained it to like an operation and you're like oh actually it's an attribute and it's gonna be like uh actually no that's not something that exists in this world um so yeah i kind of glossed over it maybe because i thought it was implied but the nice thing about having like a real language is that you get error handling immediately instead of like a miscompile or or like random c plus plus errors that just happen to pop up so kind of one of the ideas is that anything that can be handled in terms of error handling is handled immediately that that is uh exactly the way that you would expect um so this comes out in kind of everything so if you uh we're trying to use like uh an attribute for like a value uh then it's like okay an attribute is not a value or value range that doesn't work for this um because if we think about kind of like the way that table gen dr works because it's structured around like the arguments of your op and arguments are either operands or attributes people actually often get confused on where the attributes go and where the operands go which has created miscompiles and other weird behavior but the nice thing about modeling it after like the generic format is that a it's obvious and b or obvious if you use the generic form and b you know you get immediate error handling because something isn't uh happening the way it's supposed to so i guess just benefits of of not using something that's maybe not intended for that um so i kind of went fast through rewrites because conceptually they're they're the same structurally as constraints so not horribly different in terms of how you define them or use them i mean rewrites obviously you can't use on a variable because it's not constraining the variable it's transforming something uh i have a question it's a kind of a tangent um i don't know if we're ready to switch topics oh sure sean uh yeah i'm curious is the type interrupt um like you know custom types attributes like right now it's like so that stuff is like really c plus plus e but like is there any hope of you know being able to like say you have like a you know hlo dot dimension numbers or like convolution parameters or something that has you know like a bunch of fields being able to like check those easily in here and do like a little bit of math on them and stuff so you mean i guess it depends because like checking if something is like that attribute that exists that's fine um because it's kind of conceptually the same as like checking if something is like a specific um if you're gonna do i guess here's an example uh so like all of the attributes types etc that are defined in ods are are imported automatically so in terms of like is this like an hlo dimensions adder or is it like a cap interface or is it like something else uh all of this stuff is is fine um what kind of comes from there is we're kind of constrained by what we can reasonably generically support with pdl itself yeah so in terms of falling back to kind of custom code all that's easy and supportable but in terms of like doing the actual math in pdll itself then it kind of gets into like what built-ins we want to provide and what like of those types of things i think there's also one other thing which is the destructuring of the like looking inside of the attributes like say you have like a struct adder that has a bunch of fields i guess we could improve the same for types you know the the type and attribute ods thing to sort of emit an interface that allows us to sort of uh reflect on some of those things at run time so then it becomes possible to like you know check if some struck that you know field foo of some struck data is you know i32 yeah exactly i feel like a lot of things are going to come out of that which is uh pdl itself the underlying infra wants things generically because it has to support everything and it's you know anything that we want to do with it has to be through an interface or some other kind of uh generically accessible means so uh if we want to start doing that kind of stuff yeah it's gonna push like interface design or at least generic accessibility of attributes which can be beneficial and in general constant yeah yeah totally i think the python bindings would also benefit yeah uh i'll kind of quickly go over templates templates are very similar to templates in other languages except for the thing that can be a template type is something that's not an ir construct but it's something that's like statically known so these are things like constraints operation names or rewrites those can all be passed in and kind of used directly here it's like passing in different operation type for the route that's going to be replaced in this case it's a 2 and this is a bar in this case we pass in the name itself as the the template type and then pass that into an operation expression in the last case very similar to the constraint we've passed in uh kind of like a rewrite that is going to be called here so slightly rather simplistic in here we've yeah it's passing in the specific rewrite uh as the template argument so that's kind of how the the template types and like constraint arguments i guess are kind of differentiated is whether it's like an ir object or whether it's something like static right i think i have other examples of templates of templates and things like that wow okay uh so it's a kind of a final example is uh some like range expression type things um so these are concepts similar to like a range library or like a lot of the lem utilities that that might be used to constrain ranges so uh these are things like you know checking if uh all of the values of a value range satisfy some constraints or if all of the values and types within a value or type range satisfy some constraints same with things like any of checking if any element satisfies some constraints checking if ranges are empty these are just a lot of like simplistic range things that are very useful if you're kind of building more complex uh patterns none of which is very similar i think probably the more interesting ones are like the map functions which can perform like reductions over values so if you want to map all values to a constraint it's actually conceptually similar to like all of in a way they're almost identical infrastructure um but maps can also be applied to rewrites if you want to uh kind of take a value you could apply multiple different rewrites to it so like create operations that use the value or the type or things like that um and then as i kind of mentioned earlier you can perform like reductions over it which is you kind of map a value range over this and so if this rewrite returns a value in this case this copy is actually a value range which is a concatenated version of of all of the elements so whatever you created here just gets combined into like a giant list of things and the same works for like if you were to return like a type range or something like that they would all get concatenated into one giant type range um and then kind of a similar sister expression to map is this map if which applies a constraint or rewrite if a constraint holds so maybe you want to constrain like values that have one use differently than something else um so it kind of gives you like a subset or like some like structured control flow application to these things um so if you wanted to like apply some rewrite so like create operations for like all of the shaped types of the type range that's something that might be used for or like a map could be used for like creating a tuple extract for all values of about your range or something like that uh so these are kind of just like semi control flow range expressions that are added but it's kind of getting into the more complex types of rewrites and things we want to support and that's so what what can you what can you map over what's the type of the first what types for the first for the container are supported so value range type range any range is supported uh integers so like that's where it gets into what i was saying with sean which is it depends on what is a built-in construct in pdl and like what i supported so right now pdl is all like ir constructs so the types of ranges here are like value range type range and operation range for like users things like that because there's no notion of like an integer or a float in pdl itself so like these types of things depend on what we want to make is built in if you were going to like map over a range of of integers we would have to decide how what that means right because it seems that since you already give us templates you give us these ranges if you give us just basic you know next arithmetic like i don't know like axioms of natural numbers right we need a zero and a plus one then we can suddenly write all these nice unrolled things hey if i have an eight by eight vector and roll that to 64 versions of this thing right for that would be super instance i would be happy to move all the vector rewrites into pdl yeah so this is uh i can switch over to the the slide itself uh and that kind of goes into like the status of it uh so i think this will partially answer uh why those things don't exist uh which is that the the current status right now is everything in the demo works and is supported uh and kind of the initial feature set that's been focused on is uh for example like what would it take to kill table gen drr for like an example of uh the kind of features we're focusing on first uh but then after that is like getting into the the kind of crazier more interesting things so like if we had like kind of like a structure like okay i want to iterate over you know i just want to apply some map function like 10 times to like unroll something for example and other things like blocks and regions or dialect conversion type conversions things like that that all of these things are things that we want to support um but stuff that we kind of want to hopefully build on uh as we get like base things working and this also applies to like making sure that pdl itself is optimal or at least not slow because detail compiles at one time so it's kind of important um and that kind of goes into this which is to give an introduction to the language uh kind of before we kind of formalize an rfc and develop upstream because we don't want to have like a huge huge huge like drop of code uh and i also want to make sure the community can like factor in things like nicholas as you were saying like okay i want to be able to you know iterate over like a range of numbers and apply some like rewrite or map function uh because these are the types of things that we want to have conversations for and hopefully have the design like done upstream and code review done on stream and this is kind of why we've been focusing on initial feature sets so can get like a base thing agreed upon or at least ideally and then kind of move from there uh and that's all of the slides do you see this being used for name angling as well [Music] name angling yeah for example if you have multiple operations with the same name and you want to name angle them so that there's no conflicts for example given an input source that can can have multiple names at the same name like you you could have a symbol table from another language which supports similar symbol names right i think ideally anything that doesn't require like super complex c plus is in bounds to support so as long as it's it's reasonable i don't think there's a no there uh i think the things that we want to avoid is having the byte code become like basically a cpu jit in a way if that makes sense so as long as that doesn't happen anything else is kind of in realm to support that would actually be kind of cool dang it sean as long as it's not like a full x86 i think some simple things are fine but uh making sure we don't go the full route of of absolute craziness okay yeah so i would like to do like i'm literally writing this now is i have a um like a torch script module and basically you know sometimes i i might want to like interpret that and so if there was like a sufficiently powerful interpreter i know this keeps coming up then you know i could compile that at least like a compact byte code um and have that i know that's kind of a reach yeah so if i understood correctly the pdl is currently compiled to uh to emilia at runtime and then the pattern matching is also executed at runtime using an interpreter right are there any plans to also support some kind of ahead of time compilation similar to how table channels sort of functions today so the the point behind pdl is that the pdl itself is compiled at runtime um because i guess it depends like we could support modes where it's pre-compiled but there are different caveats there one of them is that often it's useful to have the dialects that you're matching linkedin to perform optimizations like removing redundant computations or or things like that and another is the cases where you have a bunch of patterns being combined together so i mean that's a lot of the benefit that you can gain from or at least some of the benefit that you gain from like a by code abstraction that's compiled at runtime is especially for myr is that you know in canonicalization you have patterns coming from like in different unknown sources um so if you compiled ahead of time you might only have like one pattern and that's not really going to give you a good benefits you're not going to have any kind of redundancy elimination or any kind of useful merging um but if you had like one pass that has all of its patterns defined there then yeah in that case compiling ahead of time is probably more useful uh i think this kind of goes into the evolution of how we want pdl to look uh it's semi-orthogonal to pdl itself because this is it's it's a use case that i find interesting but it's not something we've like scoped out in terms of uh like how to support it or how it would be supported so i think it's plausible it just hasn't been scoped out well in terms of pdl itself i have a question though uh marcus why why would you want to iot compile it um it just seems like a much easier language to work with than the table trend declarative free right at the moment and if you were and for a typical ahead of time compiler um you could then just it would be a simple replacement for a table chain driven one which currently compiles the c plus plus and if you could use the pdl definition language to also compile to c plus plus i would expect similar performance from that um different things so there's like compiling the pdl itself ahead of time and then there's compiling it in a way that replaces table gen ahead of time so it supports modes where you like generate the c plus plus just like you would with table gen to like add the patterns to your list and things like that so all of that stuff is supported so you could conceptually use it as an immediate drop-in for table gen like it has the same like user-facing api in terms of like what you would add so all of that is like supported what's different is in if you compile the pdl byte code itself ahead of time versus not because it's slightly different so marcus your concern is mostly with performance like you're afraid that the bytecode interpretation is going to be slower than the the code table chain generates sort of here i mean if if the part code would be fast enough i don't want to assume that the byte code or any runtime interpretation is going to be slow enough i was just curious whether that's something i mean some might be saying that it's worth uh watching the presentation on pdl itself uh about this but a lot of the things that were underlying the development of pdl was the idea that by jitting the bytecode first by compiling multiple patterns together in the bytecode you can do csc across matching meaning like if when you share table gen there's a lot of redundancy and we're going to match one pattern honestly on the same up there can be multiple patterns and we're going to try to rematch rematch pretty much over and over the same code with pdl because you have the global view you can create a tree of matching and reduce the amount of redundancy that's one thing another thing is that by cheating um you can start doing smart thing for example you have already the context available so if your matcher says is this attribute you know the unit attribute presence or is this uh a true like for example loop unroll if my attribute is loop and roll true because you already have the context when you jet you can get the pointer to the true attributes and replace this with a pointer comparison instead of is this a boolean and that get the value is it true you know um so there are a bunch of things that you can do in the jet that would make it even potentially uh faster much faster than this so instead of all the complicated um conversion and pattern matching uh thing that is currently done in c plus plus that ensures that the patterns are transitively work you could compile the all patterns in the sum and have faster patterns so to say as an end result due to that thing yeah i'll send you a link to the previous presentation on the the byte code because all those optimizations were presented last year actually today is mostly the frontend part on top of it i see okay it scales better if you have one c plus plus pattern it's probably faster than one byte code pattern but if you have a hundred patterns the ch the byte code stands a pretty good chance of being faster and yeah it makes a lot of sense thank you as far as compiling patterns ahead of time you could like build the pattern tree ahead of time but um because the by code will have like attributes inside of it some loading will have to be done at runtime i posted in the chat for your reference the two presentation two pass presentation the first one was uh jeff's internship results when he developed the first prototype and the second one was um reverse presentation when it was what is what has actually been cleaned up on upstream in the final version and today is really the language part of it the previous one where the underlying engine and the optimization we were anticipating today yeah and perhaps it's good to emphasize this is the it's very much working process where we'll be eliciting feedback from the community with respect to the language and features and all of that so this is not you know this is under very active development um and you know i think there's a very nice sneak peek of like how such thing would work and you know the better user experience etc etc that it could provide um but there's probably a lot of discussions still to come on this on this course on the rfc et cetera but if folks already have feedback you know uh feel free to start discussions you know that'll be awesome thanks any last question anyone otherwise we're at the end of time so thank you everyone for joining today thank you river for presenting uh this is very exciting i'm looking forward for the rfc on this course and um well see you next week everyone bye bye bye Back To Top