plus over infinity / π crash course
- get | run
- style β cmt | sep | tab | ids | space | bad
- parlance β xyz | rank | proj | +: | scope | nouns
- vector math β vβ a | v+v | v+a | v x | shp
- type system β num | char | name | date | dict | tab | Ξ» | ΓΈ,β | mix | cast
- order of eval β rtl | (1)2
- no stinking loops β over | scan | eβ£ | eβ | eβ€
- reading code β how to solve it
- writing code β three triangles
- metrics β apples and oranges
- learning plan β gladly beyond
quod tu summum putas gradus est β seneca
Computer languages have been around for a long time, but in the beginning was the Word. We will be writing code in a language called π, but it helps to talk about it first.
π is different. At first, you will be questioning its design, and it will respond by questioning things that you consider common sense, but soon enough it will become a constructive conversation, and here is how.
The first thing newcomers frown upon is why the assignment operator is :
instead of =
. But before you close the tab, try a simple thought experiment:
x = x + 1
Most programmers will agree that this expression makes perfect sense, but if you show it to a math guy, be ready to hear "no, it isn't". And once you see what makes him think that way, you will also see why we assign values with :
in k. The above expression looks nonsensical to a π programmer for the same reason it does to a math guy, and k will most always evaluate it to false. Once you can produce a π expression where x=x+1
evaluates to true, please submit a pull request and say "hello".
π offers some unusual perspectives compared to other programming languages, and they are not necessarily wrong. They are simply different - in the same way that C++ is different from Java, and Python is much unlike Javascript.
This short introduction invites you to look at those things that other way, and we hope they'll come across as obvious and natural to you, just like xβ x+1
did a few seconds ago.
This crash course is not looking to make you an expert π programmer, because that takes a lot of time and effort. Instead, it aims to give enough confidence and motivation for you to continue on your own. We value your time, so we promise it will be fast and violent:
- The text cuts a lot of corners on general programming and CS at some expense of readability.
- The course is driven entirely by densely annotated code, comments contain essential material.
- New syntax is often introduced inline, some is self-explanatory, some relies on your intuition.
- The narrative is linear, all chapters build on previous.
- Skipping exercises will halt your progress.
This might feel a bit intense, but we hope the course is still lightweight enough to be completed in one session.
This document is not a π reference. The majority of subjects are discussed at depth sufficient to give a good general overview, but by no means exhaustive.
The man behind π is a computer scientist by the name Arthur Whitney. He is the principal designer of the language, and he is an iconic figure in a community of some of the most sophisticated programmers and scientists employed by some of the most influential institutions on the planet. Since early 90s, he delivers ever more powerful revisions of a concept he has been refining throughout his career, a system to build very efficient software that transforms large amounts of data into large amounts of money.
That is, π enjoys much success in the world of finance, where this kind of problems existed long before the term big data was coined. Many people embraced the π way and made successful careers by building solutions using π, and they appreciate their tool as much as they appreciate the man behind it, and we believe they have their reasons.
There is a good chance that you have heard or read about π language β a lot of people know the story β but what is much less likely is that you have ever met a professional π programmer. This happens not just because π programmers are rare, but also because π is not fishing for cheap publicity. This is how we've all heard about a language called π, but what we mostly get to hear about is how much it sucks to be a Java programmer.
But jokes aside, implementations of similar systems in languages like C++ or Java usually involve thousands of lines of code written by large teams, built on top of complex library stacks and even more complex infrastructure. Such projects are often expensive and inflexible, go over budget and miss deadlines.
In comparison, π solutions are typically a few factors of magnitude less code, implemented by small and agile teams, rarely require external dependencies, and ship on time.
At first it could be hard to understand how this can even be true, but imagine the effort of keeping 100 lines of code in sync with rapidly changing requirements and free of bugs, compared to 10,000 lines of code that do the same thing. Against all intuition, it is not 100 times easier, but 10,000 times easier, due to a devastating effect of cyclomatic complexity.
π is a simple, expressive and powerful computer language.
Its power stems from the fact that π is designed as a tool of thought. The vocabulary, syntax and the choice of abstractions offered by the language drive you to think about problems in a focused and clear way that quickly takes you to efficient and elegant solutions. And the reason why thinking in terms of π is so effective is nothing supernatural: brevity is the soul of wit.
π programs are concise, the syntax of the language is terse, and there is no boilerplate code to write. In π, most of the time is spent on thinking about the problem rather than writing and refactoring code or browsing source.
* no, this is not a typo
π runtime environment is an incredibly compact and efficient piece of software.
The entire system is:
- a single binary executable
- without any external dependencies
- that fits in the cache of your CPU
And that gives a selection of fundamental algorithms, data structures, techniques and primitives that withstood the test of decades of production use in some of the world's most demanding data processing environments. Inner components of the system fit together and complement each other to deliver performance. It is not uncommon for π newcomers to experience shock when they first see how much can be done with a few precise keystrokes, and how fast.
All of π programming takes place in the REPL, an idea that is actually much older than many seem to think. It has been around for at least half a century, and is known as dialogue approach, a live conversation between a human and machine as a flow of questions and answers. And in π, this conversation is much more fluent than in any other modern REPL-driven system you may be familiar with, because the questions are short and the answers are fast. This is the essence of the way of π, an experience that all π programmers consider immensely satisfying. People who write π for living love their jobs.
The only known way to learn how to program is to write programs, so you will need a live environment. As all things π, it takes very little effort.
We will be using a trial version of π, which comes without any RAM or useable CPU core limitations. As any π programmer will tell you, with great power comes great responsibility, so please donβt accidentally convert too much data into too much money too early.
The π interpreter is currently available for Linux
and macOS
on x86_64
architecture, but stay tuned - ARM, WASM, RISC-V are in the pipeline, and they will not make you wait.
So, no further ado:
$ npm i @kparc/k -g
As all things π, the development of π itself is happening very fast. New builds are published up to several times a week, so make sure you always use the latest version:
$ alias kup="npm i @kparc/k -g"
$ kup
Start your very first π session like so:
The startup banner packs a lot of useful information:
it says | it means |
---|---|
2019-08-23 12:18:26 | timestamp of your π build |
40core | cpu cores you can use in π |
270gb | max workspace alloc size |
avx512 | the best your cpu can pull off |
shakti | the company behind π |
l | m for macos, l for linux |
prod | your build is test for now |
If it ever comes to that, please always include your banner in your bug reports.
At any time during a π session, you can:
\h
view k reference card
\l
view k change log
\\
quit k session
At this point we highly recommend to avoid issuing any of the above commands, especially the latter.
Practice:
Make sure you have rlwrap
utility installed, and put an alias alias k="rlwrap k"
into your rc file. This makes your spartan π development environment a lot more pleasant to use.
Type in your first π expressions, and enjoy your first answers:
2+2 /simplest face of k is a calculator
4
x:42 /: is assign so x is now the answer
x=x+1 /= is equal, and of course it isn't
0
kcc:+/β /we knew you'd want to try this one
kcc /and what it happens to be is just:
β
Wherever you see β in this document, you are invited to try something on your own.
Indeed, the title of this document seems to make sense to the π interpreter and evaluates to exactly that, and very soon you will easily infer what it actually means. It is quite profound.
As any other language, π expects a programmer to observe certain conventions on coding style and terminology in order to understand the code written by the others and let their own code be understood. While some rules of the house of π are universal, some are not.
Commenting your π code is the best way not to end up coding Java for food, unless you are Arthur Whitney. We dare to assume you are not, so comments start with /
. When used inline, prepend at least one space. Here is an annotated declaration of two variables:
/annotations are your friends
x:42
y:42 /now, always and forever
Character ;
in π is used for one thing and one thing only, to separate π expressions. As you have seen above, π doesn't force you to terminate the line explicitly with ;
because newline is also an expression separator. Separator is used the same way and means the same thing everywhere in any context (except comments), e.g. to separate expressions inside a function body, vector declaration, function arguments, etc. Later we will see that separator is also a part of certain language constructs, but it has the same meaning there as well. But by far the most frequent explicit use of the separator you will encounter in the wild is to separate expressions within one line:
x:1; y:2; z:3 /one line, three expressions
x:1;y:2;z:3 /denser version of the above
This is a tricky subject in π. Basically, what you generally want is no indentation. This means if your π expression is getting so large that you are tempted to split it into separate lines, you likely need to refactor or return to the blackboard. Sometimes, however, indentation is fine and even necessary, and it is always one space. Not two, not four, one. Tabs will be frowned upon because they eat up a lot of space, see below.
Variable names in π follow a somewhat unusual convention. Capitals are used by π programmers very sparingly, which applies both to code and comments. While identifiers in camelCase
can sometimes be tolerated, c_style
identifiers are not permitted at all, since _
is an operator. Identifiers of functions and variables are very often boiled down to an absolute minimum, names 1-3 characters long are commonplace, which does not impact readability given that their definitions are annotated. Short identifiers might sound like a bad idea to Java programmers who are used to see identifiers longer than 2^8 bytes, but, unlike Java, π source requires very little or no scrolling. When the entire program fits in your visual buffer, "cryptic" identifiers are no longer a problem because their annotated declarations are always right in front of you:
kei:42 /kenneth eugene iverson
The subject of code organization will forever remain a major point of controversy in software development. Computer languages have different philosophies on coding style, especially when it comes to three specific keystrokes: \n
, \t
and, less surprisingly, 0x20
. If we define two extremes as "tall, lean, sparse and readable" and "robust, wide, dense and cryptic", then C is a classic example of tlsr
, and π is all the way down rwdc
road.
-
The primary influence of π syntax is the most powerful instrument of human reasoning, a mathematical notation. As such, the language actively encourages the programmer to produce very dense and succinct code, which leads to a common pitfall among beginners who are tempted to sacrifice too much readability. As any good program, a good π program must remain readable and adequately annotated.
-
Comments are integral part of the code and also consume space, so boil them down to some reasonable size as well, but don't get too cryptic.
-
In π, it is possible to minimize code scrolling or even avoid it completely. When the entire program or component fits in your view, you lose no time on navigating it and switching contexts. For example, every code block in this document fits on a laptop screen and remains very readable on mobiles (try that with Java).
-
Syntax highlighting is essential, and poor highlighting is often worse than none β so choose carefully from π syntax packages available for your editor. The best is often the one you wrote yourself, and π syntax is extremely regular and simple.
-
Medium is the message, so we refer you to π code presented in this document. Please send pull requests to help us improve it, and if you like the style, it is yours to have.
Bad form in π is code bloat. Avoid writing extra code if you can β there is too much of it written already into the world. Remove inessential code, yours or not. But if you absolutely have to write more, make it useful, secure, compact, maintainable, portable and scalable.
practice
We don't know much π to practice style yet, so this one will be read-only. Here is a trivial C program formatted in two different ways:
tlsr:
#include <stdio.h>
int
main(int argc, char **argv)
{
int x = 'a';
for(int i=0; i<26; ++i){
putchar(x++);
}
return 0;
}
rwdc:
#include<stdio.h> //k.h
typedef int I;
#define O putchar
#define DO(n,x) {I i=0,_i=(n);for(;i<_i;++i){x;}}
#include"k.h"
I main(){I x='a';DO(26,O(x++))}//nsl
Compare their strengths.
The most important terminology in π revolves around functions. Functions in π are first-class citizens. π has anonymous functions, eval, apply, recursion, and then some. It takes a leap of faith to believe it, but π is probably more lispy than certain Lisps, only you don't need to get past any parens. However, since there are no linked lists under the hood, π is not Lisp, because it was designed to be fast.
This is an uncommon feature, most languages require you to explicitly declare function arguments. Of course you can also do that in π if you want to, but if you don't, a function can have up to three implicit arguments called x
, y
and z
, which means you declare them by simply referencing them in the function body. It is a very convenient feature, not nearly as scary as it sounds:
f:{x+y+z} /f[] takes three arguments
f[1;2;3] /and here is how to call f
6
e:{[a;b;c] /not so implicit, but why?
a+b+c}
e[1;2;3] /f with 9 extra keystrokes
6
f:{x*x} /f[] has only one argument
f 42 /and you can omit brackets
1764
Note that when calling a function with three arguments f[1;2;3]
we had to use square brackets and use an expression separator, because each argument passed to a function is an expression in its own right. However, second function only takes one argument, and we were allowed to omit brackets β although we could also say f[42]
.
This illustrates the core principle of π syntax β almost everything that you intuitively feel you should be able to omit, can and should be omitted. Top candidates for omission are square []
, round brackets ()
and space 0x20
. The less you type, the better your code will get.
Syntax for explicit argument declaration {[a;b]}
is just a side remark. It is good to know, but we won't see it in this text again.
Function rank is another way of saying valence, a fancy word that describes a simple idea that is extremely important to be understood well. Rank of an operator or a function is basically the maximum count of arguments they take. Two functions shown above have ranks of 3 and 1, respectively.
Two specific ranks are so important that they have their own names. A function or an operator that takes...
- one argument is monadic
- two arguments is dyadic
As you will see, the vast majority of native operators in π have exactly two completely different meanings based on the context where they are used, which is in turn defined by the number of arguments offered to the operator.
For example, when you used your first ever π operator in the expression 2+2
, you have used the +
operator in a dyadic context since it received exactly two operands to work on, left and right, so it was inferred to be dyadic x+y plus
. The monadic +x flip
will be introduced later, and has an entirely different meaning.
Also known as function views, projections can be understood as "incomplete" function calls with at least one free, or elided argument. For example, if a function of rank 3, f[x;y;z]
, only receives first and third arguments, it will return its monadic projection, which itself behaves as a function:
f:{x+y+z} /a function of rank three
p:f[1;;3] /a projection of f[1;?;3]
p 2 /is same as call f[1;2;3]
6
Projections are commonplace and very useful. To give one example, projection provides a way for a lambda or a function to capture some of their context, i.e. bind values from the current scope to the local scope of a lambda, wherever it travels next. In other words, if you prefer functional speak, projection creates a closure.
As you already know, the action of a π operator depends on the number of arguments passed to it. However, there are situations when an operator receives two operands (typically, left and right), but is intended to perform a monadic action using right argument only. To disambiguate the rank, the operator can be declared explicitly monadic by appending :
to it:
+ /context-aware, either monadic flip or dyadic plus
+: /always stays a monadic flip, disregarding context
Later on you will see how this works in practice.
You will not get far in this course without a strong grip on the idea that some things in π land are monadic and even explicitly monadic, while others are dyadic. Make sure you have it.
On a more general note, functions in π can be of rank 1 to 9:
- it is not really possible to define a function with no arguments. Rank zero, or niladic functions, do not exist in π.
- a function cannot take more than nine explicit arguments, and some say this is an overly generous limit.
Variable scoping in π is an important aspect of its design. Newcomers often expect to find so called lexical scoping in π β that is, every inner scope has access to all variables defined in all outer scopes, which is how the majority of modern programming languages treats this subject. However, π has a different take on this.
In π, variable visibility is limited to exactly two scopes: local and global.
Take your time to absorb this fact and appreciate its implications. While not easily digested by imperative crowd, the benefits of scope isolation are immediately obvious to functional folks:
-
While π is not purely functional, for as long as π function does not access or modify global state, it remains pure, i.e. free from side effects. Pure functions may behave in a mathematically sound fashion, and can be reasoned about in terms of domain and range, much like their math cousins.
-
Scope isolation relieves the program from an entire class of bugs related to shadowing and naming clashes.
-
Pure functions are best friends with immutability and distributed architectures.
The fundamental benefit of this way of thinking about functions is simplicity and composability. Code blocks that are easy to debug, test, refactor and reuse result in clean, secure and scalable systems.
The last remark on π terminology is of extreme importance. While π is a computer language, its grammar is defined in terms we normally use to denote parts of human speech. That is, π expressions are composed of verbs, nouns and adverbs. For now, let's focus on the former two and consider the following two sentences:
- "shuffle this deck of cards"
- "take three random cards from this deck"
Both are imperatives, where verbs act as the main part of the expression - they identify the action to be taken. However, there is an important difference between the two. In linguistics, the structure of the first sentence is known as verb-only predicate, while the second is a verb-plus-direct-object predicate. In π speak, we recognize the verb "shuffle" in the first sentence to be monadic, while the second is built around a dyadic verb "take".
And this is exactly what π verbs are:
- A verb instructs the interpreter to do something, or answer some question.
- A verb can be monadic or dyadic, i.e. requires one or two nouns to form a grammatically correct phrase.
Since verbs operate on nouns, they are very often called operators, and nouns passed to verbs are said to be their operands or arguments.
For newcomers, the biggest caveat with verbs is whether or not to consider a user-defined function, i.e. any expression enclosed in curly brackets, to be a verb β and the correct answer is no. This might feel counter-intuitive at first, but it helps to think about the content of curly brackets as of an independent expression, i.e. a sentence of arbitrary complexity, which can contain as many verbs as needed, or none at all. For that reason, functions are treated as nouns in π, and while in certain contexts verbs and functions can indeed be used interchangeably, functions differ from operators in the following fundamental way:
- Operators can be applied using both infix and prefix notation
- Functions can only be applied using prefix notation, also known as "functional form"
This only sounds confusing until you see what this means in practice:
a:44;b:2 /a and b are typical nouns: variables holding an number
a-b /a dyadic verb 'subtract', applied using infix notation
42
-[a;b] /the same thing in prefix notation, aka functional form
42
sub:{x-y} /a dyadic noun 'sub', a function assigned to a variable
sub[a;b] /the only way to use 'sub' is to treat it as a function
42
a sub b /putting 'sub' between its operands is not going to work
a sub b
^
class error
We shall revisit parts of π speech again, especially adverbs, and if this quick introduction left you puzzled, for now it is safe to think that verbs are simply "built-in operators".
recap
So far you know how to:
- add two integers
- assign values to variables
- declare and call basic functions
- be friends with
x
,y
andz
- tell monadic and dyadic apart
- explicitly declare monadic ops
- deal with verbs and nouns
- annotate your code
This is a good start, but tells you absolutely nothing about what π really is, and from here things will start to get real.
The word atom
is a synonym for scalar value
, or simply scalar
. Every language has them, and in π they are as useful as elsewhere. But π belongs to a family of vector languages, which means its fundamental type is an ordered set of atoms or other ordered sets.
In π parlance, terms array
, list
and vector
are often used interchangeably and refer to the same thing, but we will stick with vector
to avoid confusing you, because vectors are much more general than classic arrays and have nothing to do with linked lists.
x:(0;1;2;3;4) /one way of declaring an integer vector
y:0 1 2 3 4 /same effect using more informal syntax
x
0 1 2 3 4
y
0 1 2 3 4
a:42 /a scalar variable, an atom with a name
v:,42 /a vector of length 1, one integer item
k is strictly "pass by value", i.e. there are no references or pointers being passed around, although it is an illusion created by the underlying implementation. In reality, k avoids making copies of stuff unless it becomes absolutely necessary:
x:0 1 2 3 4 /everything is passed by value
y:x /so this will make a copy of x
y /y is a clone, not a reference
0 1 2 3 4 /(in reality it only pretends)
y:x:0 1 2 3 4 /same effect in one expression
/y and x are actually the same
/until one of them is modified
The first thing you need to know about vectors is that most operations you expect to work for atoms work equally well for vectors, too:
x:y:0 1 2 3 4 /x and y are twin copies
x+y /pairwise sum of vectors
0 2 4 6 8
x*y /product is pairwise too
0 1 4 9 16
x%y /division is %, get used
ΓΈ 1 1 1 1
0%0 /ΓΈ is nan, void and null
ΓΈ
x=y /compare x to y pairwise
1 1 1 1 1 /1 is truthy, 0 is false
Mixing atomic and vector operands makes total sense and is very useful:
x:0 1 2 3 4
x+1 /increment all elements
1 2 3 4 5
x=1 /compare each of x to 1
0 1 0 0 0
x%0 /divide each by 0, ouch
ΓΈ β β β β /βxββ (x%0) β {-β,β,ΓΈ}, beware (0%0) = ΓΈ, enjoy responsibly
3%x /divide 3 by each of x
β 3 1.5 1 0.75
Indexing is zero-based as you would expect, and if you pass a vector of indices, you get back a vector of items:
x:2 4 8 16 32
x[2] /get 3rd element of x
8
x 2 /no need for brackets
8
x[1 4] /get 2nd and 5th item
4 32 /gives another vector
x 1 4 /no need for brackets
4 32
y:1 4 /y is an index vector
x y /same as x[y] less []
4 32
Pairwise operations on vectors of incompatible shape make much less sense to π than division by zero, and will throw an error:
x:0 1 2 3 4
y:0 1 2
x+y
x+y
^
length error
Note the difference between shape and length. This reminds us that a vector can be composed not just from atoms but from other vectors as well, and there is no practical limit on the depth of nesting. In other words, vectors can have arbitrary shape:
y:(,1;1 1;1 2 1;1 3 3 1) /pascal's triangle
y
,1
1 1
1 2 1
1 3 3 1
Vector arithmetic is penetrating, which means that vector operators apply at depth for as long as operands have compatible shape. It is a good time to introduce the monadic +x
to better see how this works:
mat:(1 2 3;4 5 6;7 8 9) /shall there be mat:
mat /a square matrix 3*3
1 2 3
4 5 6
7 8 9
tam:+mat /monadic + is 'flip'
tam /y is a transposed x
1 4 7
2 5 8
3 6 9
mat=tam
1 0 0
0 1 0
0 0 1
mat>tam
0 0 0
1 0 0
1 1 0
mat+42
43 44 45
46 47 48
49 50 51
practice
First, lets make sure +x flip
operator transposes rectangular matrices just as well as squares, which would be of little surprise. Then try to flip something less obvious, and after that you have two more transformations to apply. Inspect all intermediate results and make sure you follow their logic:
mat:(1 2 3 4;6 7 8 9) /a rectangular matrix
+mat /flip it, no big deal
1 6
2 7
3 8
4 9
t:(,1;1 1;1 2 1;1 3 3 1) /t is triangle vector
t:+t /flip it, what gives?
t:t>0 /t greater than zero?
t:(+t)+t /t transposed plus t?
β
No rocket science, all pretty basic, but carry on.
Type system in π gets strict when it has to, but also agrees that implicit casts and type coercion have their strengths β especially when done right, which in π they are.
Before we see the examples, the first thing you need to know about types in π is that they are divided into two broad classes: vector types and atomic types. That is, a vector with a single element, say, 42
, is not the same type as an atomic integer of the same value. Finally, since functions and other things in π are also assignable values, they also have their place in the type system. Those are special types and we will not cover them here in much detail.
Here is a quick overview of basic π types and their symbolic names:
`i int
`f float
`j long
`c char
`n name
`D date
`t time
`a dict
`A table
`. vect
This is not very revealing, so lets see them in action. The operator to query the type of anything in π is monadic @x
, and if you are not sure what we mean by monadic
, perhaps it is a good time to start over.
@42 /int scalar
`i
@424242j /int64 atom
`j
@.5 /float atom
`f
@0 1 2 /int vector
`i
v:0 1 .5 2
@v /0.5 promotes vector to float
`f
v 1 /2nd item, f is short for 1.0
1f
Like in C, there is no dedicated type for strings in π. Strings are just char vectors:
@"k" /"k" is char atom
`c
s:"kei";@s /s is char vector
`c
s 0 /1st element of s
"k"
A type called name is the same idea as internalized string found in some other languages. This means that a single instance of an arbitrarily long string can be placed into a global hash table that persists for a lifetime of a π process and can later be referenced by its hash key as many times as necessary without creating additional copies of the string.
We could say that in case of names π actually passes references instead of values, but they are not true pointers and there is no arithmetic defined for them. Names come handy in many situations, for now lets just see how they quack:
a:`kei /"kei" is now internalized
@a /name atom is its hash key
`n
b:`kei`kei`kei /three references to "kei"
@b /vector of string pointers
`n
@`"ken iverson" /spaces in names, why not?
`n
$a /$ converts names to chars
"kei"
Temporal types in k are date
and time
:
d:1981-02-01 /yyyy-mm-dd, ok to expect iso 8601 compliance
@d /NOTE: date atom is `D same as date vector `D
`D
t:12:34:56.789 /hh:mm:ss.sss, max resolution is milliseconds
@t
`t
dt:d+t;dt /advanced date ops is a story for another day
1981-02-01T12:34:56.789
Dictionaries are maps of keys to values, also known as hashmaps or, more generally, associative arrays. They are as useful in π as elsewhere, but unlike many languages where this data structure is built in, π allows both keys and values to be of any type, both vector and scalar. This might sound a bit confusing at first β since nothing prevents you from, say, constructing a dictionary where keys are themselves dictionaries β but in able hands this oddity becomes a very powerful tool.
Dictionaries have the type `a, and the notation for defining them uses a dyadic operator ! key
:
d:`a`b!(1 2 3;4 5 6) /keys!values
d
a|1 2 3
b|4 5 6
@d /dict type is `a
`a
d`a /key lookup (aka d[`a] or d `a)
1 2 3
!d /dict keys
`a`b
. d /dict vals (note the space!)
1 2 3
4 5 6
Tables are flipped dictionaries, and they require a separate large discussion. Here, we will only describe their syntax for the sake of completeness. Table type is `A, and notation is the same as dict, only with +x flip
operator prepended. As common sense implies, a dictionary won't flip unless all values are of the same length.
No comments on any of this for now β but if you can follow the logic of what is going on here, you'll agree that in some rare cuircumstances a technology is indeed indistinguishable from magic. See for yourself:
x:`goo`apl`amz
y:1 2 3 4 5 6
t:+`s`d`p!(x,x;`D$y;1.5*y) /table is a transposed dict
@t
`A
t /trades: stock, date, price
s d p
--- ---------- ---
goo 2024-01-02 1.5
apl 2024-01-03 3
amz 2024-01-04 4.5
goo 2024-01-05 6
apl 2024-01-06 7.5
amz 2024-01-07 9
select avg p by s from t /pretend you didn't see this
s |p
---|----
amz|6.75
apl|5.25
goo|3.75
Lambdas are assignable values and have their own type, and more than one:
@{x+y} /type of lambda gives away its rank
`2
nil:{0} /don't expect nil to have rank zero
@nil /niladic functions don't exist in k
`1
a:{x},{x+x},{x*x} /a vector of lambdas, sure, why not
a[2]16 /calls 3rd lambda, same as a[2][16]
256
Null values in π are typed, integer null is Γ
and float null is ΓΈ
. Infinity is a scalar float β
. Working with nulls and infinities can be very tricky, and it is very important to pay attention to their types:
n:ΓΈ /float null is type float
@n
`f
N:Γ /int null is type integer
@N
`i
n=N /equal, but not the same!
1
@β /infinity is a float atom
`f
It is evident that nulls and infinities are Unicode glyphs. Although it is very easy to set up keyboard shortcuts for them, there are idiomatic ways to enter them using plain ASCII (but avoid if you can):
0%0 /float null is zero div by zero
ΓΈ
42e /missing exponent is float null
ΓΈ
`i$0%0 /int null is rounded float null
Γ
_0e /int null via `monadic _x floor'
Γ
1%0 /inf is just reciprocal of zero
β
%0 /inf via `%x 'inverse` operator
β
PI:3.14159265358979323846264338327950288
Ο=PI /synonym of Ο in ascii, but why
1
Composite vector type, or you could also say mix vector, is of special mention. Such vectors are either a mixture of atoms of disparate types, or contain something more complex than atoms, e.g. other vectors:
c:0,1,"a",2,3 /a char impostor among ints, c is mix
@c /type of a mixed list is backtick-dot
`.
x:(1 2 3;4 5 6;7 8 9) /a vector of vectors is composite too
x
1 2 3
4 5 6
7 8 9
@x
`.
Type casting, both explicit and implicit, is demonstrated by the following examples which also give a general feel of how type coercion behaves. The cast
operator in k is a dyadic t$x
, where t
is a type name and x
is a subject of cast:
1+.5 /int plus float is float, no surprises here
1.5
1f*2 /any float operand promotes result to float
2f
`i$42.99 /explicit cast from `f to `i drops mantissa
42
`i$42.0 42.99 /`f to `i will round down the entire vector
42 42
0+"abc" /integer operand demotes `c vector to ascii
97 98 99
`c$1+"HAL9000" /add 1 to ascii, cast back to `c, surprise:
"IBM:111"
"012"+"345" /sum ascii codes of chars, result stays `c
"ceg"
a:1 2 3 /lets start with a nice uniform int vector
a[0]:1f /now, replace its 1st element with a float
@a /whoops, our int vector got demoted to mix
`.
a:`f$a;@a /explicit cast to float solves the problem
`f
`i$1981-02-01 /dates represented as ints look like this:
-15674
15674+1981-02-01 /and they are simply offsets in days from:
2024-01-01
`i$2024-01-01 /so the epoch date itself must be exactly:
0
1+`kei /no math for names, this type is immutable
^
type error
@@42 /what is type name of a type name of int?
β
There are things left to be said about the type system, but the expression @@42
above (which evaluates to some kind of wordplay, type name of a type name is name
) urges us to the next section which is all about how to make sense of this expression.
As you must have noticed, the syntax for indexing vectors and calling functions is identical:
l:{x+x} /some monadic function l[x]
t:2 4 8 16 /some random integer vector
r:0 3 /index vector: indices of t
l[t] /apply function l to each t
4 8 16 32
t[r] /items of t at indexes in r
2 16
l[t[r]] /compose: apply l to t at r
4 32
What we also know that π actively encourages us to omit brackets whenever possible, so lets do exactly that:
l t r /exactly the same as l[t[r]]
4 32
And here is comes: once we drop the brackets, it suddenly becomes absolutely natural to read this expression right to left. Take your time to contemplate and absorb this fact. In very little time you will see how it works in practice, and once you put it to practice yourself, you will agree that this way of functional composition is simple, elegant and intuitive:
k expressions are read, written and evaluated right to left.
But when we say "expressions" we don't mean "programs", and this is a very important distinction:
k programs are read, written and evaluated left to right.
This might sound confusing, but look at the diagram of a small π program that consists of three identical expressions l t r
, with parens added for clarity. Further down is the order of evaluation of the entire program, which leaves no room for confusion:
/ L T R
/(l t r);(l t r);(l t r)
/ I > II > III
/(3 2 1);(6 5 4);(9 8 7)
And now that we know the way the rivers flow in π land, we are equipped to discuss the next subject.
Very early on in our lives, we are taught there is a good reason for multiplication and division to bind stronger than addition and subtraction, so they must be computed first. Later on, we are taught that most computer languages require us to learn a lot more complex systems of operator precedence to do anything useful with them to begin with, and much later on β all the deadly caveats hidden in those systems. But in π, the question of operator precedence is fully answered by their order of evaluation, so buckle up. Here is the one and only rule ever to be learned about precedence in π:
There is no operator precedence in π, unless it is explicitly defined by round brackets.
By default, all operators in a π expression are treated equally and evaluated strictly from right to left. Obviously, this includes arithmetic operators as well, and before you ask yourself how *
can possibly have no precedence over +
, consder a few basic math expressions annotated with their order of evaluation:
3+2+1 /"take 1, add 2, add 3"
6
3*2+1 /"take 1, add 2, multiply by 3"
9
(3*2)-1 /"take 2, multiply by 3, sub 1"
5
-1+3*2 /same thing, without overrides
5
As you see, it is much easier to get used to the lack of precedence than it appears at first β indeed, simplicity is the worst enemy of complexity. The last two examples demonstrate the basic strategy of avoiding parens entirely, and there is a good reason for ditching them β it makes the order of evaluation completely linear.
Although precedence overrides are sometimes inevitable and can be beneficial, they have an adverse effect on readability. Basically, while reading a π expression, what you generally want is to go fast and uninterrupted β and precedence overrides interrupt the natural flow of comprehension. Good π programmers think of others before themseves, and seek to produce code which follows the natural order of evaluation by minimizing the use of round brackets.
Now we can revisit the last expression from the previous chapter:
@@42 /"type name of a type name of 42" actually reads backwards:
`n /"get 42, apply monadic @, get `i, apply monadic @, get `n"
A convincing proof that type name of a type name in π is indeed a name
.β
practice
Let's revisit the code from the first snippet in this document:
x:(1 2 3;4 5 6;7 8 9)
x=x+1 /how can a universal truth be so false?
β
Too easy, but we'll make up for it.
This part might be easier to digest than the previous, especially if you are familiar with functional programming. The title, borrowed without permission from the legendary k resource, says it all - you will not find a π construct that resembles an explicit for
loop, and although there is a while
construct in π, it is almost never used in practice.
And this is not just to avoid untold damages from trivial errors people keep making in their loop definitions. The main reason explicit loops are banned from π is because it offers something better. The idea that displaces them is a simple and strong abstraction called adverbs, but before we see them in action, it helps to understand why they are called that way:
An adverb is a modifier that accepts a user-defined function or a native operator and returns a new monadic or dyadic verb which acts on one or two input operands in some desirable way to produce an output. Input and output can be scalar values or vectors, depending on the adverb and the action it modifies.
The formal definition sounds a bit dry, so let's consider a classic example of how adverbs replace loops:
"Compute a sum of elements of a given integer vector."
Thinking in implicit loops suggests something we've all done a million times:
int sum(int[]in){
int i=0,acc=0;
for(;i<sizeof(in);++i)
acc+=in[i];
return acc;}
But what if we could state the problem to a computer like this:
"Put a +
between all adjacent items and give me the grand total."
And that is the simplest way to describe what π adverb over
does when it is used to modify dyadic +
. Only over
, as all other adverbs, is general, and will happily modify any dyadic operator or function. Described more formally, over
looks more like this pseudocode:
- if
x
is an atom, returnx
- set
acc
to 0 (a.k.a. accumulator) - while
next x
, setacc
to the result ofv[acc;next x]
- return
acc
The above is nothing else but a general case of the explicit loop found in sum()
, as well as of all other explicit loops of this particular family. In functional speak, one would say adverb over
folds a vector of values and reduces them into a scalar.
And since over
is just v/x
π, this is how sum
function looks like:
s:{+/x} /s is 'plus over x'
s 1 2 3 4 5
15
It is a good moment to look back at the C version, one last time β and be surprised to hear that its for
loop declaration contains an ancient, but ever so popular bug, which π version does not because spotting bugs in +/x
is considerably easier. Besides, even if the C code wasn't broken, it would only work for 32-bit integers.
You could be tempted to see of what other use over
could be. Let's introduce a new k operator, !x til
, and implement another obvious candidate for over
:
x:!9 /! is til, get first n integers
x /tada, we have all ints up to 8
0 1 2 3 4 5 6 7 8
fact:{*/1j+!x} /fact x 'mul over 1 plus til x'
fact 20
2432902008176640000j
Now that we have parted ways with loops, and discussed over
in detail, it is time to meet the rest of six π adverbs. Please welcome the magnificent six, and note that only most trivial use cases are shown:
adverb over is f/x
adverb scan is f\x
where f
is a dyadic
verb and x
is an input vector
a:0 1 2 3 4 /some data
+/a /puts a + between items, 0+1+2+3+4
10 /and returns the final result only
+\a /scan returns intermediate results
0 1 3 6 10 /running sum, aka debugger of over
adverb each is f'x
where f
is a monadic
verb and x
is an input vector
d:42,"a",`kei /some composite vector
fn:{@x} /some monadic function
fn'd /apply fn to each of d
`i`c`n /the type of each item
f:{1%x},{x*x} /reciprocal and square
{x 25}'f /call each of f for 25
0.04 625
adverb eachleft is x f\:y
adverb eachright is x f/:y
where f
is a dyadic
verb and x
and y
are left and right inputs,
either vectors or atoms
10 20 30-\:5 /eachleft does (10-5),(20-5),(30-5)
5 15 25 /each of left, substracted by right
5-/:0 20 30 /eachright does (5-0),(5-20),(5-30)
5 -15 -25 /left, substracted by each of right
adverb eachprior is x f':y
and (f':)x
first form is seeded eachprior
where f
is a dyadic
verb, and x
is a
seed value and y
is an input vector
second form is seedless eachprior
where f
is a dyadic
verb, and x
is
an input vector
2+':4 8 16 /seeded eachprior gives (2+4),(4+8),(8+16)
6 12 24 /sum 1st item and seed, then sum each item and its prior
(+':)4 8 16 /seedless eachprior gives (4),(4+8),(8+16)
4 12 24 /first item stays as is, then sum each item and its prior
This doesn't seem like much, adverbs seem to be doing pretty basic stuff. But hold that thought for a minute.
recap
We have seen:
- what π type system looks like
- how basic vector and atom math works
- which way to read and comprehend π code
- what is the only existing precedence rule
- why there is no
for
and why there areadverbs
practice
Consider a slightly less trivial example of eachleft
working together with x*y multiply
on two vector operands:
x:!9 /a vector of integers from 0 til 8
x+:1;x /add 1 to each of x, same as x:x+1
1 2 3 4 5 6 7 8 9
x*\:x /what would be x times eachleft x?
β
Bonus questions:
kcc:+/β /how come k sums up to infinity this fast?
kcc
β
+\(1+!42)%0 /how about a running sum of 42 infinities?
β
Make sure you can follow the logic of these examples before advancing to the next chapter, where things will get a lot less innocent. But if you didn't find the exercise challenging enough, here is another one:
Extra bonus:
q:7;GF:!q;q\GF*\:GF /what does GF stand for?
β
/`c$97+2/:+2\:32 458 1172 1443 275
The title of this chapter is borrowed from a legendary book published in 1945, a small volume by mathematician George PΓ³lya where he shows how to tackle problems and arrive to solutions. It is a very inspiring read.
Let's tackle a little problem. We will look at a π function that actually does something very useful and implements a familiar algorithm. The subject of the game is to figure out how it is implemented in π and to identify the algorithm. It is very useful to dissect all of it on paper, so put your interpreter aside for now.
So, here is the code:
/what is f[], and can we read it?
f:{$[2>#?x;x;,/f'x@&:'~:\x<*1?x]}
This little monster is deliberately designed to make as little sense as possible at first glance, but once we take it apart, we hope you'll agree it is actually very simple and readable:
f:{$[2>#?x;x;,/f'x@&:'~:\x<*1?x]}
f:{...} /f is a function, that is a good start
f:{.x.} /f takes only one implicit argument, x
f:{.f.} /f clearly calls f, so it is recursive
$[?;?;?] /the entire body of f is nothing else but
$[c;t;f] /a ctf cond, aka if-then-else aka ternary
2>#?x /c: boolean condition
x /t: do this if c is 1
,/f'x@&:'~:\x<*1?x /f: do that if c is 0
2>#?x /we don't how to read this, but it is clear that f[]
/halts recursion when it evaluates to 1, returning x
/so lets find out what it means, going right to left
/we have two new operators, both in monadic context:
/monadic ?x is 'distinct' unique elements of x
/monadic #x is 'count' count the elements of x
#?x /'count distinct' count unique elements of x
2>#?x /'greater' true if count distinct x is less than 2
$[2>#?x;x;...] /"if x has <2 unique items, return x, otherwise ..."
Coffee break, here is what we know so far:
- the function is recursive
- its overall control flow
- the condition that stops recursion
This gives us confidence to wrestle down the last part, the recursion step:
,/f'x@&:'~:\x<*1?x /this must be the recursion step, read right to left:
x:4 0 1 2 /a small dataset to help us see what is going on here
rnd:1?x /x?y is 'find': picks x random elements from vector y
rnd /list with one random item from x
,2
rnd:*rnd;rnd /*x is 'first': return the head element of a vector x
2
cmp:x<rnd /bool vector of 1s where x[n]<rnd, 0s where otherwise
cmp
0 1 1 0
/~x is 'not': boolean Β¬x, non-0 turns 0, all 0 turn 1
mask:~:\cmp /explicit 'not' scan: returns cmp and negation of cmp
mask
0 1 1 0
1 0 0 1
/&x is 'where': return indices of x where x are not 0
idx:&:'mask /explicit 'where' each: apply 'where' to each of mask
idx
1 2 /mask[0] has 1s at indices 1 and 2
0 3 /mask[1] has 1s at indices 0 and 3
pts:x@idx /dyadic @ is 'index': elements of x at indices in idx
pts
0 1 /list of items in x less than pivot rnd
4 2 /list of items in x greater or equal to pivot rnd
pts:x@&:'~:\x<*1?x /"partition x by pivot: items < rnd and items >= rnd"
x:f'pts /adverb each: apply f to each partition, recurse down
,/x /,/ is 'raze': unnest aka flatten a vector of vectors
,/(1 2 3;4 5 6) /in other words, raze flattens first level of nesting
1 2 3 4 5 6
Note that it is the first time we have seen rank override in action. In both cases, explicit monadics are passed to an adverb β and this is a very typical use case.
Now that we know what every specific part does, we can zoom out and see the big picture. Feel free to use the interpreter to play around and test your ideas.
And of course, f
is nothing else but:
qs:{$[2>#?x;x;,/qs'x@&:'~:\x<*1?x]} /qsort on rand pivot
i:9 2 5 5 1 8 1 3 6 1 /a hairy int shuffle
f:2.6 -β 8.6 Ο 1.7 β 3.5 5.6 /a Ο in a float soup
c:"edrofgtnljgrpliifp" /a char entropy pool
mess:(i;f;c)
qs'mess /restore order quick
β
And of course this is not the quickest quicksort
ever written, but this is just a toy. In real life you would simply use the built-in operator which is a lot more efficient:
^3 2 1 /monadic ^x is 'sort'
1 2 3
sort:^:' /explicit 'sort each'
(qs'mess)~(sort mess) /x~y 'match' operands
1
\t:10000 qs'mess /apply qs 10000 times
β
\t:10000 sort mess /fix 10000x more mess
β
As you see, native sort is incomparably faster. But what our DIY sort function is very good for is to demonstrate the principle of doing more with less, and that is what π is all about.
Check out examples of quicksort
in the wild in C++, Python, JavaScript and Java.
recap
Previously we have seen:
- monadic
!x til (first x integers)
- monadic
%x inverse (reciprocal)
- monadic
+x flip (transpose)
- monadic
$x string
- monadic
@x type
- monadic
^x sort
- dyadic
t@x cast
- dyadic
x:y assign
- dyadic
x=y equal
And qs
code brought a few more:
- ctf cond
$[c;t;f]
- monadic
?x distinct
- monadic
#x count
- monadic `*x first
- monadic
~x not
- monadic
&x where
- monadic
,/x raze
- dyadic
x@y index
- dyadic
x~y match
- dyadic
x?y find
Finally, you are now equipped with the most ubiquitous system routine:
\t:n expr
benchpress an expressionn
times, result is in milliseconds
Although this is still a small part of π operator arsenal, if you can do quicksort
with this much, you can do a lot more. And then add vector arithmetic, and then take everything to the power of six adverbs.
practice
- take another good look at the code of
qs
function - retrace the steps of the code analysis
- in a new π session, reproduce
qs
from scratch
It sounds much harder than it really is. It might take more than one attempt, but you will be amazed how fast you will get there. However, before advancing to the next chapter, make sure that you do.
Among other things, the annotated breakdown of qs
code gives a good impression of what is typically going on inside of π programmer's head β but tells you nothing about how fast it usually happens. A proficient π programmer would read and understand qs
in well under two minutes. With a bit more practice, you will agree that reading π programs is easy and fun.
It is time to write our first π program, and this time around there will be a lot less hand-holding. We will solve the classic Project Euler p18, also known as p67:
By starting at the top of the triangle below and
moving to adjacent numbers on the row below, the
maximum total from top to bottom is 23:
β3
β7 4
2 β4 6
8 5 β9 3
------------------
9 + 4 + 7 + 3 = 23
Problems 18 and 67 are simply two bigger triangles, and the challenge is to find the sum of maximum paths in them. While 18 can be solved by bruteforce, 67 can not, but the efficient algorithm is absolutely trivial. It is given away in the example above: we simply need to fold rows going bottom up, like so:
8 5 9 3
8 9 9 /max
+ + + /sum
2 4 6
10 13 15 /out
13 15 /max
+ + /sum
7 4
20 19 /out
20 /max
+ /sum
3
23 /out
It is easy to see that the key to the solution is a function that reduces the current row (max
) and merges it into the next (sum
). It expects two arguments, i.e. both rows to work with, and returns out
. That's all we need to implement it:
r4:8 5 9 3 /take two bottom rows to assist thinking
r3:2 4 6
0|42 /x|y is max: returns largest of operands
42
1_1 2 3 /x_y is drop: discards x items of a list
2 3
|':r4 /max eachprior: get max element pairwise
8 8 9 9
1_|':r4 /drop first result of seedless eachprior
r3+1_|':r4 /pairwise sum: merge bottom row into top
10 13 15
row:{y+1_|':x} /row reduction operation, x bottom, y up
row[r4;r3] /r4 and r3 are fine, so should be others
10 13 15
Great, we have the reduction function, now let's apply it over the test triangle to make sure it folds it into what we expect. Since we are reducing a vector into an atom, it is abundantly clear which adverb we want to use:
t:(,3;7 4;2 4 6;8 5 9 3)
|t /monadic |x is reverse the order of vector
8 5 9 3
2 4 6
7 4
,3
row/|t /apply row reductor over reversed triangle
,23
*row/|t /*x is first: simply return the first item
23
mxpath:{*{y+1_|':x}/|x} /maximum path in triangle vector
mxpath t
23
Looks like the mxpath
is doing pretty well. Let's fetch the input file for the problem 67, load it, parse it and solve it:
/backslash cmd executes an os command directly from k:
\curl https://projecteuler.net/project/resources/p067_triangle.txt > p67.txt
lines:0:"p67.txt" /0:x reads a text file as vector of lines
lines 2 /just to make sure we have something real
"52 40 09"
t:`k?'lines /parse each line of input as k expression
t 2
52 40 9
#t /a bigger triangle, but not a bigger deal
100
mxpath t /returns max path sum, the solution to 67
β
We didn't tell you do this, but the complete program can also be written as a single π expression:
*{y+1_|':x}/|`k?'0:"p67.txt" /load, parse and fold
β
recap
We have seen some new stuff:
- shell
\cmd
- dyadic
x|y max
- dyadic
x_y drop
- monadic
|x reverse
- monadic
0:x load text
practice
-
Reproduce
mxpath
from scratch in a new π session, same way you did withqs
. -
Find a way to load and parse the triangle from problem 18, and solve it using your own code.
-
Verify solutions for 18 and 67 on Project Euler.
-
As you know, once you provide a correct answer to an Euler problem, you can browse its discussion forum. You might want to check out some other solutions to 18 and 67 in other computer languages.
-
Measure
\t:n
for 100, 1000, 10000 runs and estimate the time complexity of the algorithm. Make sure disk I/O is excluded from the measurement.
There is no new material in this small chapter, so we can go straight to practice.
practice
Many things in life can only be understood in comparison. Compare functionality of these two programs:
package com.less.with.more.doing.sort;
public final class qs{public void s(int[] x){}}
qs:{$[2>#?x;x;,/qs'x@&:'~:\x<*1?x]}
Now, compare the source code of these two:
import java.util.Arrays;
public class S{public static void main(String[] a){
int[] x={5,4,3,2,1};Arrays.sort(x);
System.out.printf("%s",Arrays.toString(x));}}
$ echo "^5 4 3 2 1" | k
1 2 3 4 5
Finally, compare the size of their runtimes:
252M May 17 13:59 jdk-8u211-macosx-x64.dmg
79M May 17 13:50 jre-8u211-macosx-x64.dmg
109K May 17 13:54 k.tgz
We have covered a lot of ground, good time to put things into perspective. Below is a complete map of π operators, and those marked with bullets you have already seen and used at least once:
x+y +x
-------------------------
: β assign
+ β add β flip
- β subtract β negate
* β multiply β first
% β divideby β inverse
& β¦ min|and β where
| β max|or β reverse
< β less β¦ up
> β more β¦ down
= β equal β¦ group
~ β match β not
! β key β enum
, β catenate β list
^ β¦ except β sort
# β¦ take β count
_ β drop β floor
? β find β distinct
@ β index β type
. β¦ apply β value
$ β pad|cast β string
It really feels like we have explored more than we didn't, and it is huge progress indeed. But many things remain to be discovered, because operators is only one aspect of π β and this short introduction could not possibly cover everything.
We conclude with a list of subjects that you are now ready to explore on your own:
π language | π platform |
---|---|
additional π operators | debugging and securing π systems |
tables and k-sql language | building clients and servers in π |
vector aggregates | benchmarking, testing and tracing |
entropy sources, math primitives | disk i/o, persistence and streaming |
advanced use of adverbs, threads | ipc and distributed workloads |
native csv, tsv, json and utf | fault tolerance and monitoring |
integrated cryptography core | scripting, deployment, os tuning |
nanosecond time, datetime math | interop with python, julia and c |
π-expressions, `0 | tech support and user community |
design of internal components | π resources, tools and packages |
Although ee cummings opened his famous poem with words somewhere i have never travelled, it seems that some π programmers prefer to read poetry backwards. That explains a lot about the title of our final chapter.
dream a little bigger β