diff --git a/blog/we-have-go-2.markdown b/blog/we-have-go-2.markdown
new file mode 100644
index 0000000..fef399d
--- /dev/null
+++ b/blog/we-have-go-2.markdown
@@ -0,0 +1,896 @@
+---
+title: We Already Have Go 2
+date: 2022-05-25
+tags:
+ - golang
+ - generics
+ - context
+ - modules
+---
+
+I've been using Go since Go 1.4. Since I started using Go then (2014-2015 ish),
+I’ve seen the language evolve significantly. The Go I write today is roughly the
+same Go as the Go I wrote back when I was still learning the language, but the
+toolchain has changed in ways that make it so much nicer in practice. Here are
+the biggest things that changed how I use Go on a regular basis:
+
+* The compiler rewrite in Go
+* Go modules
+* The context package
+* Generics
+
+This is a good thing. Go has had a lot of people use it. My career would not
+exist in its current form without Go. My time in the Go community has been
+_catalytic_ to my career goals and it’s made me into the professional I am
+today. Without having met the people I did in the Go slack, I would probably not
+have gotten as lucky as I have as consistently as I have.
+
+Releasing a "Go 2" has become a philosophical and political challenge due to the
+forces that be. "Go 2" has kind of gotten the feeling of "this is never going to
+happen, is it?" with how the political forces within and without the Go team are
+functioning. They seem to have been incrementally releasing new features and
+using version gating in `go.mod` to make it easier on people instead of a big
+release with breaking changes all over the standard library.
+
+This is pretty great and I am well in favour of this approach, but with all of
+the changes that have built up there really should be a Go 2 by this point. If
+only to make no significant changes and tag what we have today as Go 2.
+
+Take everything I say here with a grain
+of salt the size of east Texas. I am not an expert in programming language
+design and I do not pretend to be one on TV. I am also not a member of the Go
+team nor do I pretend to be one or see myself becoming one in the
+future.
If you are on the Go team and think that something I said
+here is demonstrably wrong, please [contact me](/contact) so I can correct it. I
+have tried to contain my personal feelings or observations about things to these
+conversation snippets.
+
+This is a look back at the huge progress that has been made since Go 1 released
+and what I'd consider to be the headline features of Go 2.
+This is a whirlwind tour of the huge progress in improvement to the Go compiler,
+toolchain, and standard library, including what I'd consider to be the headline
+features of Go 2. I highly encourage you read this fairly large post in chunks
+because it will feel like _a lot_ if you read it all at once.
+
+## The Compiler Rewrite in Go
+
+When the Go compiler was first written, it was written in C because the core Go
+team has a background in Plan 9 and C was its lingua franca. However as a result
+of either it being written in C or the design around all the tools it was
+shelling out to, it wasn’t easy to cross compile Go programs. If you were
+building windows programs on a Mac you needed to do a separate install of Go
+from source with other targets enabled. This worked, but it wasn’t the default
+and eventually the Go compiler rewrite in Go changed this so that Go could cross
+compile natively with no extra effort required.
+
+This has been such an amazingly productive
+part of the Go toolchain that I was shocked that Go didn’t have this out of the
+gate at version 1. Most people that use Go today don’t know that there was a
+point where Go didn’t have the easy to use cross-compiling superpower it
+currently has, and I think that is a more sure marker of success than anything
+else.
+
+The cross compliation powers are why
+Tailscale uses Go so extensively throughout its core product. Every Tailscale
+client is built on the same Go source tree and everything is in lockstep with
+eachother, provided people actually update their apps. This kind of thing would
+be at the least impossible or at the most very difficult in other languages like
+Rust or C++.
+
+This one feature is probably at the heart of more CI flows, debian package
+releases and other workflows than we can know. It's really hard to understate
+how simple this kind of thing makes distributing software for other
+architectures, especially given that macOS has just switched over to aarch64
+CPUs.
+
+Having the compiler be self-hosting does end up causing a minor amount of
+grief for people wanting to bootstrap a Go compiler from absolute source code
+on a new Linux distribtion (and slightly more after the minimum Go compiler
+version to compile Go will be raised to Go 1.17 with the release of Go 1.19
+in about 6 months from the time of this post being written). This isn't too
+big of a practical issue given how fast the compiler builds, but it is a
+nonzero amount of work. The bootstrapping can be made simpler with
+[gccgo](https://gcc.gnu.org/onlinedocs/gccgo/), a GCC frontend that is mostly
+compatible with the semantics and user experience of the Go compiler that
+Google makes.
+
+Another key thing porting the compiler to Go unlocks is the ability to compile
+Go packages in parallel. Back when the compiler was written in C, the main point
+of parallelism was the fact that each Go package was compiled in parallel. This
+lead to people splitting up bigger packages into smaller sub-packages in order
+to speedhack the compiler. Having the compiler be written in Go means that the
+compiler can take advantage of Go features like its dead-simple concurrency
+primitives to spread the load out across all the cores on the machine.
+
+The Go compiler is fast sure, but
+over a certain point having each package be compiled in a single-threaded manner
+adds up and can make build times slow. This was a lot worse when things like the
+AWS, GCP and Kubernetes client libraries had everything in one big package.
+Building those packages could take minutes, which is very long in Go
+time.
+
+## Go Modules
+
+In Go's dependency model, you have a folder that contains all your Go code
+called the `GOPATH`. The `GOPATH` has a few top level folders that have a
+well-known meaning in the Go ecosystem:
+
+* bin: binary files made by `go install` or `go get` go here
+* pkg: intermediate compiler state goes here
+* src: Go packages go here
+
+`GOPATH` has one major advantage: it is ruthlessly easy to understand the
+correlation between the packages you import in your code to their locations on
+disk.
+
+If you need to see what `within.website/ln` is doing, you go to
+`GOPATH/src/within.website/ln`. The files you are looking for are somewhere in
+there. You don’t have to really understand how the package manager works (mostly
+because there isn’t one). If you want to hack something up you just go to the
+folder and add the changes you want to see.
+
+You can delete all of the intermediate compiler state easily in one fell swoop.
+Just delete the `pkg` folder and poof, it’s all gone. This was great when you
+needed to free up a bunch of disk space really quickly because over months the
+small amount of incremental compiler state can really add up.
+
+The Go compiler would fetch any missing packages from the internet at build time
+so things Just Worked™️. This makes it utterly trivial to check out a project and
+then build/run it. That combined with `go get` to automatically just figure
+things out and install them made installing programs written in Go so easy that
+it’s almost magic. This combined with Go's preference for making static binaries
+as much as possible meant that even if the user didn't have Go installed you could
+easily make a package to hand off to your users.
+
+The GOPATH was conceptually simple to reason about. Go code goes in the GOPATH. The
+best place for it was in the GOPATH. There's no reason to put it anywhere else.
+Everything was organized into its place and it was lovely.
+
+This wasn’t perfect though. There were notable flaws in this setup that were
+easy to run into in practice:
+
+* There wasn't a good way to make sure that everyone was using the _same copies_
+ of every library. People did add vendoring tools later to check that everyone
+ was using the same copies of every package, but this also introduced problems
+ when one project used one version of a dependency and another project used
+ another in ways that were mutually incompatible.
+* The process to get the newest version of a dependency was to grab the latest
+ commit off of the default branch of that git repo. There was support for SVN,
+ mercurial and fossil, but in practice Git was the most used one so it’s almost
+ not worth mentioning the other version control systems. This also left you at
+ the mercy of other random people having good code security sense and required
+ you to audit your dependencies, but this is fairly standard across ecosystems.
+* Dependency names were case sensitive on Linux but not on Windows or macOS.
+ Arguably this is a "Windows and macOS are broken for backwards compatibility
+ reasons" thing, but this did bite me at random times without warning.
+* If the wrong random people deleted their GitHub repos, there's a chance your
+ builds could break unless your GOPATH had the packages in it already. Then you
+ could share that with your coworkers or the build machine somehow, maybe even
+ upload those packages to a git repository to soft-fork it.
+* The default location for the GOPATH created a folder in your home directory.
+
+Yeah, yeah, this default was added later
+but still people complained about having to put the GOPATH somewhere at first.
+Having to choose a place to put all the Go code they would use seemed like a big
+choice that people really wanted solid guidance and defaults on. After a while
+they changed this to default to `~/go` (with an easy to use command to influence
+the defaults without having to set an environment variable). I don't personally
+understand the arguments people have for wanting to keep their home directory
+"clean", but their preferences are valid regardless.
+
+Overall I think GOPATH was a net good thing for Go. It had its downsides, but as
+far as these things go it was a very opinionated place to start from. This is
+something typical to Go (much to people's arguments), but the main thing that it
+focused on was making Go conceptually simple. There's not a lot going on there.
+You have code in the folder and then that's where the Go compiler looks for
+other code. It's a very lightweight approach to things that a lot of other
+languages could learn a lot from. It's great for monorepos because it basically
+treats all your Go code as one big monorepo. So many other languages don’t
+really translate well to working in a monorepo context like Go does.
+
+### Vendoring
+
+That making sure everyone had the same versions of everything problem ended up
+becoming a big problem in practice. I'm assuming that the original intent of the
+GOPATH was to be similar to how Google's internal monorepo worked, where
+everyone clones and deals with the entire GOPATH in source control. You'd then
+have to do GOPATH juggling between monorepos, but the intent was to have
+everything in one big monorepo anyways, so this wasn't thought of as much of a
+big deal in practice. It turns out that people in fact did not want to treat Go
+code this way, in practice this conflicted with the dependency model that Go
+encouraged people to use with how people consume libraries from GitHub or other
+such repository hosting sites.
+
+The main disconnect between importing from a GOPATH monorepo and a Go library
+off of GitHub is that when you import from a monorepo with a GOPATH in it, you
+need to be sure to import the repository path and not the path used inside the
+repository. This sounds weird but this means you'd import
+`github.com/Xe/x/src/github.com/Xe/x/markov` instead of
+`github.com/Xe/x/markov`. This means that things need to be extracted _out of_
+monorepos and reformatted into "flat" repos so that you can only grab the one
+package you need. This became tedious in practice.
+
+In Go 1.5 (the one where they rewrote the compiler in Go) they added support for
+[vendoring code into your
+repo](https://medium.com/@freeformz/go-1-5-s-vendor-experiment-fd3e830f52c3).
+The idea here was to make it easy to get closer to the model that the Go authors
+envisioned for how people should use Go. Go code should all be in one big happy
+repo and everything should have its place in your GOPATH. This combined with
+other tools people made allowed you to vendor all of your dependencies into a
+`vendor` folder and then you could do whatever you wanted from there.
+
+One of the big advantages of the `vendor` folder was that you could clone your
+git repo, create a new process namespace and then run tests without a network
+stack. Everything would work offline and you wouldn't have to worry about
+external state leaking in. Not to mention removing the angle of someone deleting
+their GitHub repos causing a huge problem for your builds.
+
+Save tests that require internet access or
+a database engine!
+
+This worked for a very long time. People were able to vendor their code into
+their repos and everything was better for people using Go. However the most
+critical oversight with the `vendor` folder approach was that the Go team didn't
+create an official tool to manage that `vendor` folder. They wanted to let tools
+like `godep` and `glide` handle that. This is kind of a reasonable take, Go
+comes from a very Google culture where this kind of problem doesn't happen, so
+as a result they probably won't be able to come up with something that meets the
+needs of the outside world very easily.
+
+I can't speak for how `godep` or `glide`
+works, I never really used them enough to have a solid opinion. I do remember
+using [`vendor`](https://github.com/bmizerany/vendor) in my own projects though.
+That had no real dependency resolution algorithm to speak of because it assumed
+that you had everything working locally when you vendored the code.
+
+### `dep`
+
+After a while the Go team worked with people in the community to come up with an
+"official experiment" in tracking dependencies called `dep`. `dep` was a tool
+that used some more fancy computer science maths to help developers declare
+dependencies for projects in a way like you do in other ecosystems. When `dep`
+was done thinking, it emitted a bunch of files in `vendor` and a lockfile in
+your repository. This worked really well and when I was working at Heroku this
+was basically our butter and bread for how to deal with Go code.
+
+It probably helped that my manager was on
+the team that wrote `dep`.
+
+One of the biggest advantages of `dep` over other tools was the way that it
+solved versioning. It worked by having each package declare
+[constraints](https://golang.github.io/dep/docs/the-solver.html) in the ranges
+of versions that everything requires. This allowed it to do some fancy
+dependency resolution math similar to how the solvers in `npm` or `cargo` work.
+
+This worked fantastically in the 99% case. There were some fairly easy to
+accidentally get yourself in cases where you could make the solver loop
+infinitely though, as well as ending up in a state where you have mutually
+incompatible transient dependencies without any real way around it.
+
+`npm` and `cargo` work around this by
+letting you use multiple versions of a single dependency in a
+project.
+
+However these cases were really really rare, only appearing in much, much larger
+repositories. I don't think I practically ran into this, but I'm sure someone
+reading this right now found themselves in `dep` hell and probably has a hell of
+a war story around it.
+
+### vgo and Modules
+
+This lead the Go team to come up with a middle path between the unrestricted
+madness of GOPATH and something more maximal like `dep`. They eventually called
+this Go modules and the core reasons for it are outlined in [this series of
+technical posts](https://research.swtch.com/vgo).
+
+These posts are a very good read and I'd
+highly suggest reading them if you've never seem then before. It outlines the
+problem space and the justification for the choices that Go modules ended up
+using. I don't agree with all of what is said there, but overall it's well
+worth reading at least once if you want to get an idea of the inspirations
+that lead to Go modules.
+
+Apparently the development of Go modules came out as a complete surprise,
+even to the core developer team of `dep`. I'm fairly sure this lead my
+manager to take up woodworking as his main non work side hobby, I can only
+wonder about the kind of resentment this created for other parts of the
+`dep` team. They were under the impression that `dep` was going to be the
+future of the ecosystem (likely under the subcommand `go dep`) and then had
+the rug pulled out from under their feet.
+
+The `dep` team was as close as we've
+gotten for having people in the _actual industry_ using Go _in production_
+outside of Google having a real voice in how Go is used in the real world. I
+fear that we will never have this kind of thing happen again.
It's
+also worth noting that the fallout of this lead to the core `dep` team leaving
+the Go community.
+
+Well, Google has to be using Go modules in
+their monorepo, right? If that's the official build system for Go it makes sense
+that they'd be dogfooding it hard enough that they'd need to use the tool in the
+same way that everyone else did.
+
+lol nope. They use an overcomplicated
+bazel/blaze abomination that has developed in parallel to their NIH'd source
+control server. Google doesn't have to deal with the downsides of Go modules
+unless it's in a project like Kubernetes. It's easy to imagine that they just
+don't have the same problems that everyone else does due to how weird Google
+prod is. Google only has problems that Google has, and statistically your
+company is NOT Google.
+
+Go modules does solve one very critical problem for the Go ecosystem though: it
+allows you to have the equivalent of the GOPATH but with multiple versions of
+dependencies in it. It allows you to have `within.website/ln@v0.7` and
+`within.website/ln@0.9` as dependencies for _two different projects_ without
+having to vendor source code or do advanced GOPATH manipulation between
+projects. It also adds cryptographic checksumming for each Go module that you
+download from the internet, so that you can be sure the code wasn't tampered
+with in-flight. They also created a cryptographic checksum comparison server so
+that you could ask a third party to validate what it thinks the checksum is so
+you can be sure that the code isn't tampered with on the maintainer's side. This
+also allows you to avoid having to shell out to `git` every time you fetch a
+module that someone else has fetched before. Companies could run their own Go
+module proxy and then use that to provide offline access to Go code fetched from
+the internet.
+
+Wait, couldn't this allow Google to see the
+source code of all of your Go dependencies? How would this intersect with
+private repositories that shouldn't ever be on anything but work
+machines?
+
+Yeah, this was one of the big privacy
+disadvantages out of the gate with Go modules. I think that in practice the
+disadvantages are limited, but still the fact that it defaults to phoning home
+to Google every time you run a Go build without all the dependencies present
+locally is kind of questionable. They did make up for this with the checksum
+verification database a little, but it's still kinda sus.
I'm not
+aware of any companies I've worked at running their own internal Go module
+caching servers, but I ran my own for a very long time.
+
+The earliest version of Go modules basically was a glorified `vendor` folder
+manager named `vgo`. This worked out amazingly well and probably made
+prototyping this a hell of a lot easier. This worked well enough that we used
+this in production for many services at Heroku. We had no real issues with it
+and most of the friction was with the fact that most of the existing ecosystem
+had already been using `dep` or `glide`.
+
+There was a bit of interoperability glue
+that allowed `vgo` to parse the dependency definitions in `dep`, `godep` and
+`glide`. This still exists today and helps `go mod init` tell what dependencies
+to import into the Go module to aid migration.
+
+If they had shipped this in prod, it probably would have been a huge success. It
+would also let people continue to use `dep`, `glide` and `godep`, but just doing
+that would also leave the ecosystem kinda fragmented. You’d need to have code
+for all 4 version management systems to parse their configuration files and
+implement algorithms that would be compatible with the semantics of all of them.
+It would work and the Go team is definitely smart enough to do it, but in
+practice it would be a huge mess.
+
+This also solved the case-insensitive filesystem problem with
+[bang-casing](https://go.dev/ref/mod#goproxy-protocol). This allows them to
+encode the capital letters in a path in a way that works on macOS and Windows
+without having to worry about horrifying hacks that are only really in place for
+Photoshop to keep working.
+
+### The Subtle Problem of `v2`
+
+However one of the bigger downsides that came with Go modules is what I've been
+calling the "v2 landmine" that Semantic Import Versioning gives you. One of the
+very earliest bits of Go advice was to make the import paths for version 1 of a
+project and version 2 of a project different so that people can mix the two to
+allow more graceful upgrading across a larger project. Semantic Import
+Versioning enforces this at the toolchain level, which means that it can be the
+gate between compiling your code or not.
+
+Many people have been telling me that
+I’m kind of off base for thinking that this is a landmine for people, but I am
+using the term “landmine” to talk about this because I feel like it reflects the
+rough edges of unexpectedly encountering this in the wild. It kinda feels like
+you stepped on a landmine.
+
+It's also worth noting that the protobuf
+team didn't use major version 2 when making an API breaking change. They
+defended this by saying that they are changing the import path away from GitHub,
+but it feels like they wanted to avoid the v2 problem.
+
+The core of this is that when you create major version 2 of a Go project, you
+need to adjust all your import paths everywhere in that project to import the
+`v2` of that package or you will silently import the `v1` version of that
+package. This can end up making large projects create circular dependencies on
+themselves, which is quite confusing in practice. When consumers are aware of
+this, then they can use that to more gradually upgrade larger codebases to the
+next major version of a Go module, which will allow for smaller refactors.
+
+This also applies to consumers. Given that this kind of thing is something that
+you only do in Go it can come out of left field. The go router
+[github.com/go-chi/chi](https://github.com/go-chi/chi/issues/462) tried doing
+modules in the past and found that it lead to confusing users. Conveniently they
+only really found this out after the Go modules design was considered final and
+Semantic Import Versioning has always been a part of Go modules and the Go team
+is now refusing to budge on this.
+
+My suggestion to people is to never
+release a version `1.x.x` of a Go project to avoid the "v2 landmine". The Go
+team claims that the right bit of tooling can help ease the pain, but this
+tooling never really made it out into the public. I bet it works great inside
+Google's internal monorepo though!
+
+When you were upgrading a Go project that already hit major version 2 or
+higher to Go modules, adopting Go modules forced maintainers to make another
+major version bump because it would break all of the import paths for every
+package in the module. This caused some maintainers to meet Go modules with
+resistance to avoid confusing their consumers. The workarounds for people that
+still used GOPATH using upstream code with Semantic Import Versioning in it
+were also kind of annoying at first until the Go team added "minimal module
+awareness" to GOPATH mode. Then it was fine.
+
+It feels like you are overly focusing on the
+`v2` problem. It can't really be that bad, can it? `grpc-gateway` updated to v2
+without any major issues. What's a real-world example of this?
+
+The situation with
+[github.com/gofrs/uuid](https://github.com/gofrs/uuid/issues/61) was heckin'
+bad. Arguably it's a teething issue as the ecosystem was still moving to the new
+modules situation, but it was especially bad for projects that were already at
+major version 2 or higher because adding Go modules support meant that they
+needed to update the major version just for Go modules. This was a tough sell
+and rightly so.
This was claimed to be made a non-issue by the right
+application of tooling on the side, but this tooling was either never developed
+or not released to us mere mortals outside of Google. Even with automated
+tooling this can still lead to massive diffs that are a huge pain to review,
+even if the only thing that is changed is the version number in every import of
+every package in that module. This was even worse for things that have C
+dependencies, as if you didn't update it everywhere in your dependency chain you
+could have two versions of the same C functions try to be linked in and this
+really just does not work.
+
+Overall though, Go modules has been a net positive for the community and for
+people wanting to create reliable software in Go. It’s just such a big semantics
+break in how the toolchain works that I almost think it would have been easier
+for the to accept if _that_ was Go 2. Especially since the semantic of how the
+toolchain worked changed so much.
+
+Wait, doesn’t the Go compiler have a
+backwards compatibility promise that any code built with Go 1.x works on go
+1.(x+1)?
+
+Yes, but that only applies to _code you
+write_, not _semantics of the toolchain_ itself. On one hand this makes a lot of
+sense and on the other it feels like a cop-out. The changes in how `go get` now
+refers to adding dependencies to a project and `go install` now installs a
+binary to the system have made an entire half decade of tool installation
+documentation obsolete. It’s understandable why they want to make that change,
+but the way that it broke people’s muscle memory is [quite frustrating for
+users](https://github.com/golang/go/issues/40276#issuecomment-1109797059) that
+aren’t keeping on top of every single change in semantics of toolchains (this
+bites me constantly when I need to quick and dirty grab something outside of a
+Nix package). I understand _why_ this isn’t a breaking change as far as the
+compatibility promise but this feels like a cop-out in my subjective
+opinion.
+
+## Contexts
+
+One of Go’s major features is its co-operative threading system that it calls
+goroutines. Goroutines are kinda like coroutines that are scheduled by the
+scheduler. However there is no easy way to "kill" a goroutine. You have to add
+something to the invocation of the goroutine that lets you signal it to stop and
+then opt-in the goroutine to stop.
+
+Without contexts you would need to do all of this legwork manually. Every
+project from the time before contexts still shows signs of this. The best
+practice was to make a "stop" channel like this:
+
+```go
+stop := make(chan struct{})
+```
+
+And then you'd send a cancellation signal like this:
+
+```go
+stop <- struct{}{}
+```
+
+The type `struct{}` is an anonymous
+structure value that takes 0 bytes in ram. It was suggested to use this as your
+stopping signal to avoid unneeded memory allocations. A `bool` needs one whole
+machine word, which can be up to 64 bits of ram. In practice the compiler can
+smoosh multiple bools in a struct together into one place in ram, but when
+sending these values over a channel like this you can't really cheat that
+way.
+
+This did work and was the heart of many event loops, but the main problem with
+it is that the signal was only sent _once_. Many other people also followed up
+the stop signal by closing the channel:
+
+```go
+close(stop)
+```
+
+However with naïve stopping logic the closed channel would successfully fire a
+zero value of the event. So code like this would still work the way you wanted:
+
+```go
+select {
+ case <- stop:
+ haltAndCatchFire()
+}
+```
+
+### Package `context`
+
+However if your stop channel was a `chan bool` and you relied on the `bool`
+value being `true`, this would fail because the value would be `false`. This
+was a bit too brittle for comfortable widespread production use and we ended
+up with the [context](https://pkg.go.dev/context) package in the standard
+library. A Go context lets you more easily and uniformly handle timeouts and
+giving up when there is no more work to be done.
+
+This started as something that existed
+inside the Google monorepo that escaped out into the world. They also claim to
+have an internal tool that makes
+[`context.TODO()`](https://pkg.go.dev/context#TODO) useful (probably by showing
+you the callsities above that function?), but they never released that tool as
+open source so it’s difficult to know where to use it without that added
+context.
+
+One of the most basic examples of using contexts comes when you are trying to
+stop something from continuing. If you have something that constantly writes
+data to clients such as a pub-sub queue, you probably want to stop writing data
+to them when the client disconnects. If you have a large number of HTTP requests
+to do and only so many workers can make outstanding requests at once, you
+want to be able to set a timeout so that after a certain amount of time it gives
+up.
+
+Here's an example of using a context in an event processing loop (of course while
+pretending that fetching the current time is anything else that isn't a contrived
+example to show this concept off):
+
+```go
+t := time.NewTicker(30 * time.Second)
+ctx, cancel := context.WithCancel(context.Background())
+defer cancel()
+
+for {
+ select {
+ case <- ctx.Done():
+ log.Printf("not doing anything more: %v", ctx.Err())
+ return
+ case data := <- t.C:
+ log.Printf("got data: %s", data)
+ }
+}
+```
+
+This will have the Go runtime select between two channels, one of them will
+emit the current time every 30 seconds and the other will fire when the
+`cancel` function is called.
+
+Don't worry, you can call the `cancel()`
+function multiple times without any issues. Any additional calls will not do
+anything special.
+
+If you want to set a timeout on this (so that the function only tries to run
+for 5 minutes), you'd want to change the second line of that example to this:
+
+```go
+ctx, cancel := context.WithTimeout(context.Background(), 5 * time.Minute)
+defer cancel()
+```
+
+You should always `defer cancel()` unless
+you can prove that it is called elsewhere. If you don't do this you can leak
+goroutines that will dutifully try to do their job potentially forever without
+any ability to stop them.
+
+The context will be automatically cancelled after 5 minutes. You can cancel it
+sooner by calling the `cancel()` function should you need to. Anything else in
+the stack that is context-aware will automatically cancel as well as the
+cancellation signal percolates down the stack and across goroutines.
+
+You can attach this to an HTTP request by using
+[`http.NewRequestWithContext`](https://pkg.go.dev/net/http#NewRequestWithContext):
+
+```go
+req, err := http.NewRequestWithContext(ctx, http.MethodGet, "https://christine.website/.within/health", nil)
+```
+
+And then when you execute the request (such as with `http.DefaultClient.Do(req)`)
+the context will automatically be cancelled if it takes too long to fetch the
+response.
+
+You can also wire this up to the `Control-c` signal using a bit of code
+[like this](https://medium.com/@matryer/make-ctrl-c-cancel-the-context-context-bd006a8ad6ff).
+Context cancellation propagates upwards, so you can use this to ensure that things
+get stopped properly.
+
+Be sure to avoid creating a "god context"
+across your entire app. This is a known anti-pattern and this pattern should only
+be used for small command line tools that have an expected run time in the minutes
+at worst, not hours like production bearing services.
+
+This is a huge benefit to the language because of how disjointed the process of
+doing this before contexts was. Because this wasn’t in the core of the language,
+every single implementation was different and required learning what the library
+did. Not to mention adapting between libraries could be brittle at best and
+confusing at worst.
+
+I understand why they put data into the context type, but in practice I really
+wish they didn’t do that. This feature has been abused a lot in my experience.
+At Heroku a few of our production load bearing services used contexts as a
+dependency injection framework. This did work, but it turned a lot of things
+that would normally be compile time errors into runtime errors.
+
+I say this as someone who maintains a
+library that uses contexts to store [contextually relevant log
+fields](https://pkg.go.dev/within.website/ln) as a way to make logs easier to
+correlate between.
Arguably you could make the case that people are misusing the
+tool and of course this is what will happen when you do that but I don't know if
+this is really the right thing to tell people.
+
+I wish contexts were in the core of the language from the beginning. I know that
+it is difficult to do this in practice (especially on all the targets that Go
+supports), but having cancellable syscalls would be so cool. It would also be
+really neat if contexts could be goroutine-level globals so you didn’t have to
+"pollute" the callsites of every function with them.
+
+At the time contexts were introduced,
+one of the major arguments I remember hearing against them was that contexts
+"polluted" their function definitions and callsites. I can't disagree with this
+sentiment, at some level it really does look like contexts propagate "virally"
+throughout a codebase.
I think that the net improvements to
+reliability and understandability of how things get stopped do make up for this
+though. Instead of a bunch of separate ways to cancel work in each individual
+library you have the best practice in the standard library. Having contexts
+around makes it a lot harder to "leak" goroutines on accident.
+
+## Generics
+
+One of the biggest ticket items that Go has added is "generic types", or being
+able to accept types as parameters for other types. This is really a huge ticket
+item and I feel that in order to understand _why_ this is a huge change I need
+to cover the context behind what you had before generics were added to the
+language.
+
+One of the major standout features of Go is interface types. They are like Rust
+Traits, Java Interfaces, or Haskell Typeclasses; but the main difference is that
+interface types are _implicit_ rather than explicit. When you want to meet the
+signature of an interface, all you need to do is implement the contract that the
+interface spells out. So if you have an interface like this:
+
+```go
+type Quacker interface {
+ Quack()
+}
+```
+You can make a type like `Duck` a `Quacker` by defining the `Duck` type and a
+`Quack` method like this:
+
+```go
+type Duck struct{}
+
+func (Duck) Quack() { fmt.Println("Quack!") }
+```
+
+But this is not limited to just `Ducks`, you could easily make a `Sheep` a
+`Quacker` fairly easily:
+
+```go
+type Sheep struct{}
+
+func (Sheep) Quack() { fmt.Println("*confused sheep noises*") }
+```
+
+This allows you to deal with expected _behaviors_ of types rather than having to
+have versions of functions for every concrete implementation of them. If you
+want to read from a file, network socket, `tar` archive, `zip` archive, the
+decrypted form of an encrypted stream, a TLS socket, or a HTTP/2 stream they're
+all [`io.Reader`](https://pkg.go.dev/io#Reader) instances. With the example
+above we can make a function that takes a `Quacker` and then does something with
+it:
+
+```go
+func main() {
+ duck := Duck{}
+ sheep := Sheep{}
+
+ doSomething(duck)
+ doSomething(sheep)
+}
+
+func doSomething(q Quacker) {
+ q.Quack()
+}
+```
+
+If you want to play with this example,
+check it out on the Go playground [here](https://go.dev/play/p/INK8O2O-D01). Try
+to make a slice of Quackers and pass it to `doSomething`!
+
+You can also embed interfaces into other interfaces, which will let you create
+composite interfaces that assert multiple behaviours at once. For example,
+consider [`io.ReadWriteCloser`](https://pkg.go.dev/io#ReadWriteCloser). Any
+value that matches an `io.Reader`, `io.Writer` and an `io.Closer` will be able
+to be treated as an `io.ReadWriteCloser`. This allows you to assert a lot of
+behaviour about types even though the actual underlying types are opaque to you.
+
+This means it’s easy to split up a [`net.Conn`](https://pkg.go.dev/net#Conn)
+into its reader half and its writer half without really thinking about
+it:
+
+```go
+conn, _ := net.Dial("tcp", "127.0.0.1:42069")
+
+var reader io.Reader = conn
+var writer io.Writer = conn
+```
+
+And then you can pass the writer side off to one function and the reader side
+off to another.
+
+There’s also a bunch of room for "type-level middleware" like
+[`io.LimitReader`](https://pkg.go.dev/io#LimitReader). This allows you to set
+constraints or details around an interface type while still meeting the contract
+for that interface, such as an `io.Reader` that doesn’t let you read too much,
+an `io.Writer` that automatically encrypts everything you feed It with TLS, or
+even something like sending data over a Unix socket instead of a TCP one. If it
+fits the shape of the interface, it Just Works.
+
+However, this falls apart when you want to deal with a collection of _only one_
+type that meets an interface at once. When you create a slice of `Quacker`s and
+pass it to a function, you can put both `Duck`s and `Sheep` into that slice:
+
+```go
+quackers := []Quacker{
+ Duck{},
+ Sheep{},
+}
+
+doSomething(quackers)
+```
+
+If you want to assert that every `Quacker` is the same type, you have to do some
+fairly brittle things that step around Go's type safety like this:
+
+```go
+func doSomething(qs []Quacker) error {
+ // Store the name of the type of first Quacker.
+ // We have to use the name `typ` because `type` is
+ // a reserved keyword.
+ typ := fmt.Sprintf("%T", qs[0])
+
+ for i, q := range qs {
+ if qType := fmt.Sprintf("%T", q); qType != typ {
+ return fmt.Errorf("slice value %d was type %s, wanted: %s", qType, typ)
+ }
+
+ q.Quack()
+ }
+
+ return nil
+}
+```
+
+This would explode at runtime. This same kind of weakness is basically the main
+reason why the Go standard library package [`container`](https://pkg.go.dev/container)
+is mostly unused. Everything in the `container` package deals with
+`interface{}`/`any` values, which is Go for "literally anything". This means
+that without careful wrapper code you need to either make interfaces around
+everything in your lists (and then pay the cost of boxing everything in an
+interface, which adds up a lot in practice in more ways than you'd think) or
+have to type-assert anything going into or coming out of the list, combined
+with having to pay super close attention to anything touching that code
+during reviews.
+
+Don't get me wrong, interface types
+are an _amazing_ standout feature of Go. They are one of the main reasons that
+Go code is so easy to reason about and work with. You don't have to worry
+about the entire tree of stuff that a value is made out of, you can just
+assert that values have behaviors and then you're off to the races. I end up
+missing the brutal simplicity of Go interfaces in other languages like Rust.
+
+
+### Introducing Go Generics
+
+In Go 1.18, support for adding types as parameters to other types was added.
+This allows you to define constraints on what types are accepted by a function,
+so that you can reuse the same logic for multiple different kinds of underlying
+types.
+
+That `doSomething` function from above could be rewritten like this with
+generics:
+
+```go
+func doSomething[T Quacker](qs []T) {
+ for i, q := range qs {
+ q.Quack()
+ }
+}
+```
+
+However this doesn't currently let you avoid mixing types of `Quacker`s at
+compile time like I assumed while I was writing the first version of this
+article. This does however let you write code like this:
+
+```go
+doSomething([]Duck{{}, {}, {}})
+doSomething([]Sheep{{}, {}, {}})
+```
+
+And then this will reject anything that _is not a `Quacker`_ at compile time:
+
+```go
+doSomething([]string{"hi there this won't work"})
+```
+
+```
+./prog.go:20:13: string does not implement Quacker (missing Quack method)
+```
+
+### Unions
+
+This also lets you create untagged union types, or types that can be a range of
+other types. These are typically useful when writing parsers or other similar
+things.
+
+It's frankly kind of fascinating that
+something made by Google would even let you _think_ about the word "union" when
+using it.
+
+Here's an example of a union type of several different kinds of values that you
+could realistically see in a parser for a language like [LOLCODE](http://www.lolcode.org/):
+
+```go
+// Value can hold any LOLCODE value as defined by the LOLCODE 1.2 spec[1].
+//
+// [1]: https://github.com/justinmeza/lolcode-spec/blob/master/v1.2/lolcode-spec-v1.2.md#types
+type Value interface {
+ int64 // NUMBR
+ float64 // NUMBAR
+ string // YARN
+ bool // TROOF
+ struct{} // NOOB
+}
+```
+
+This is similar to making something like an
+[`enum`](https://doc.rust-lang.org/book/ch06-01-defining-an-enum.html) in Rust,
+except that there isn't any tag for what the data could be. You still have to do
+a type-assertion over every value it _could_ be, but you can do it with only the
+subset of values listed in the interface vs any possible type ever made. This
+makes it easier to constrain what values can be so you can focus more on your
+parsing code and less on defensively programming around variable types.
+
+This adds up to a huge improvement to the language, making things that were
+previously very tedious and difficult very easy. You can make your own
+generic collections (such as a B-Tree) and take advantages of packages like
+[`golang.org/x/exp/slices`](https://pkg.go.dev/golang.org/x/exp/slices) to avoid
+the repetition of having to define utility functions for every single type you
+use in a program.
+
+I'm barely scratching the surface with
+generics here, please see the [type parameters proposal
+document](https://go.googlesource.com/proposal/+/refs/heads/master/design/43651-type-parameters.md)
+for a lot more information on how generics work. This is a well-written thing
+and I highly suggest reading this at least once before you try to use generics
+in your Go code. I've been watching this all develop from afar and I'm very
+happy with what we have so far (the only things I'd want would be a bit more
+ability to be precise about what you are allowing with slices and maps as
+function arguments).
+
+---
+
+In conclusion, I believe that we already have Go 2. It’s just called Go 1.18 for
+some reason. It’s got so many improvements and fundamental changes that I
+believe that this is already Go 2 in spirit. There are so many other things that
+I'm not covering here (mostly because this post is so long already) like
+fuzzing, RISC-V support, binary/octal/hexadecimal/imaginary number literals,
+WebAssembly support, so many garbage collector improvements and more. This has
+added up to make Go a fantastic choice for developing server-side applications.
+
+I, as some random person on the
+internet that is not associated with the Go team, think that if there was
+sufficient political will that they could probably label what we have as Go 2,
+but I don’t think that is going to happen any time soon. Until then, we still
+have a very great set of building blocks that allow you to make easy to maintain
+production quality services, and I don’t see that changing any time soon.