<A>

The opening keynote from the inaugural HTML Special held before CSS Day 2016 in Amsterdam.

It’s all starting to come together.

The world exploded into a whirling network of kinships, where everything pointed to everything else, everything explained everything else.
— Umberto Eco, Foucault’s Pendulum

A is for Apophenia

Apophenia is the name for that tendency in humans to see patterns where none exist, to draw connections, to make links.

Every conspiracy theory is an example of apophenia. But you don’t have to be a conspiracy theorist to experience it. We do it all the time. We see shapes in the constellations in the night sky. We see faces in just about everything.

Stars

Today, I would like to engage the apopheniac in you.

A is for Anchor

An anchor is an odd device to represent a link. I can even remember seeing anchor symbols used in the interfaces of rich text editors. If it wasn’t an anchor, it was a chain. I suppose that was meant to represent a link …because chains have links.

Anyway, why A? Why Anchor?

It goes back to how the A element was originally used. When I say originally, I mean originally. Let’s look at the first A element in the first web page at the first URL.

<A NAME=0 HREF="WhatIs.html">
hypermedia
</A>

This looks pretty familiar to us today. There’s an A Element with opening and closing tags, some text in between, and an HREF attribute for the destination. But there’s also a NAME attribute. This has since been deprecated—now we can just use an ID attribute on any element. The idea was that A elements could be used to create destinations for inbound links. They were, if you like, anchors within a page that other pages could tether themselves to. Each anchor is given a unique identifier (unique within the page, that is). Here, the identifier is simply the number zero, because this page was created by a programmer and in the mind of a programmer, counting begins with zero.

This use of the A element—using NAME attributes to create in-page anchors—never really took off. But the other attribute, the HREF attribute, that spread like wildfire. It’s short for hypertext reference, and in this particular instance, the reference is to another page in the same directory on the same server. It’s a page about hypertext.

Hypertext is text which is not constrained to be linear.

The term was first coined by Ted Nelson. He didn’t just talk about text either. He also coined the term hypermedia. He coined lots of interesting words. He talked about things being deeply intertwingled. He also coined the term teledildonics, but that is not directly relevant to hypertext or hypermedia.

If hypertext is text which is not constrained to be linear, how did we ever manage with good old-fashioned non-hyper text and non-hyper media? We used archives.

A is for Archive

In 1965, JCR Licklider was commissioned to write a report on Libraries Of The Future. He—and his company Bolt, Beranek and Newman—would prove to be instrumental in creating our modern interconnected hypermedia world. More on that later. For now, let’s look at this report. It has two parts:

  1. Concepts and Problems of Man’s Interaction with the Body of Recorded Knowledge
  2. Explorations in the Use of Computers in Information Storage, Organization, and Retrieval.

I love the scope of that first part, looking at the body of recorded knowledge.

It’s interesting that he talks about knowledge, not information, not data, but knowledge. How does data become information? How does information become knowledge?

The Library Of Babel

The Library Of Babel is a short story by Jorge Luis Borges, who I think of as the poet laureate of hypertext . He imagines a vast library that is filled with data, but frustratingly short on knowledge …because this library contains not only all the books ever written, but all the books that could ever possibly be written, with every possible permutation.

Here’s how it works:

The universe (which others call the Library) is composed of an indefinite, perhaps infinite number of hexagonal galleries… The arrangement of the galleries is always the same: Twenty bookshelves, five to each side, line four of the hexagon’s six sides… each bookshelf holds thirty-two books identical in format; each book contains four hundred ten pages; each page, forty lines; each line, approximately eighty black letters.

Let’s figure out how many books are in the Library of Babel. First, we need to know how much data each book holds.

  • There are eighty symbols (or letters) per line, and
  • 40 lines per page.
  • There are 410 pages in each book.

Multiplying 80 by 40 gives us 3200, the number of symbols on each page. Multiply that by 410 and we get a total number of 1,312,000 symbols in each book.

We have two other pieces of information to work with. Borges tells us:

The orthographic symbols are twenty-five in number.

That’s 22 letters, the comma, the period, and the space.

Here’s the crucial bit of information that ensures that the library has boundaries:

In the vast library there are no two identical books.

Knowing that, we can calculate the number of books in the library. It’s the number of symbols (25) raised to the power of the number of symbols in each book (1,312,000).

25 to the power of 1,312,000 expressed in base ten is 10 to the power of 1,834,097. Remember that’s just the number of books: a figure that’s over 1,834,097 digits long. That number wouldn’t fit inside one book in the library (which, if you remember only holds 1,312,000 symbols).

So the number of books in the Library of Babel is not infinite …but it is really, really, really big. To give you some idea of just how big 10 to the power of 1,834,097 is, it has been calculated that the observable universe contains approximately 10 to the power of 80 atoms. There are more books in the Library of Babel than there are atoms in the universe.

And yet, thanks to the World Wide Web, you can theoretically peruse every single one of them.

Jonathan Basile has created libraryofbabel.info—an online representation of Borges’s creation. It contains all possible text. You can browse by hexagon, then shelf, then row, then book, then page. Or you can search for a specific piece of text, because—of course—that text must be in there somewhere.

The very words I am speaking now must be somewhere in the library.

There’s also a plug-in for Chrome so you can highlight any piece of text on the web, and then find its corresponding page in the Library of Babel.

This library has every possible piece of data …but it’s sorely lacking in information (although the online version helps).

A is for All

Having all the data isn’t enough. It needs to be organised—turned into information—for us to make use of it (and hopefully further transform that information into knowledge).

There have been many attempts to organise information. When those attempts are limited to a subset of data—instead of trying to create a Library of Babel—then they can be quite successful.

Carl Linnaeus organised the natural world using a naming convention for describing species—binomial nomenclature.

Charles Messier catalogued astronomical objects.

Melvil Dewey created the Dewey Decimal System to help librarians organise their collections. But this was a proprietary system, not an open standard. So the Belgian librarian Paul Otlet devised a Universal Decimal Classification system. I think it’s fair to think of him as the father of information architecture.

The crazy old uncle of information architecture would be bishop John Wilkins, who in 1668 wrote An Essay towards a Real Character and a Philosophical Language. Centuries later, this would pique the interest of Borges who described Wilkins’s madcap idea in an essay called The Analytical Language of John Wilkins. The idea was that the world could be classified into sounds.

He divided the universe in forty categories or classes, these being further subdivided into differences, which was then subdivided into species. He assigned to each class a monosyllable of two letters; to each difference, a consonant; to each species, a vowel. For example: de, which means an element; deb, the first of the elements, fire; deba, a part of the element fire, a flame.

It didn’t scale well.

Gottfried Wilhelm Leibniz—inventor of calculus and nemesis to Newton—had a similar idea to Wilkins, but whereas Wilkins was trying to classify information using sounds, Leibniz wanted to use symbols: characteristica universalis. His leap of genius was to realise that if you could do this—represent the world with symbols—then you could perform calculations on those symbols. He described the conceptual framework for performing such operations as a Calculus ratiocinator. Centuries later, Norbert Wiener, the creator of cybernetics would say:

The general idea of a computing machine is nothing but a mechanization of Leibniz’s calculus ratiocinator.

A is for Ada

Leibniz’s calculus ratiocinator was an idea, rather than a real machine. Charles Babbage was a Victorian inventor and entrepreneur who was given seed funding by the British government to create his Difference Engine: a machine for computing logarithmic tables …a computer, if you will.

The Difference Engine

The Wozniak to Babbage’s Jobs was Ada Lovelace, the daughter of the notorious Lord Byron. Ada’s mother did everything in her power to steer her daughter away from following in her father’s footsteps of becoming a poet. Instead Ada immersed herself in the world of mathematics. It was through mathematics that Ada hit upon the same insight as Leibniz—if we can perform calculations upon symbols, and those symbols don’t have to just represent numbers, then we can perform calculations on anything …maybe even poetry.

In enabling mechanism to combine together general symbols in successions of unlimited variety and extent, a uniting link is established between the operations of matter and the abstract mental processes of the most abstract branch of mathematical science.

The Difference Engine was abandoned in favour of an even more ambitious project: The Analytical Engine. Finally, the world would get a mechanical machine capable of performing calculations on symbols that could represent concepts or thoughts—a thinking machine!

As with so many start-ups, Lovelace and Babbage never quite managed to make it to market. But at least they did genuinely have an incredible journey before going broke.

A is for Alan

The Analytical Engine, like the calculus ratiocinator, remained an idea. It would be another century before we’d get a real nuts’n’bolts computing machine, thanks to the brilliant Alan Turing.

But as well as having a hand in creating the world’s first physical computer, he also presented us with an imaginary machine. This theoretical machine was described as having:

…an infinite tape marked out into squares, on each of which a symbol could be printed.

Like Leibniz’s calculus ratiocinator, a Turing machine would operate on symbols. How many symbols are we talking about here? Remember the Library Of Babel had twenty five symbols to work with.

Two. Two symbols. One and zero. On and off. True and false. With an infinitely long piece of tape and infinite amount of time, two symbols are enough to calculate literally anything.

This reduction of the world into its smallest pieces was the brainchild of Claude Shannon. He coined the term "bit" to describe this indivisible unit of information.

By the way, Turing himself referred to his theoretical machine as an automatic machine, or a-machine. A is for A-machine.

A is for Atlantic

There’s another imaginary machine that serves as a wonderful conceptual prototype for working with hypertext and hypermedia. In 1945 Vannevar Bush published an article in the Atlantic Monthly called As We May Think. In this article, Bush describes a machine called the Memex, a contraction of Memory Index.

Memex

The Memex is built into a desk. Screens and switches on the surface of the desk allow the user to interact with huge amounts of information stored within the desk on microfilm. This would provide an "enlarged intimate supplement to one’s memory."

Because everyone’s mind is different, no two people would use the Memex in quite the same way. Bush described these individualistic approaches to linking concepts together as associative trails:

Wholly new forms of encyclopedias will appear, ready made with a mesh of associative trails running through them, ready to be dropped into the memex and there amplified.

The information stored in the Memex is the same for everyone, but the associative trails created by the user in navigating this information are unique. Bush also proposed that these associative trails could be shared. Users of the Memex could follow the breadcrumbs left by others.

This is hypertext.

A is for Augmentation

So many ideas about hypertext confined to the imagination! When do we get a working demo?

Ladies and gentlemen, on December 9th, 1968, we get the mother of all demos, courtesy of Douglas Engelbart. Six years previously, he had set out his goals for human-computer interaction in a work entitled Augmenting Human Intellect: A Conceptual Framework. To achieve this aim of augmenting humanity, he created a working demo of his oNLine System, abbreviated to NLS.

Oh, and for this demo he just happened to invent video conferencing, the graphical user interface, and the mouse. He also implemented hypertext.

With Bush’s Memex and Engelbart’s NLS, concepts could be linked together, turning them from text storage devices to hypertext storage devices. But for hypertext to reveal its true power, we need a network greater than anything seen up to this point. Towards the end of the mother of all demos, we hear the first rumblings of just such a network.

A is for Arpa

ARPA stands for the Advanced Research Projects Agency.

This government agency turned to our old friend JCR Licklider, author of Libraries Of The Future. His company—Bolt, Beranek and Newman—set about designing a communications system that used this new-fangled packet switching that Leonard Kleinrock was so excited about. With packet switching, information was broken down into discrete chunks, routed around a network independent of each other, and then re-assembled at the destination.

At the same time, this idea of packet switching was independently discovered by Paul Baran at the Rand Corporation, who was trying to find a resilient network architecture capable of surviving nuclear attack.

The packet-switching idea was put to the test with the creation of a new network called the ARPANET. The very first message sent over the ARPANET was at 10:30pm on October, 29th, 1969. It was the command: LOG IN.

The message was sent and… the system crashed after the first two characters. Fittingly, then, the real first message sent over this proto-internet was "LO".

But they fixed the bugs and they kept working on making the system better as it grew bigger. It morphed from being a single network, the ARPANET, into being a network of networks, or an Inter-network, soon shortened to simply Internet.

For this internet to work, it was essential that all the individual networks connecting to it were using the same protocols to communicate. That’s what really makes the internet the internet—regardless of what kind of hardware is being used, there’s an agreement on how to switch those packets around. That’s really all a protocol is: an agreement. They’re more like treaties than code.

Bob Kahn, Vint Cerf, Jon Postel

Bob Kahn and Vint Cerf were the statesmen crafting the internet’s protocols, and Jon Postel was the diplomat ensuring adoption went smoothly. They wanted a robust network, resilient not to nuclear attack, but to any kind of top-down control. They set out to create a protocol that would work for a network with no centre.

Together they crafted TCP/IP: the Transmission Control Protocol and Internet Protocol. TCP/IP is a deliberately dumb set of protocols. The protocols care not a whit for the contents of the packets being switched around the network. It’s a simple low-level agreement. They used to joke that you should be able to implement TCP/IP using two tin cans and a piece of string.

You can then create more complex protocols on top of this simple, low-level, dumb foundation. You can create protocols for sending and receiving email, protocols for telnet, gopher, file transfer protocols that sit atop TCP/IP. Best of all, you don’t need to ask anyone for permission. If you want to create a new protocol today, you can just go ahead and do it. All you need is rough consensus and running code. It turns out that running code isn’t the hard part. The hard part is convincing people to use your protocol. Otherwise you don’t benefit from Metcalfe’s Law:

The value of a network is proportional to the square of the number of connected users of the system.

Think about it. The first person to own a telephone had a completely useless object. As soon as one other person had a telephone, it suddenly become exponentially more useful. That was the challenge facing the creator of a new protocol at the start of the 90s: that protocol was HTTP—the HyperText Transfer Protocol.

HTTP is one part of a three-part stack: HTTP is the protocol, URLs are the identifiers, and HTML is the format. Together they form the World Wide Web project.

The project was the brainchild of a young computer scientist named Tim Berners-Lee. This wasn’t his first attempt at creating a hypertext system.

In the 1980s, he created a system called ENQUIRE. It was named after a Victorian book of manners called Enquire Within Upon Everything, which I always thought would be a great name for the web.

Enquire didn’t work out in the end, but it would influence the design of the World Wide Web project.

Another influence on the design of the system was the place where Tim Berners-Lee was working. CERN—the European Centre for Nuclear Research. It’s an amazing place. The greatest experiment in the history of our species is being conducted beneath the border between Switzerland and France. In the 16-mile wide ring of the Large Hadron Collider at CERN, human beings are recreating the conditions from the start of our universe. Protons are smashed together at velocities approaching the speed of light. It’s a truly awe-inspiring endeavour.

CERN

When I visited CERN, I expected to be blown away by the science, and I was. I also expected to be blown away by simply being at the birthplace of the web, and I was. But what I wasn’t expecting was to be blown away by how things get done at CERN. There is very little hierarchy. People from all kinds of backgrounds—from Nobel prize winning physicists to students on a Summer internship—collaborate on experiments for pure scientific research.

Trying to manage the flow of information in this collaborative but chaotic place was the challenge that Tim Berners-Lee was trying to solve. You can’t just mandate a particular operating system or piece of software—people at CERN can and will use whatever they want. In the same way that the internet is a network of networks, what CERN needed was some way of allowing all these different computers with different operating systems to share information with each other.

Tim Berners-Lee submitted a paper to his supervisor, Mike Sendall. It had the uninspiring title Information Management: A Proposal. Well, Mike Sendall must have seen some potential, because he scrawled across the top:

Vague, but exciting.

The proposal described:

…a solution based on a distributed hypertext system.

Tim Berners-Lee was very familiar with previous hypertext systems. All of these fed into his project:

  • Vannevar Bush’s Memex,
  • Douglas Engelbart’s oNline System,
  • Ted Nelson’s Xanadu, although it was still vapourware at this point,
  • Apple’s Hypercard system for the Mac,
  • and his own Enquire project.

But just creating the code wasn’t enough. He—and his colleague and collaborator, Robert Cailliau—needed to convince the scientists at CERN to use this technology. To start with, they needed a catchy name.

For a while, they floated the idea of calling it the Mesh.

Then they kick around the idea of calling it The Information Mine. But Tim Berners-Lee wasn’t keen on this one. He knew that, whatever name they chose, it would end up getting abbreviated, and he was worried it would look a bit egotistical.

And so they settled on World Wide Web. You have to admire the chutzpah of calling it World Wide Web when, at that point, it only existed on one person’s computer.

Sure enough, it did end up getting abbreviated. Except in this case, there are actually more syllables in the so-called abbreviation—WWW—than there are in the full name.

WWW

They even made a logo. Graphic designers they are not. But there is a reason for the green colour of the Ws. Robert Cailliau is a synesthesiac—he “hears” the W sound as the colour green.

And so the web was born. Good job!

Where the web was born

Learning from the lessons of TCP/IP, Tim Berners-Lee made sure to keep the individual parts of the system as simple as possible (but no simpler). The World Wide Web didn’t succeed because the technology was the best; far from it. It succeeded because the technology was just simple enough—but also powerful enough—for people to get started with straight away.

Take HTML, for example. There was no official Version 1 specification for this hypertext markup language.

Instead there was a document called simply HTML Tags, presumably written by Tim Berners-Lee. This document listed the entirety of HTML, which was a grand total of 21 elements.

Most of those elements weren’t even invented by Tim Berners-Lee. Instead, he borrowed the vocabulary already being used by scientists at CERN. They were used to writing documents in GML, which is supposed to stand for Generalised Markup Language, but was coincidentally created by three people whose last names were Goldfarb, Mosher, and Lorie: G, M, and L.

There was one element that was completely new to HTML:

A

This one single element is what enables the HT part of HTML. With this element, and its href attribute, anyone could link to anything on the web. It is brilliant in its simplicity.

Pleased with the way their project was progressing, Tim Berners-Lee and Robert Cailliau submitted a proposal to present their World Wide Web at a hypertext conference.

They were rejected. Hypertext experts thought the World Wide Web was stupid.

As Ted Nelson put it:

Today’s one-way hypertext—the World Wide Web—is far too shallow. The Xanadu project foresaw world-wide hypertext decades ago, and endeavored to create a much deeper system. The Web, however, took over with a very shallow structure.

He’s not wrong. Hypertext on the web is shallow. It is stupid. Like TCP/IP, it is not smart.

The thing is, if you’re hoping to get mass adoption, being smart is a bug. Being stupid is a feature.

Just about every other hypertext system embodied the idea of two-way linking. There was an awareness at both ends of the link. If the resource being linked to were to move or change, the link could be updated. It’s robust, but it’s complicated.

On the World Wide Web, by contrast, links only work in one direction. If the resource being linked to ends up moving or changing, well, tough luck. The result is link rot. That’s the price we pay for a very simple hypertext system.

But, now that the web has been around a couple of decades, there is a sort-of, kind-of implementation of two-way linking.

It uses the humble rel attribute.

Rel is short for relationship. The value inside the rel attribute describes the relationship of the linked resource to the current document.

<a href="…" rel="…">

Some rel values were officially canonised in HTML.

rel="prev" means that the linked resource has the relationship of being the previous document to the current document.

<a href="…" rel="prev">

rel="next" means that the linked resource has the relationship of being the next document after the current document.

<a href="…" rel="next">

rel="author" means that the linked resource has the relationship of representing the author of the current document.

<a href="…" rel="author">

You get the idea.

Incidentally, there was once a corresponding rev attribute that described the reverse relationship. In other words, the relationship of the current document to the linked resource.

<a href="…" rev="…">

So you could simultaneously say the current document has a relationship of being the previous document to the linked resource and that the linked resource has a relationship of being the next document to the current document.

<a href="…" rev="prev" rel="next">

Confused? Well, that’s why the rev attribute was eventually dropped from HTML. It was just too complicated.

There was an attempt to use the rev attribute in an early microformat called vote-links. This was proposed way back in 2004. Blogs were hot. Political blogs were very hot indeed: Bush Jr.—no relation to Vannevar—was running for re-election in the States, and John Kerry was the challenger. There was much blogging and gnashing of teeth.

The problem was the rise of PageRank, the algorithm that drove Google’s search engine. The fundamental premise of PageRank was that linking to something counts as an endorsement. But there were many bloggers linking to articles that they disagreed with very strongly.

By using a rev value of “vote-for”, authors could explicitly say that this document is a vote for the resource being linked to.

Or by using a rev value of “vote-against” they could make it clear that this document is a vote against the linked resource.

It never really took off because, as I said, the rev attribute was just too hard to grok.

And that’s okay. The whole point of microformats is that they are the very embodiment of the motto of the Internet Engineering Task Force: rough consensus and running code. There just wasn’t enough take-up of vote-links for it to thrive.

Another early idea—that actually preceded the official creation of microformats—was XFN, which stood for XHTML Friends Network: the most Web 1.0 format name ever.

XFN built on existing behaviour. Bloggers would often have a list of links in their sidebar pointing to other bloggers they had some kind of relationship with. If you assume that a URL can represent a person, then the rel attribute is perfect for encoding that relationship information.

I can link to a friend’s website and say that the person represented by the linked resource has a relationship of being a friend to the person represented by the current document: me!

<a href="…" rel="friend">

Or I can link to a colleague’s website and say that they have a relationship of being a colleague to me.

<a href="…" rel="colleague">

And because rel values—like class values—can be space-separated, I can combine rel values into one attribute. I can link to someone and say that they are both a friend and a colleague.

<a href="…" rel="friend colleague">

I still have XFN values in the sidebar of my blog, but again, it never really took off.

Except for one value, that seems at first glance to be completely pointless:

A rel value of "me": the linked resource has a relationship of being …me?

<a href="…" rel="me">

A is for Adactio

My website is adactio.com. I love my website. Even though it isn’t a physical thing, I think it might be my most prized possession.

It’s a place for me to think and a place for me to link.

But my web presence isn’t limited to adactio.com—I have profiles on lots of different sites: Twitter, Facebook, Instagram, Medium, Dribbble, Flickr, Tumblr, CodePen …the list goes on.

None of those places are as important to me as my own website, but they are representations of me.

I link out to these profiles from my own website using that rel value of "me". That’s me on Twitter. That’s me on Github. That’s me on Flickr.

<a rel="me" href="https://twitter.com/adactio">
<a rel="me" href="https://github.com/adactio">
<a rel="me" href="https://flickr.com/adactio">

Nothing unusual there. These are regular one-way hyperlinks.

What’s interesting is that many of those profiles on other websites provide a URL field where I can enter my own website. These third-party profiles then link back to my website also using a rel value of "me".

<a rel="me nofollow" href="https://adactio.com">

(They also use a rel value of "nofollow" to discourage spammers. The phrase "nofollow" makes absolutely no sense as a rel value—you can’t have a relationship of "nofollow" to anything—but it was invented by Google. We don’t get to argue with the 900 pound Google gorilla.)

Anyway, the result of having these reciprocal links, both using rel="me" means that we’ve kinda, sorta got two-way linking on the World Wide Web.

But …so what?

A is for Authentication

Some of those third-party profiles I’m linking to—Twitter, Github, Flickr—have something in common. They allow third-party authentication using OAuth.

If I can log into my Twitter, or Github, or Flickr profile using OAuth, and those profiles have two-way links with my website, then I can “borrow” that authentication flow for my own site.

That’s the idea behind IndieAuth. I enter the URL of my own website. It finds the links from there to my other profiles using rel="me". Then I can choose which one of those profiles I want to authenticate against. Once I’ve authenticated with that service, I’ve also authenticated my own website.

Just by adding a short rel value to some links, I can use my website as a log-in.

My website also has a small write API. The API is called micropub. By combining IndieAuth and micropub, I can log into somebody else’s posting interface using my website, and then use that interface to post to my own website.

These building blocks—microformats, IndieAuth, micropub—have emerged from a small but dedicated group of people.

We gather together at fun little events called Indie Web Camps.

An Indie Web Camp takes place over two days. The first day is split into discussions of technology and design. The second is all about implementing what we’ve discussed on the first day. I’m always amazed by how much gets done when you’re in the same space as like-minded people.

Still, the technologies being discussed and implemented aren’t the real focus of Indie Web Camp. The core tenet of Indie Web Camp is an idea. It’s a very simple idea, that at one point would have been uncontroversial. That idea is that you should have your own website.

But who’s got time for that? Especially when it’s so much easier to write and share and link using platforms like Twitter, Facebook, and Medium. That’s a good point. A lot of the time at an Indie Web Camp is spent reverse-engineering what those services are doing so well, and applying them to personal websites.

Also, we don’t necessarily want to stop using those services. After all, that’s where the people are (for now). Instead we want to figure out ways of making use of these services, while still keeping the canonical versions of what we create under our own control.

There’s a fantastic little Indie Web service called Brid.gy that allows you to not only post from your own site out to third party networks, but also receive replies, and likes, and retweets back at your own site, using another Indie Web building block called webmention.

In a way, webmentions allow a kind of two-way linking. I can cross-post something from adactio.com to Twitter, or Instagram, or Facebook. Then when someone replies on Twitter, or Instagram, or Facebook, I get notified with a ping back to my own site.

All of this is possible because I’ve verified the identity of those third-party profiles using nothing more than a simple rel attribute on a hyperlink.

The Indie Web uses a grab-bag of deliberately simple technologies that combine to create something so much more powerful than the sum of its parts …just like the web itself.

We often talk about places like Facebook, or Instagram, or Medium as walled gardens. Walled gardens are as old as the web itself. AOL was a well-cultivated walled garden back in the day.

To The Garden

There’s nothing wrong with walled gardens. They’re safe spaces. They take care of your enjoyment and entertainment, so you don’t have to.

But there also a bit boring. I certainly don’t relish the idea of spending my days within the boundaries of someone else’s vision.

There’s a different kind of garden. It takes its name from another short story by Borges.

The Garden of Forking Paths. It is uncontrolled. It is full of possibilities. It’s a bit scary. It takes more dedication to explore. You might get lost. But is that so bad? When was the last time you were truly lost on the World Wide Web, when you clicked through link after link—no cheating by opening new tabs, now—until you ended up somewhere, blinking and asking yourself “what I was looking for?”

I would like us all to spend more time in the garden of forking paths. I would like us all to continue to grow this garden of forking paths. Add your own website to this garden of forking paths. Use it to make more links.

On your website, you can link to this thing over here and that thing over there, and in doing so create an entirely new forking path.

Remember, the web, like the internet, has no centre. In theory I could start from any single A element, and by following all the forking paths, traverse the entire World Wide Web.

That opening hyperlink could be on your own website. One single A element can be the portal to an entire universe of knowledge.

Also on Medium

Responses

UIE

Awesome stuff. @adactio’s keynote from the inaugural HTML Special held before CSS Day 2016 in Amsterdam. ow.ly/t7Zx302ihbZ

# Posted by UIE on Tuesday, July 19th, 2016 at 4:31pm

Chris Aldrich

by Jeremy Keith from adactio.com

The opening keynote from the inaugural HTML Special held before CSS Day 2016 in Amsterdam.

If possible, click to play, otherwise your browser may be unable to play this audio file. The world exploded into a whirling network of kinships, where everything pointed to everything else, everything explained everything else. — Umberto Eco, Foucault’s Pendulum

I wasn’t able to attend the original presentation, but I think it’s even more valuable to listen to it all alone rather than in what was assuredly a much larger crowd. There is a wonderful presence in this brief history of the internet, made all the more intriguing by Jeremy’s performance as if it were poetry about technology. I find that he’s even managed to give it an interesting structured format, which, in many senses, mirrors the web itself.

I hope that if you’re starting your adventure on the web, that you manage to find this as one of the first links that starts you off on your journey. It’s a great place to start.

Syndicated copies to:

collect.readwriterespond.com

If possible, click to play, otherwise your browser may be unable to play this audio file.

Jeremy Keith provides a different introduction to the #IndieWeb. He maps a path from the beginning of the web, discussing apophenia, anchors, archive, all, Ada Lovelace, Alan Turing, Atlantic, augmentation, ARPA, attribute, Adactio and authentication.

Keith invites people to the world of forking paths:

I would like us all to spend more time in the garden of forking paths. I would like us all to continue to grow this garden of forking paths. Add your own website to this garden of forking paths. Use it to make more links.

On your website, you can link to this thing over here and that thing over there, and in doing so create an entirely new forking path.

# Tuesday, April 24th, 2018 at 3:20am

@rem

…is exactly where I read about it. I was cruising your blog late last night and came across it. My problem is that 22 letters means there must be phrases missing, which is making me sad.

# Posted by @rem on Sunday, March 24th, 2019 at 3:09pm

Chris Aldrich

A stark statement to make about one’s website. I feel exactly the same way.

Frank Meeuwsen

Wouter schrijft op zijn blog (en reageert zo automatisch op mijn blogpost, waardoor ik er van weet, magie!) over een vreemde gewoonte van sommige bloggers om álles te documenteren op het eigen domein. Als voorbeeld geeft Wouter mijn verzameling van bieren die ik drink en deel op Untappd. Maar tegelijk hier post door de magie van API’s.

that, in my opinion, ultimately end up polluting my RSS reader instead of enriching it. https://brainbaking.com/notes/2021/05/19h15m25s12/

Wouter vraagt zich vervolgens af of het bookmarken en liken van sites op je eigen site wel iets bijdraagt voor je lezers. Hij vraagt zich af: “Is dat wel bloggen?”

Blogs als logboeken

Nou Wouter… Als je de geschiedenis van bloggen en weblogs bekijkt, zul je ontdekken dat blogs juist zo zijn begonnen. Als de logboeken van nieuwsgierige techies die het internet ontdekken midden jaren ’90. Zij deden verslag van hun online reizen om zo anderen te laten zien wat er op het web is te vinden. Mijn eigen blog begon ik in 2000 met precies dat, links naar andere sites. Niets meer en niets minder. Over het algemeen wordt Robot Wisdom aangehaald als de eerste weblog. Wat niets anders was dan de links die schrijver Jorn Barger vond en wilde delen.

Zoals ik in de footer van de site schrijf en zo nu en dan hier aanhaal, ik probeer mijn site als een commonplace book te gebruiken. Dat gebeurt eerlijk gezegd nog te weinig. Als ik hier écht alles zou publiceren wat ik zo op een dag deel en doe, dan wordt het een firehose van informatie. Dat is een andere manier van bloggen dan we de afgelopen 10 jaar zien. Die is veel meer gericht op niches, kennisdeling, en op het goed vindbaar zijn op specifieke zoektermen. Daar is niets mis mee, maar er is meer mogelijk.

Verschillende feeds

Maar ik begrijp dat niet iedereen is geïnteresseerd in mijn bier-avonturen. Of de likes en bookmarks die hier opsla. Daar zijn trouwens wel aparte feeds voor! Hier is de feed voor likes en dit is de feed voor bookmarks.

Als auteur van deze blog kan ik eens kijken hoe ik de verschillende posts een betere taxonomie kan geven. Zodat je als lezer kunt kiezen of je alles wilt lezen minus een specifieke categorie. Of je wilt juist alleen een specifieke categorie lezen in je feedreader. De Post Kinds plugin biedt hier wel oplossingen, maar recent komen we tot de conclusie dat die voor een zeker lock-in zorgt die ongewenst is. Ik ga denken over een oplossing!

Als lezer kun je er voor kiezen om een feedreader te gebruiken die filters en regels ondersteunt. Zo kun je in je feedreader mijn algemene feed filteren op alles waar de titel start met “Untappd check-in” om de bierposts niet meer te zien. Ik begrijp dat dit een suboptimale oplossing is omdat je nu als lezer aan de slag moet. Aan de andere kant, in het dagelijkse leesleven filter je net zo goed onbewust wat je om je heen ziet.

Frankopedia

Deze site is primair mijn plek om te schrijven en publiceren wat ik wil bewaren en terugvinden. Bijna al mijn blogposts van de afgelopen 20 jaar zijn er te vinden. Er komt alleen maar meer bij en dat betekent dat voor verschillende lezers er verschillende ingangen moeten zijn. Dat is iets om aan te werken. Maar tegelijk, deze site is wie ik ben en wat ik tof vind. Het is de Frankopedia.

Jeremy Keith schreef het al eens mooi

My website is adactio.com. I love my website. Even though it isn’t a physical thing, I think it might be my most prized possession.

It’s a place for me to think and a place for me to link.

Jeremy Keith https://adactio.com/articles/10887

Dank je wel voor de feedback Wouter, ik waardeer het enorm. Het brengt mij weer nieuwe ideeën om als auteur te blijven doen wat ik het liefste doe en tegelijk mijn gewaardeerde lezers niet te overspoelen met mijn schrijvelarij, als ze dat niet willen!

James

While perusing the IndieWeb wiki, I stumbled upon a quote from a blog post written on adactio.com. The quote was:

My website is adactio.com. I love my website. Even though it isn’t a physical thing, I think it might be my most prized possession.

These words stayed with me after reading them. Jeremy (adactio.com) eloquently summarises a feeling I too have about my website: it is a home for my content that, despite being digital, means a great deal to me.

The phrase “internet home” is one that has stuck with me for a while. One could compare a website to a home in many ways. A “home page” is a page on the web that is yours. A personal website, by extension, is a set of pages that you own. Whether or not you built your website from scratch is not important: what matters is that you have something on the web that is on a domain you control. Having a website gives you freedom to control how your content is represented on the web, from the design down to the typography.

My personal home has expanded in scope a lot. I love working on new additions to my site, big and small. I am presently working on adding JSON Feed support to my site. Yes, adding another feed type to my site is a violation of the Don’t Repeat Yourself (DRY) principle but this is my home, a place for me to learn. I love my website and want to improve it in the ways that matter to me.

I know many bloggers have written about why they blog and have a website. I have in the past, too. I wanted to take a moment to share a reason that I don’t talk about as much, thanks to the prompt from Jeremy: running a website, while sometimes tedious, is a labour of love. I can write content and publish it using a design that I personally really like. I don’t worry about content accessibility on this site because everything is designed in such a way as to be easy for people to read and navigate. My site is a place where I can explore coding in a way that I otherwise cannot. I can work on things that matter to me on this site, whether it is new pages, ways to publish content, or anything else related to the inner workings behind this website.

If you have not yet started your own website, I would encourage you to do so and publish some content. See what you think. Blogging or having a website is not for everyone but you’ll never know if blogging will become your next big obsession — just like coffee is to me! — until you give it a try. If you do start a blog, let me know. I’d love to look around and see the way in which you choose to represent your content on your site.

Also posted on IndieNews.

# Posted by James on Tuesday, November 16th, 2021 at 12:00am

James

I am writing a blog post every day from December 1st to December 24th, 2021, about a blogger whose writing or site I follow. My aim for this series is to help you discover new blogs and to help get the word out about content creators whose blogs I appreciate. You can read more about this series in the inaugural post where I chat more about this series. Let’s begin!

Adactio (adactio.com)

Adactio is a blog run by Jeremy Keith, the founder of Clearleft. On the Adactio blog, you will find four different types of content: a journal with shorter articles, a list of links to resources Jeremy has liked, articles Jeremy has written, and shorter form notes. All of these categories have their own feed so you can pick and choose which one(s) you enjoy the most and subscribe. I have subscribed to three out of the four categories in my RSS feed.

Jeremy blogs mostly about the web, covering topics such as web tracking, using the right tools for the job, and his work writing for the web.dev blog. I have enjoyed posts on all of these topics. Jeremy’s extensive experience shines through in every article, presenting informed and well-articulated arguments for many areas that I find interesting.

I have found some terrific web resources on adactio.com over the last week. Posting links on quite a regular cadence, you’ll almost certainly find a link to a web resource or article that you will find interesting, as long as you are interested in the web. If you are in the mood for a long-form piece of content, look no further than Jeremy’s articles section. You’ll find plenty of written talks, presentations, and long-form content to perus.e

Adactio.com also comes with a terrific set of themes. I in awe of the themes available, which cover a wide range of historical web aesthetics and made me feel confident in offering a “retro” version of my site.

If you’re looking for a place to start on Adactio.com, I would recommend:

Today I also learned that Jeremy has been blogging for 20 years, making his blog one of the longest-running that I have seen to date. Amazing!

Wrapping Up

I love writing this series. For the last three days, one of the first things on my mind after waking up is “what blog am I going to feature today?” I have seen so many interesting websites in the last few years. If you ever feel like the web is all the same, I’d recommend checking out the IndieWeb or clicking through the websites I feature in this series. You’ll realise there is still a great deal of creative content on the web written by independent bloggers: you just have to know where to start looking. And this series will give you a good place to begin if you haven’t already!

I shall see you in tomorrow’s edition of this series. Until then, happy reading and exploring!

# Posted by James on Friday, December 3rd, 2021 at 12:00am

1 Share

# Shared by Giulia Laco on Tuesday, July 20th, 2021 at 9:26am

15 Likes

# Liked by Jay Robinson on Sunday, July 10th, 2016 at 7:17pm

# Liked by Jay Robinson on Sunday, July 10th, 2016 at 7:17pm

# Liked by Daniel Hudson on Thursday, May 24th, 2018 at 3:11am

# Liked by Christian Schaefer on Thursday, November 1st, 2018 at 11:13am

# Liked by @matthiasott@mastodon.social on Thursday, November 1st, 2018 at 11:41am

# Liked by CSS Day conference on Thursday, November 1st, 2018 at 11:41am

# Liked by Dale Blackburn on Sunday, March 24th, 2019 at 2:56pm

# Liked by Benjamin Read on Sunday, March 24th, 2019 at 4:45pm

# Liked by Responsive Web Design, or RWD on Monday, March 25th, 2019 at 8:03am

# Liked by Jonathan Basile on Thursday, April 11th, 2019 at 10:12pm

# Liked by Francis Rubio is coding CSS on Tuesday, July 20th, 2021 at 7:30am

# Liked by Florens Verschelde on Tuesday, July 20th, 2021 at 8:20am

# Liked by Losty on Tuesday, July 20th, 2021 at 9:22am

# Liked by Giulia Laco on Tuesday, July 20th, 2021 at 10:05am

# Liked by Maciej J on Wednesday, July 21st, 2021 at 8:46am

2 Bookmarks

# Bookmarked by Frank Meeuwsen on Sunday, May 23rd, 2021 at 12:00am

# Bookmarked by Elly Loel on Tuesday, June 27th, 2023 at 12:00am

Have you published a response to this? :