Summary: David Shanske and I recap the recent IndieWeb Summit 2018 in Portland Oregon including recent developments like microsub, readers, Vouch, and even the comeback of webrings!
Summary: Our first episode since January. We get caught up on some recent IndieWebCamps, an article about IndieWeb in The New Yorker, changes within WordPress, and upcoming events.
Post Kinds and new exclude functionality (🎧 00:48:15)
widgets
titleless posts
On this day
David’s list of 24 IndieWebCamps he’s attended
Looking back at past IndieWebCamp sessions and wiki pages for interesting ideas and new itches
Date and time stamps on webmentions Call for tickets in WordPress
Subscribing to h-cards with WebSub
Is Mastodon IndieWeb?
Fixing IndieAuth
Improving scoping, particularly for multi-user sites
Feel free to send us your questions or topic suggestions for upcoming episodes. (Use the comments below or your own site using Webmention).
Perhaps a future episode on Micro.blog?
I’m still tinkering away at pathways for following people (and websites) on the open web (in my case within WordPress). I’m doing it with an eye toward making some of the UI and infrastructure easier in light of the current fleet of Microsub servers and readers that will enable easier social reading without the centralized reliance on services like Facebook, Twitter, Instagram, Snapchat, Medium, LinkedIn, et al.
If you haven’t been following along, here are some relevant pieces for background:
Generally I’ve been adding data into my Following Page (aka blogroll on steroids) using the old WordPress Links Manager pseudo-manually. (There’s also a way to bulk import to it via OPML, using the WordPress Tools Menu or via /wp-admin/import.php?import=opml). The old Links Manager functionality in WordPress had a bookmarklet to add links to it quickly, though it currently only seems to add a minimal set–typically just the URL and the page title. Perhaps someone with stronger JavaScript skills than I possess could improve on it or integrate/leverage some of David Shanske’sParse This work into such a bookmark to pull more data out of pages (via Microformats, Schema.org, Open Graph Protocol, or Dublin Core meta) to pre-fill the Links Manager with more metadata including page feeds, which I now understand Parse This does in the past month or so. (If more than one feed is found, they could be added in comma separated form to the “Notes” section and the user could cut/paste the appropriate one into the feed section.) Since I spent some significant time trying to find/dig up that old bookmarklet, I’ll mention that it can be found in the Restore Lost Functionality plugin (along with many other goodies) and a related version also exists in the Link Library plugin, though on a small test I found it only pulled in the URL.
Since it wasn’t completely intuitive to find, I’ll include the JavaScript snippet for the Links Manager bookmarklet below, though note that the URL hard coded into it is for example.com, so change that part if you’re modifying for your own use. (I haven’t tested it, but it may require the Press This plugin which replaces some of the functionality that was taken out of WordPress core in version 4.9. It will certainly require one to enable using the Links Manager either via code or via plugin.)
Since I’ve been digging around a bit, I’ll note that Yannick Lefebvre’s Link Library plugin seems to have a similar sort of functionality to Links Manager and adds in the ability to add a variety of additional data fields including tags, which Ton Zijlstra mentions he would like (and I wouldn’t mind either). Unfortunately I’m not seeing any OPML functionality in the plugin, so it wins at doing display (with a huge variety of settings) for a stand-alone blogroll, but it may fail at the data portability for doing the additional OPML portion we’ve been looking at. Of course I’m happy to be corrected, but I don’t see anything in the documentation or a cursory glance at the code.
In the most ideal world, I’d love to be able to use the Post Kinds Plugin to create follow posts (see my examples). This plugin is already able to generally use bookmarklet functionality to pull in a variety of meta data using the Parse This code which is also built into Post Kinds.
It would be nice if these follow posts would also copy their data into the Links Manager (to keep things DRY), so that the blogroll and the OPML files are automatically updated all at once. (Barring Post Kinds transferring the data, it would be nice to have an improved bookmarklet for pulling data into the Links Manager piece directly.)
Naturally having the ability for these OPML files be readable/usable by Jack Jamieson’s forthcoming Yarns Microsub Server for WordPress (for use with social readers) would be phenomenal. (I believe there are already one or two OPML to h-feed converters for Microsub in the wild.) All of this would be a nice end -to-end solution for quickly and easily following people (or sites) with a variety of feeds and feed types (RSS, Atom, JSONfeed, h-feed).
An additional refinement of the blogroll display portion would be to have that page display as an h-feed of h-entries each including properly marked up h-cards with appropriate microformats and discoverable RSS feeds to make it easier for other sites to find and use that data. (This may be a more IndieWeb-based method of displaying such a page compared with the OPML spec.) I’ll also note that the Links Manager uses v1 of the OPML spec and it would potentially be nice to have an update on that as well for newer discovery tools/methods like Dave Winer’sShare Your OPML Subscription list, which I’m noting seems to be down/not functioning at the moment.
Micropub Plugin work for WordPress
It will include a Media endpoint
Code for integration with the WordPress REST API
rel=”alternate”
This sketch solution may be an end-around the issue of getting WordPress (or potentially other CMSes) Themes to be microformats 2 compatible, and allow a larger range of inter-compatibility for websites and communication.
Last week was the 8th annual IndieWeb Summit held in Portland, Oregon. While IndieWeb Camps and Summits have traditionally been held on weekends during people’s free time, this one held in the middle of the week was a roaring success. With well over 50 people in attendance, this was almost certainly the largest attendance I’ve seen to date. I suspect since people who flew in for the event had really committed, the attendance on the second day was much higher than usual as well. It was great to see so many people hacking on their personal websites and tools to make their personal online experiences richer.
The year of the Indie Reader
Last year I wrote the post Feed Reader Revolution in response to an increasingly growing need I’ve seen in the social space for a new sort of functionality in feed readers. While there have been a few interesting attempts like Woodwind which have shown a proof-of-concept, not much work had been done until some initial work by Aaron Parecki and a session at last year’s IndieWeb Summit entitled Putting it all Together.
I spent a few minutes in a session at the end of Tuesday and managed to log into Aperture and create an account (#16, though I suspect I may be one of the first to use it besides the initial group of five developers). I also managed to quickly and easily add a microsub endpoint to my website as well. Sadly I’ve got some tweaks to make to my own installation to properly log into any of the reader app front ends. Based on several of the demos I’ve seen over the past months, the functionality involved is not only impressive, but it’s a properly large step ahead of some of the basic user interface provided by the now-shuttered Woodwind.xyz service (though the code is still available for self-hosting.)
Several people have committed to make attempts at creating a microsub server including Jack Jamieson who has announced an attempt at creating one for WordPress after having recently built the Yarns reader for WordPress from scratch this past year. I suspect within the coming year we’ll see one or two additional servers as well as some additional reading front ends. In fact, Ryan Barrett spent the day on Wednesday hacking away at leveraging the News Blur API and leveraging it to make News Blur a front end for Aperture’s server functionality. I’m hoping others may do the same for other popular readers like Feedly or Inoreader to expand on the plurality of offerings. Increased competition for new reader offerings can only improve the entire space.
Even more reading related support
Just before the Summit, gRegor Morrillunveiled the beta version of his micropub client Indiebookclub.biz which allows one to log in with their own website and use it to post reading updates to their own website. For those who don’t yet support micropub, the service saves the data for eventual export. His work on it continued through the summit to continue to improve an already impressive product. It’s the fist micropub client of its kind amidst a growing field of websites (including WordPress and WithKnown which both have plugins) that offer reading post support. Micro.blog has recently updated its code to allow users of the platform the ability to post reads with indiebookclub.biz as well. As a result of this spurt of reading related support there’s now a draft proposal to add read-of and read-status support as new Microformats. Perhaps reads will be included in future updates of the post-type-discovery algorithm as well?
Given the growth of reading post support and a new micropub read client, I suspect it won’t take long before some of the new microsub-related readers begin supporting read post micropub functionality as well.
IndieAuth Servers
In addition to David Shanske’s recent valiant update to the IndieAuth plugin for WordPress, Manton Reece managed to finish up coding work to unveil another implementation of IndieAuth at the Summit. His version is for the micro.blog platform which is a significant addition to the community and will add several hundred additional users who will have broader access to a wide assortment of functionality as a result.
The Future
While work continues apace on a broad variety of fronts, I was happy to see that my proposal for a session on IndieAlgorithms was accepted (despite my leading another topic earlier in the day). It was well attended and sparked some interesting discussion about how individuals might also be able to exert greater control over what they’re presented to consume. With the rise of Indie feed readers this year, the ability to better control and filter one’s incoming content is going to take on a greater importance in the very near future. With an increasing number of readers to choose from, more people will hopefully be able to free themselves from the vagaries of the blackbox algorithms that drive content distribution and presentation in products like Facebook, Twitter, Instagram and others. Based on the architecture of servers like Aperture, perhaps we might be able to modify some of the microsub spec to allow more freedom and flexibility in what will assuredly be the next step in the evolution of the IndieWeb?
Diversity
While there are miles and miles to go before we sleep, I was happy to have seen a session on diversity pop up at the Summit. I hope we can all take the general topic to heart to be more inclusive and actively invite friends into our fold. Thanks to Jean for suggesting and guiding the conversation and everyone else for continuing it throughout the rest of the summit and beyond.
Other Highlights
Naturally, the above are just a few of the bigger highlights as I perceive them. I’m sure others will appear in the IndieNews feed or other blogposts about the summit. The IndieWeb is something subtly different to each person, so I hope everyone takes a moment to share (on your own sites naturally) what you got out of all the sessions and discussions. There was a tremendous amount of discussion, debate, and advancement of the state of the art of the continually growing IndieWeb. Fortunately almost all of it was captured in the IndieWeb chat, on Twitter, and on video available through either the IndieWeb wiki pages for the summit or directly from the IndieWeb YouTube channel.
I suspect David Shanske and I will have more to say in what is sure to be a recap episode in our next podcast.
Photos
Finally, below I’m including a bunch of photos I took over the course of my trip. I’m far from a professional photographer, but hopefully they’ll give a small representation of some of the fun we all had at camp. (The photos can be found at the original post.)
Final Thanks
People
While I’m thinking about it, I wanted to take a moment to thank everyone who came to the summit. You all really made it a fantastic event!
I’d particularly like to thank Aaron Parecki, Tantek Çelik, gRegor Morrill, Marty McGuire, and David Shanske who did a lot of the organizing and volunteer work to help make the summit happen as well as to capture it so well for others to participate remotely or even view major portions of it after-the-fact. I would be remiss if I didn’t thank Martijn van der Ven for some herculean efforts on IRC/Chat in documenting things in real time as well as for some serious wiki gardening along the way. As always, there are a huge crew of others whose contributions large and small help to make up the rich fabric of the community and we wouldn’t be who we are without your help. Thank you all! (Or as I might say in chat: community++).
And finally, a special personal thanks to Greg McVerry for kindly letting me join him at the Hotel deLuxe for some late night discussions on the intersection of IndieWeb and Domain of One’s Own philosophies as they dovetail with the education sector. With growing interest and a wealth of ideas in this area, I’m confident it’s going to be a rapidly growing one over the coming years.
Sponsors
I’d also like to take a moment to say thanks to all the sponsors who helped to make the event a success including Name.com, GoDaddy, Okta, Mozilla, DreamHost, and likely a few others who I’m missing at the moment.
I’d also like to thank the Eliot Center for letting us hosting the event at their fabulous facility.
In the recent Twitter Migration, in addition to trying out Mastodon, I’ve been seeing some people go back to blogs or platforms like Micro.blog, WordPress, Tumblr, WriteFreely (like Mastodon it’s a part of the Fediverse, but built for blogging instead of short posts) and variety of others. They’re looking for a place where they can truly own and share their content, often in healthier and more humane ways. Many are extolling the virtues of posting on their own website so that they own their content to protect against the sort of platform problems many are now seeing and experiencing on the rapidly dying birdsite. I’ve seen a growing number of people in/on several platforms reviving the early Twitter practice of #FollowFriday to help people discover new and interesting people to follow.
As a result, while everyone is exploring new platforms and new online spaces for maintaining their identities and communicating, I’m going to suggest something else interesting to shift our online social patterns: Instead of spending time on Twitter, Mastodon, Instagram, or other major social platforms, start practicing #FeedReaderFriday by carving out some time to find and follow people’s websites directly with a feed reader or social reader. Then engage with them directly on their own websites.
I already spend a reasonable amount of time in a variety of readers looking at both longform articles as well as social media posts (status updates, notes, bookmarks, and photos), but starting this Friday, I’m going to practice #FeedReaderFriday. Instead of opening up Twitter or Mastodon, I’ll actively and exclusively reach for one of my feed readers to read people’s content and respond to them directly.
As part of the effort, I’ll share people’s sites I follow and enjoy. I’ll also suggest some feed readers to try out along with other related resources. I’ll use the tag/hashtag #FeedReaderFriday to encourage the website to website conversation. If you’re interested in the experiment, do come and join me and help to spread the word.
Just as many people are doing the sometimes difficult but always rewarding emotional labor of helping people migrate from the toxicity of Twitter and its algorithmic feeds, perhaps those of us who have websites and use social readers could help our friends and family either set up their own spaces or onboard them to social readers in this effort? Mastodon’s decentralized nature is an improvement and provides a reasonable replacement for Twitter, but eventually people will realize some of the subtle issues of relying on someone else’s platform just as they’ve seen issues with Twitter, Facebook, Instagram, or the now defunct Google+.
Feel like you’ll miss people’s content on traditional social media? There are definitely a variety of ways to follow them in a variety of feed and social readers. Not sure what RSS is? Feel free to ask. Know of some interesting tricks and tools you use to make discovering and subscribing to others’ blogs easier? Share them! Have fantastic resources for discovering or keeping up with others’ websites? Share those too. Not quite sure where to begin? Ask for some help to better own your online identity and presence.
It may be a slow start, but I think with some care, help, and patience, we can help to shift both our own as well as others’ online social reading and correspondence habits to be kinder, smarter, and more intentional.
What will you read on #FeedReaderFriday? Who will you recommend following?
For some brief background, my following page is built on some old functionality in WordPress core that has since been hidden. I’m using the old Links Manager for collecting links and feeds of people, projects, groups, and institutions. This link manager creates standard OPML files, which WordPress can break up by categories, that can easily be imported/exported into most standard feed readers. Even better, some feed readers like Inoreader, support OPML subscriptions, so one could subscribe to my OPML file, and any time I update it in the future with new subscriptions, your feed reader would automatically update to follow those as well. I use this functionality in my own Inoreader account, so that any new subscriptions I add to my own site are simply synced to my feed reader without needing to be separately added or updated.
The best part of creating such a list and publishing it in a standard format is that you, dear reader, don’t need to spend the several hours I did to find, curate, and compile the list to recreate it for yourself, but you can now download it, modify it if necessary, and have a copy for yourself in just a few minutes. (Toward that end, I’m also happy to update it or make additions if others think it’s missing anyone interesting in the space–feedback, questions, and comments are heartily encouraged.) You can see a human-readable version of the list at this link, or find the computer parse-able/feed reader subscribe-able link here.
To make it explicit, I’ll also note that these lists also help me to keep up with people and changes in the timeframe between conferences.
Anecdotal Domains observations
In executing this OPML project I noticed some interesting things about the Domains community at large (or at least those who are avid enough to travel and attend in person or actively engage online). I’ll lay these out below. Perhaps at a future date, I’ll do a more explicit capture of the data with some analysis.
The largest majority of sites I came across were, unsurprisingly, WordPress-based, which made it much easier to find RSS feeds to read/consume material. I could simply take a domain name and add /feed/ to the end of the URL, and voilà, a relatively quick follow!
There are a lot of people whose sites didn’t have obvious links to their feeds. To me this is a desperate tragedy for the open web. We’re already behind the eight ball compared to social media and corporate controlled sites, why make it harder for people to read/consume our content from our own domains? And as if to add insult to injury, the places on one’s website where an RSS feed link/icon would typically live were instead populated by links to corporate social media like Facebook, Twitter, and Instagram. In a few cases I also saw legacy links to Google+ which ended service and disappeared from the web along with a tremendous number of online identities and personal data on April 2, 2019. (Here’s a reminder to remove those if you’ve forgotten.) For those who are also facing this problem, there’s a fantastic service called SubToMe that has a universal follow button that can be installed or which works well with a browser bookmarklet and a wide variety of feed readers.
I was thrilled to see a few people were using interesting alternate content management systems/site generators like WithKnown and Grav. There were also several people who had branched out to static site generators (sites without a database). This sort of plurality is a great thing for the community and competition in the space for sites, design, user experience, etc. is awesome. It’s thrilling to see people in the Domains space taking advantage of alternate options, experimenting with them, and using them in the wild.
I’ll note that I did see a few poor souls who were using Wix. I know there was at least one warning about Wix at the conference, but in case it wasn’t stated explicitly, Wix does not support exporting data, which makes any potential future migration of sites difficult. Definitely don’t use it for any extended writing, as cutting and pasting more than a few simple static pages becomes onerous. To make matters worse, Wix doesn’t offer any sort of back up service, so if they chose to shut your site off for any reason, you’d be completely out of luck. No back up + no export = I couldn’t recommend using.
If your account or any of your services are cancelled, it may result in loss of content and data. You are responsible to back up your data and materials. —Wix Terms of Use
I also noticed a few people had generic domain names that they didn’t really own (and not even in the sense of rental ownership). Here I’m talking about domain names of the form username.domainsproject.com. While I’m glad that they have a domain that they can use and generally control, it’s not one that they can truly exert full ownership over. (They just can’t pick it up and take it with them.) Even if they could export/import their data to another service or even a different content management system, all their old links would immediately disappear from the web. In the case of students, while it’s nice that their school may provide this space, it is more problematic for data portability and longevity on the web that they’ll eventually lose that institutional domain name when they graduate. On the other hand, if you have something like yourname.com as your digital home, you can export/import, change content management services, hosting companies, etc. and all your content will still resolve and you’ll be imminently more find-able by your friends and colleagues. This choice is essentially the internet equivalent of changing cellular providers from Sprint to AT&T but taking your phone number with you–you may change providers, but people will still know where to find you without being any the wiser about your service provider changes. I think that for allowing students and faculty the ability to more easily move their content and their sites, Domains projects should require individual custom domains.
If you don’t own/control your physical domain name, you’re prone to lose a lot of value built up in your permalinks. I’m also reminded of here of the situation encountered by faculty who move from one university to another. (Congratulations by the way to Martha Burtis on the pending move to Plymouth State. You’ll notice she won’t face this problem.) There’s also the situation of Matthew Green, a security researcher at Johns Hopkins whose institutional website was taken down by his university when the National Security Agency flagged an apparent issue. Fortunately in his case, he had his own separate domain name and content on an external server and his institutional account was just a mirrored copy of his own domain.
If you’ve got it, flaunt it.
—Mel Brooks from The Producers (1968), obviously with the it being a referent to A Domain of One’s Own.
Also during my project, I noted that quite a lot of people don’t list their own personal/professional domains within their Twitter or other social media profiles. This seems a glaring omission particularly for at least one whose Twitter bio creatively and proactively claims that they’re an avid proponent of A Domain of One’s Own.
And finally there were a small–but still reasonable–number of people within the community for whom I couldn’t find their domain at all! A small number assuredly are new to the space or exploring it, and so I’d give a pass, but I was honestly shocked that some just didn’t.
(Caveat: I’ll freely admit that the value of Domains is that one has ultimate control including the right not to have or use one or even to have a private, hidden, and completely locked down one, just the way that Dalton chose not to walk in the conformity scene in The Dead Poet’s Society. But even with this in mind, how can we ethically recommend this pathway to students, friends, and colleagues if we’re not willing to participate ourselves?)
Too much Twitter & a challenge for the next Domains Conference
One of the things that shocked me most at a working conference about the idea of A Domain of One’s Own within education where there was more than significant time given to the ideas of privacy, tracking, and surveillance, was the extent that nearly everyone present gave up their identity, authority, and digital autonomy to Twitter, a company which actively represents almost every version of the poor ethics, surveillance, tracking, and design choices we all abhor within the edtech space.
Why weren’t people proactively using their own domains to communicate instead? Why weren’t their notes, observations, highlights, bookmarks, likes, reposts, etc. posted to their own websites? Isn’t that part of what we’re in all this for?!
One of the shining examples from Domains 2019 that I caught as it was occurring was John Stewart’s site where he was aggregating talk titles, abstracts, notes, and other details relevant to himself and his practice. He then published them in the open and syndicated the copies to Twitter where the rest of the conversation seemed to be happening. His living notebook– or digital commmonplace book if you will–is of immense value not only to him, but to all who are able to access it. But you may ask, “Chris, didn’t you notice them on Twitter first?” In fact, I did not! I caught them because I was following the live feed of some of the researchers, educators, and technologists I follow in my feed reader using the OPML files mentioned above. I would submit, especially as a remote participant/follower of the conversation, that his individual posts were worth 50 or more individual tweets. Just the additional context they contained made them proverbially worth their weight in gold.
Perhaps for the next conference, we might build a planet or site that could aggregate all the feeds of people’s domains using their categories/tags or other means to create our own version of the Twitter stream? Alternately, by that time, I suspect that work on some of the new IndieWeb readers will have solidified to allow people to read feeds and interact with that content directly and immediately in much the way Twitter works now except that all the interaction will occur on our own domains.
Seeing good examples of existing domains is crucial for showing students what is possible in creating their own domain, says @CassieNooyen#domains19
As educators, one of the most valuable things we can and should do is model appropriate behavior for students. I think it’s high time that when attending a professional conference about A Domain of One’s Own that we all ought to be actively doing it using our own domains. Maybe we could even quit putting our Twitter handles on our slides, and just put our domain names on them instead?
Of course, I wouldn’t and couldn’t suggest or even ask others to do this if I weren’t willing and able to do it myself. So as a trial and proof of concept, I’ve actively posted all my interactions related to Domains 2019 that I was interested in to my own website using the tag Domains 2019. At that URL, you’ll find all the things I liked and bookmarked, as well as the bits of conversation on Twitter and others’ sites that I’ve commented on or replied to. All of it originated on my own domain, and, when it appeared on Twitter, it was syndicated only secondarily so that others would see it since that was where the conversation was generally being aggregated. You can almost go back and recreate my entire Domains 2019 experience in real time by following my posts, notes, and details on my personal website.
So, next time around can we make an attempt to dump Twitter!? The technology for pulling it off certainly already exists, and is reasonably well-supported by WordPress, WithKnown, Grav, and even some of the static site generators I noticed in my brief survey above. (Wix obviously doesn’t even come close…)
I’m more than happy to help people build and flesh out the infrastructure necessary to try to make the jump. Even if just a few of us began doing it, we could serve as that all-important model for others as well as for our students and other constituencies. With a bit of help and effort before the next Domains Conference, I’ll bet we could collectively pull it off. I think many of us are either well- or even over-versed in the toxicities and surveillance underpinnings of social media, learning management systems, and other digital products in the edtech space, but now we ought to attempt a move away from it with an infrastructure that is our own–our Domains.
Just as I was getting sick last week, Colin Walker wrote “There has to be a better way to subscribe to sites.” He’s definitely hit the nail right on the head. The process is currently painful and disorganized, it’s also working on technology that’s almost two decades old and difficult for newcomers at best.
I’ve always posited that one of the reasons that social media silos have been so successful is that they’ve built some fantastic readers. Sure their UI is cleaner and just dead simple, but to a great extent 95% of their product is an evolved feed reader while the other 5% is a simple posting interface that makes it easy to interact. To compare, most CMSes are almost completely about posting interface, and spend very little time, if any, worrying about providing a reading experience.
The IndieWeb has been making some serious strides on making cross-site interactions easier with the Webmention and Micropub protocols, but the holy grail is still out there: allowing people to have an integrated feed reader built into their website (or alternately a standalone feed reader that’s tightly integrated with their site via Micropub or other means).
For those watching the space with as much interest as I have, there are a couple of interesting tools in the space and a few on the immediate horizon that are sure to make the process a whole lot easier and create a new renaissance in the open web.
SubToMe: a Universal Subscribe Button
First, for a relatively simple one-size-fits-all subscribe button, I recommend people take a look at SubToMe which touts itself as a “Universal Follow button” because it “makes it easy for people to follow web sites,because browsers don’t do it.” The button is fairly straightforward and has an awful lot of flexibility built in. In the simplest sense it has some solid feed detection so it finds available feeds on a web page and then provides a handful of recommended major readers to the user. With two clicks, one can pretty quickly and almost immediately subscribe to almost any feed in their reader of choice.
For publishers, one can quickly install a simple button on their site. They can further provide a list of specific feeds they want to advertise, and they can even recommend a particular feed reader if they choose.
For consumers, the service provides a simple browser bookmarklet so that if a site doesn’t have a button, they can click a subscribe button in their browser. Then click on a provider. Done. One can also choose a preferred provider to shorten the process.
Almost all the major feed readers are supported out of the box and the process of adding new ones is relatively simple.
Microsub
Since last June there’s been a quietly growing new web spec called Microsub that will assuredly shake up the subscription and reader spaces. In short it provides a standardized way for clients to consume and interact with feeds collected by a server.
While it gets pretty deep pretty quickly, the spec is meant to help decouple some of the heavy architecture of building a feed reader. In some way it’s analogous to the separation of content and display that HTML and CSS allows, but applied to the mechanics of feed readers and how readers display their content.
There are already a few interesting projects by the names of Together and Indigenous that are taking advantage of the architecture
I can’t wait to see how it all dovetails together to make a more integrated reading and posting interface as well as the potential it has for individual CMSs to potentially leverage the idea to include integrated interfaces into their products. I can’t wait for the day when my own personal website is compatible with Microsub, so that I can use any Microsub client to read my timeline and follow people.
I’m also sure that decoupling the idea of displaying posts from actually fetching remote feeds will make it easier to build a reader clients in general. I hope this has a Cambrian explosion-type of effect on the state of the art of feed readers.
I’d recommend those interested in a high level discussion to have a listen to the following thee short episodes of Aaron Parecki’sPercolator microcast.
I’ve been slowlybutsurely working on compiling a list of people I’m following online. In older iterations of the web, this would have been known as a blogroll, but I think it’s time to update the concept and potentially add some new features and functionality to it. It’s also time to upgrade its status on my site, so I’m moving it from a widgetized sidebar area on my front page to its own page under my “About” menu.
I’ve read books like Clay Johnson’s Information Diet: a Case for Conscious Consumption and P.M. Forni’s excellent Thinking Life: How to Thrive in the Age of Distractionwhich describe an overwhelming media and online social atmosphere with some prescriptive measures for cutting down on the noise. More people obviously need this type of advice and I’m regularly thinking about how to cut down on the noise and get more valuable signal out of my online tools.
An Inventory of Sources
As a result of all this noise from too many sources and social platforms, I’ve found that having a manifest or complete inventory of all my online reading sources can be immensely valuable. It will make it easier to see what I’m reading and consuming on a regular basis and therefor easier to prune or update this list based on how often I’m reading these sources compared to the value I’m getting out of them.
I can look at the titles of the sources and better get a feel for exactly what I’m consuming and possibly how much. Those I don’t read as often can be pruned out of the list or can serve as a reminder of why I wanted to add them in the first place and what I wanted to get out of them. Better that I be nagged to read things I know I’ll get value out of than defaulting to the fast food-esque fluff that, like many others, I turn to on Facebook, Twitter, and Instagram because it’s “easy” to consume.
I’ve also now compiled a year’s worth of reading data for things that I’ve read online. I’ve saved links to literally everything I’ve read in the past full calendar year to my website (though I only choose to show a subsection of those links to the public). This has given me a more solid data set of what I’ve read and interacted with to better guide my decisions about what I should put on the list and what I shouldn’t.
Notifications
As for the notification overload, by moving some of my reading onto my site via the excellent PressForward reader, I can drastically cut down on the number of notifications I get in email or via phone. I can more directly control exactly which notifications (and when they’re sent) that are originating from my own website.
Fighting Algorithms (and winning!)
Over the past few years, we’ve seen the rapid rise of algorithms. In some cases they’ve provided worthwhile improvements to our lives, while in others they’re downright malicious and destructive. This has become drastically more apparent in the past year or so, and I invite those who aren’t aware of their dramatic effects on our lives to read Cathy O’Neil’s book Weapons of Math Destruction, which does a great job of outlining them for the lay person with no technical background.
Every day these black box algorithms are choosing more and more of what we read and consume. (The only thing worse than the lack of a free press coupled with government controlled media is a corporate algorithmically controlled media which gives you the illusion of freedom.) Because most companies that are using these algorithms in the social space are doing so to keep us more “engaged” and on their sites for longer and clicking their ads with out any transparency, I can no longer trust them. My goals and ideals when reading online content are drastically different than theirs. I want to become more informed, challenged, and made to think. I don’t want their programmatic “reversion to the mean” forcing me to read more memes, jokes, political vitriol, and useless content.
To fight these algorithms, particularly those found in Facebook and to a lesser extent in Twitter, I’m going to cut them off at the knees and consciously choose a set of specific streams to read and engage with. Because I control what goes in to the system, I’ll know exactly what comes out. To touch on the food analogy again, when I cook for myself, I know exactly what the ingredients are and can thus eat a more healthy and well-balanced diet compared to going out and eating fast-food where I’m not ever quite sure if the “beef” is really beef, much less if it’s safe. Yes, I’ll say it, I’m going to go both organic as well as farm-to-table in my online social life.
Twitter thought experiment
Initially I had contemplated declaring Twitter bankrupcy. It seemed like a brilliant and cathartic-ly wonderful idea! But cleaning out my Twitter feed to a much smaller subset ultimately seemed like way too much work. I can only think about the hours and hours of time I’ve spent even creating and categorizing Twitter feeds into lists on my account. (Fortunately others can also follow those curated lists to find some value, so it’s not a total loss.) Starting over again from scratch on my main feed seemed untenable. Even if I did clean it all out, I would potentially have a better feed, but it’s still a feed on a silo which I don’t own or control and it doesn’t have any effect on needing to repeat the same work on dozens of other silos. Heavy pruning and weeding within someone else’s walled garden seemed like a painful and unscalable time-suck that I would potentially need to repeat on an ongoing basis. It’s akin to the sharecropping of content that I had previously been doing for them and refuse to continue to do so.
The better option seems to be to use open web technologies to create and maintain my own personal list. It’s something I own and can control. I can update it as often as I want. Even better, I only need to do it in one place instead of dozens and the results can be distributed across multiple sites almost instantaneously!
As I’ll also discuss below, my open list is still easily shareable and modifiable by others. So I’m not accruing benefits just to myself, but my work can become scale-able and usable by others.
In creating my list I wanted to go above the traditional blogroll and add additional context that most of them often didn’t originally have. I’ve tried to add a photo, logo, or avatar of some sort for all the sources to provide some visual context. I’ve also added either a description of the site or a snippet from the site’s owner to give an idea of what it is about (in addition to categorizing them by one or more tags) as well as an optional reason why I’m following them. I’ve also included a link to the site as well as an RSS, atom, or h-entry feed for the site to make subscribing easier for others. Where appropriate, I’ve added the microformats XFN data to these sites as well so others will have an idea of my relationship to those entities or people I’m following. Disclosure is a good thing, right? Just ask a journalist. (Viewing this last part is currently only available via parsers or by viewing the page source within a browser, but it’s there for potential future use.) In aggregate, these bits of context are not only valuable for page viewers who are considering subscribing/following them for themselves, but they also make a statement about me as a reader, a topic I’ll touch on further below.
Promotion of position: from sidebar to a full page
Given the value of social following/friending in the past decade, it’s long overdue to promote the old-school blogroll, which was traditionally placed in a diminutive position in one’s sidebar, to a more prominent position on its own page (or others may even choose to span it over multiple pages).
Social media platforms do their best to hide our social graphs from us thereby making more of what they do seem magical. Many have even bent over backwards to prevent other possibly competing social startups from leveraging our own social graphs on their platform to help build them up. Just where do they think that data came from initially? It came from me! I own it and should continue owning it.
To that end, my follow list in some sense is an implicit statement of me owning that data once again. While it may take me a bit to import and arrange it all, I’ll have ownership and agency over it. Perhaps an outside service may want pieces or parts of it, and in some cases having it open and portable may provide continued future value to me.
As an analogy for what this means, think back to the days of arduously making mix tapes in the 80’s. You’d spend hours and hours diligently copying and pasting songs together onto a cassette tape to give to a favored someone. The gift usually meant more than just the songs on the tape. This type of thing is far easier now with digital music services to the point of devaluing part of the original meaning of a mix tape. However, almost no modern music service will allow you to take your hand-crafted playlists out of their service to other competing services to make it easier to switch from something like iTunes to Amazon Music or Google Music. It’s painful and annoying in an age chock full of digital exhaust. I’m hoping that my open following list might be a lot like the portable digital music play list I wish I had.
Identity
I’m placing my follow list as a submenu item underneath my “About Me” page. Why? On most social networks there are a few simple fields, typically in a profile or on an explicit profile page, which give others some basic data about who the account holder is and what they do. Often people use this data to make relatively quick decisions about whether they should follow (or follow back) another person. Sadly I’m of the opinion that the amount and richness of the data on these pages is too sparse to be of much use. Fortunately by owning my own site, I can remedy this problem for others who visit it.
My website has thousands to potentially hundreds of thousands of posts. What data can I easily provide people who are interested in learning more about me without reading the whole book as it were? My About page is a good quick place to start, but it can’t necessarily give the whole picture. I’ve also got a few other sub-pages under my About page which helps to round out the snapshot picture of who I am. These include:
my /now page, which tells others what I’m up to most recently, but at a higher level than reading a month’s worth of status updates;
my /Favorites page, which is a list of some of my favorite things and things I use on a regular basis; it’s not dissimilar to a “What I’m Using” page or regular posts concept;
my /Bucketlist page, which is a list of some things I’ve done or would like to do before I “kick the bucket”;
I’ve also recently added an /AMA or Ask Me Anything page, so that if there’s something pressing you need to know that isn’t written or find-able on my site, you can easily ask it.
Finally, there’s now also a source for others to quickly see what I’m regularly reading and find valuable enough on the web to have created a list of it all.
I think that in evaluating others, this last page (the following page) may actually provide the most value, and so I hope it does to others in return. I can’t help noting here how I’ll often judge others by which books they have on their shelves at home, or this great judgmental quote from John Waters:
“If you go home with somebody, and they don’t have books, don’t fuck ’em!”
I hope others I’m following will follow suit and create their own following pages as I’d honestly love nothing more than to know who and what they find valuable, and to be able to extract it quickly to add to my own list! The value of discovery here can be tremendous.
Intellectual Antecedents
I know that academics like to give credit to their sources when writing papers, though they often do so in explicit footnote form. Abstracted out to a more general form, I’m hoping that my following page can also help to provide some meta data about which sources I regularly find valuable and which ones are most likely influencing me even if they’re not explicitly footnoted within my writings.
Benefit of following members of the IndieWeb
Having been using a version of my following page for a while, I’ve found one particularly nice feature of following people who are adherents of the IndieWeb movement. Because they’ve chosen to post on their own site first (and optionally syndicate to other silos), their internet presence is far more centralized for subscription and consumption. I don’t have to follow them on dozens of multiple social silos to attempt to capture all their content. I can subscribe in one place and get as much or as little as I like! You can do much the same with my site, which I’ve discussed in the past.
Now of course this isn’t the case for everyone yet, and there can be some exceptions (since not everyone owns every post-type yet nor has quit all their silos), but it does tremendously cut down on the noise, cruft, and duplicated messages that live on multiple platforms.
I’ve experimented in the past with following even a subset of researchers and their work online. The amount of time needed to catalog them all, find their various presences in sites like Facebook, Twitter, LinkedIn, Google+, Academia, ResearchGate, etc., etc. was painful, but then setting up notifications and creating a workflow was even worse–particularly since I want to read or see everything they’re putting out over time. I think I’d have been better off building them all custom websites to publish their content instead.
OPML means sharing
OPML really stands for Outline Processor Markup Language. It is an XML-based format and standard used for feed lists interchange. All this to mean that it’s a standardized specially formatted document that allows one to share all the data in it easily by means that make sense to certain machines that would want it.
The most common example is that most feed readers allow you to import and export OPML files (with the .xml extension) so that you can quickly and easily move all of your feeds from one reader to another. (This is kind of like the playlist analogy for music that I mentioned–it’s just a playlist, or readlist if you prefer, for feed readers.) This is great if you want to try out a or move to a new feed reader.
Even better, because you can find and save a copy of my list, others can easily port it into their feed readers and sample the things that I’m seeing and often reading.
But wait! There’s more…
Many modern feed readers are supporting OPML subscription functionality! (What’s that you ask?) It’s fine to download my OPML list and import it into your reader. But what happens when I update it next week with three new great sources and remove a dead feed that no longer works? You’re stuck missing out on the new stuff and have to manually find and remove the broken one yourself. Instead, if you’ve subscribed to my OPML in your feed reader, the reader knows the URL where my list lives and checks it frequently for updates so you don’t have to worry about syncing the changes yourself! Shazam! It’s now a lot like a shared/synced playlist for articles. For those who are familiar with Twitter lists and following those, it’s very similar to how those work, except in this case they’re open and work on multiple sites and apps instead of being stuck in a proprietary service.
How
Now the part you’ve been waiting for: How can I do this myself?
For those who are on WordPress, much of the base functionality is already built into WordPress core. Below I’ll provide a few means and tips for getting you most of the way while still having some flexibility in where and how you choose to display your particular version.
(For those not on WordPress, check out some of the details and documentation on the IndieWeb wiki and ask in their chat how you might go about doing it.)
Re-enable Links Manager interface
The code for the WordPress blogroll functionality was built into core and was known as the Links Manager, but it was removed in version 3.5 for new installs that didn’t have any pre-existing links. I’ll note that the functionality was removed in late 2012 long after social media had already begun to make functionality like blogrolls (and even blogs themselves) fall out of fashion.
Fortunately, while it’s now hidden for most, it can be brought back with one line of code. (Hooray for backwards compatibility!) You can bring this functionality back to your website by adding the following snippet of code into your theme’s functions.php file:
You can do this manually in the administrative user interface of your WordPress install by going to Appearance » Editor, which will bring up your theme files. Then in the right hand sidebar there should be a link for editing your functions.php file. Cut and paste the line of code into the file on its own line and then click Update File.
That’s easy enough, but what do you do if you’re scared of code? (You shouldn’t be, by the way…) The same functionality can be brought back with the Link Manager plugin. Just download it in the admin UI under Plugins » Installed Plugins and click Add New at the top of the page. Search for the plugin name Link Manager to download and then activate. That’s it.
Note: some may worry at the fact that the details for this plugin include the warning words:
This plugin hasn’t been updated in over 2 years. It may no longer be maintained or supported and may have compatibility issues when used with more recent versions of WordPress.
On a scale of 1-10 for warnings, this one is really less than a 1. This has to be one of the simplest plugins in all of WordPress because it really only includes the single line of code above. There’s really almost nothing with it that could change, break, or need to be updated. It’s old, but it will work.
You’ve now re-enabled the Links Manager which will put a Links tab into your admin UI. You can click on it to start adding your links, feeds, photos, and data. The WordPress codex has great documentation for how to do this: https://codex.wordpress.org/Links_Manager
Within the admin UI you can now display a blogroll widget by going to Appearance » Widgets and moving the Links widget into one of the widgetizable areas in your theme.
Put your following list onto a page by itself
Sadly, because the Links Manager is so old and is now hidden, development on it seems to have long since stalled. This means you’ll require some simple code to get things working a bit better in terms of display. I’ll do my best to give you instructions for cutting and pasting with as little code as possible.
Plugin and Code
There’s a convenient plugin called Links Page which will get us most of the way. Go ahead and download and activate it. From the plugin interface, click the edit link for the Links Page.
The Links Page plugin displayed in the Plugin page of the WordPress admin UI.
The editor will pop up with the code for the plugin, which looks like this:
function linkspage($text)
if (preg_match("
add_filter('the_content', 'linkspage', 2);
In between the parenthesis for the function wp_list_bookmarks(), you’ll want to add something like the following code snippet I’ve customized for my following lists:
Yours doesn’t necessarily need to be exactly the same, but it should reflect how you’d like your own list to look. To accomplish this take a look at the documentation and examples for this function to pick and choose among the options you’d like to display. You’ll just string the options together between two single quotes and separate them with an ampersand (&) as in my example above.
Caveat: Since we’ve done some “cowboy coding” here and modified the code directly in the plugin, we run the risk of the accidentally updating the plugin and overwriting our changes. I would suggest that this risk is fairly low given the simplicity of the plugin and the unlikelihood that it would need an update. More advanced WordPress users will know that the better option is to roll up all the code in the plugin and all their changes and put it into their functions.phpfile or just fork the plugin with a new name and go from there.
CSS for styling and display
You may want to put in a bit of CSS to modify how our following list is displayed on the page. Without some tweaks or taking some extreme care when uploading or linking to the photos/avatars, we may run into some display issues.
As a result I’ve added the following snippet of CSS to my theme’s style.css file:
ul.xoxo.blogroll > li > a > img
width: 20px;
height: 20px;
ul.xoxo.blogroll li
list-style-type: none;
You can accomplish this by going to Appearance » Editor in the admin UI and editing the file by cutting and pasting the segment above into it and clicking Update file when you’re done.
The Page itself
We’ve now got all the big pieces in place. If you haven’t already, add some data into the Links Manager (documented here). You then need to create a new page on your site in the admin UI. I’ve named mine Following, but you can name yours Blogroll, Links, or anything you’d like really–it is your site after all.
Next, as described in the instructions for the Links Page plugin add the following text into the body of your post:
<!--links-page-->
When you’re done, save the page. The plugin will then replace the text above with your following list based on the output properties you specified.
Optionally you may want to go to Appearance » Menus to modify your menu to show your follow page in your menu structure so people can easily get to it.
Other Options
Those who’d like a different way of doing all of the above might also consider trying out other blogroll-related plugins in the WordPress repository. There are likely some other excellent options and methods to accomplish some of this functionality in a way that’s acceptable for your needs.
Future
So where do we go from here? This is certainly not complete by any means and there could be additional functionalities built on top of and even beside all of this.
I haven’t delved into it deeply, but I know there are developers like Dave Winer who have created services like Share Your OPML which allow you to upload your own file and then get recommendations of similar feeds in which you might also have some interest. Services like this that take advantage of my open data to provide me with value in return could be truly awesome.
I’m sure others smarter than I will come up with better UI. I’d personally love to have a bookmarklet similar to SubToMe that allows me to quickly and easily scrape a page and post the data from a person’s site to my following list (SubToMe currently redirects one to third party readers instead.)
The future might also bring increased ease-of-use as well as expanded functionality. I’m curious what value might be extracted by adding microfomats like h-cards to my follow lists? What could parsers do with a microformat like ‘p-following’ to more quickly create social graphs like Ryan Barrett’sIndie Map?
What might we expect with simpler formats than OPML, which could likely be done with microformat classes the same way that h-entry and h-feed have made supporting clunkier specs like RSS and Atom far easier?
I’m feeling itchy with all the potential possibilities…
Comments
I’d love to hear people’s thoughts and comments on the usefulness of any of the above. Is it something you’d attempt to do yourself? (If you attempted it, did it actually work?) What would you change? How could it be extended? What UI/UX improvements could be added? Other interactivity suggestions? How can the discoverability of such a thing be improved? What could be built on top of it all?
Also feel free to share your following pages, blogrolls, and OPML feeds in the comments below. Have you added your examples to the IndieWeb wiki to help others improve?
Are there people or sources missing from my following page that you’d recommend? (Keep in mind I’m far from done adding sources…)
Recently I discovered the IndieWeb project, and I… think I am a lot more intrigued by it than by other Better Social Media Platform pipe dreams and decentralization projects I’ve seen? Because it’s not a monolithic platform that has to be all things to all people, or even one that has to gain a critical mass of userbase before it’s useful for anything. It’s just a bunch of people, making sites that work for them, and banging out protocols so their sites can talk to each other and hook up to the social-media hangouts du jour.
The basic idea:
- Have a personal website, preferably a personal domain name, that is the hub for your online identity and stuff. Posts, tweets, pictures, links, reading list, events, whatever you’d normally be posting to social media. You host it, you control it, you own it. You tweak it to fit your needs, no Xkit required.
- Once the original archival copy is up on your personal site, cross-post it to whatever social media sites it belongs on. You don’t have to quit your Tumblr habit, or convince your friends to quit theirs, or give up the audience you can reach on a large site.
- Use a pingbacks-on-steroids tool to collect all the responses (likes, reblogs, comments, etc) from the various sites you’ve cross-posted to. Ideally, display them at the bottom of the post back on your website.
As an idea, I like it a lot. In practice, a lot depends on what tools are already available, how useable they are, how capable you are of coding/templating/configuring to fill in the gaps, and how difficult large sites make it to push/pull from them automatically. That’s pretty much what I’m interested in exploring in the near future, for my own use if nothing else. I already have most of my Tumblr content backed up to a Wordpress install on my own shared hosting account, so I’m kinda curious see how much IndieWeb compatibility I can manage using plugins and template tweaks.
Indieweb and fandom:
As a potential tool for fandom to wean ourselves off the various hellsites we’ve inhabited over the years… okay, it’s an interesting thought. One with lots of unanswered questions, but interesting.
Lots of unanswered questions, so the rest of this is going under a cut.