Skip to main content
 

Every company is a community

Silhouettes of people gathering at sunset

There’s a piece in the latest Harvard Business Review which starts with a premise I’d like to challenge:

It’s well-known that firms where strategy and culture align outperform firms where they do not. It follows, then, that if the two aren’t aligned, you most likely need to change your culture.

The rest of the article goes on to describe how storytelling is an integral part of establishing a strong culture in a company — and it absolutely is. A cohesive, supporting culture and the ability to tell strong stories are things every organization needs if it wants to succeed.

What I want to challenge is this idea that if strategy and culture don’t align, it’s the culture that needs changing. To be sure, quite often it does: particularly in situations where not enough time has been spent building a supportive culture to begin with. But culture is made of people, relationships, norms, and stories. The premise above hinges on the idea that if the people in your organization aren’t aligned with an organization’s strategy, you need to change the people. The strategy is paramount. But what if that’s not true? What if the strategy really is at fault, and organizations need to put more trust in their people?

I believe that the strongest organizational cultures are the most equitable ones. Whether you agree with my belief or not, research backs me up: the most productive work cultures are the ones where everyone feels empowered to speak up and be heard, where management genuinely listens to and acts on both the needs and ideas of their workforce. There’s a necessary underlying respect that you can’t simply storytell your way around.

In turn, that respect is built by distributing equity: giving people real ownership, both figurative and literal, in their workplace. There’s a reason burnout is driven by people not feeling like they can affect the choices that impact their work; people want to have control, and to make real progress on meaningful work.

If you don’t have those things, then, sure, your culture needs to be changed. But if you find yourself wishing that everyone would just go along with what you’re telling them to do, perhaps the first change needs to be a little more personal.

Building this level of interpersonal respect necessitates approaching building your workforce like a community. In turn, this means prioritizing strong interpersonal relationships. Can people talk with each other openly? Are they able to bring their whole selves to work? Does management listen and act? Are there rules and norms that foster emotional safety, particularly among people who may feel underrepresented and therefore alone in the organization? Does the organization treat the people who work inside it as fully-realized, three dimensional human beings, or are they fungible line items on a spreadsheet?

Is your company a diverse, happy, healthy community of trusted experts?

Was your strategy co-developed with your community?

Are your community’s needs and ideas represented in your decisions?

Will your community directly experience any upside that is an outcome of their work?

From here, other, related ideas become more obvious. If your community is co-developing your strategy, your want as many diverse ideas as possible. Hiring people from different contexts and backgrounds becomes an integral part of setting strategy. It becomes important to ask, when considering any new hire, whether they’ll bring a new perspective to your community. A homogenous workforce becomes a liability, because you’ll have a narrower set of ideas to work with.

Obviously, this mindset of collaborative inclusion is not commonly employed among business leaders. If it were, we’d see more diverse organizations with happy workforces, rather than the stark monocultures that engage in union-busting we see in tech (and everywhere) today. It runs completely counter to Elon Musk’s “extremely hardcore” work culture and wild firing rampages, for example. Not to mention the screaming fits associated with software leaders like Bill Gates and Steve Jobs.

Often, managers want to be mavericks: the smartest people in the room, who bend reality with their singular vision and bring everyone along from the ride. The truth is, it’s kind of bullshit: a story egotists tell themselves to justify being antisocial. You can’t hypnotize people into working for you through sheer charisma. The only way to scale an organization is to set a really strong internal culture first, and then empower everyone you add to your community to help you build it.

If you’ve hired great people and built a strong community, and those people are telling you that your strategy is off, you should believe them. And then you should shut up and listen to them, work together, and build something better together.

If you haven’t hired great people, and you haven’t built a great community, you’ve failed at business-building 101 and need to go back to the drawing board.

Every industry, including tech, comes down to people. Every company is a community. And every community is built on trust, respect, and equity.

· Posts · Share this post

 

Unions work

Hands joining together in solidarity

The Writer’s Guild of America seems to have received everything it asked for:

Delivering on issues that many scribes saw as core to their profession, the deal contains big leaps in AI guardrails, residuals and data transparency for writers — leaps that could be transplanted into the upcoming negotiations between the AMPTP and SAG-AFTRA, which could start in the next week.

This is a great example of how unions can really work for their members. Hollywood is a broadly-unionized industry, as we’ve seen for the last five months, and the result has been real gains in writer equity and compensation in the face of technology changes like streaming and AI.

Of course, at least in America, most industries are not highly-unionized. 10.1% of wage and salary workers were unionized in 2022, down from a peak of about a third, which coincided with income inequality’s lowest point. Generally, unionized workers make 18% more (20% for African Americans, 23% for women, and 34% for Hispanic workers).

Tech is often the home of a particular kind of libertarian thinking that is often anti-union. But that, thankfully, is changing. In 2004, a third of tech workers were in favor of unionization; twelve years later, it was 59%. These days, prominently recognized tech unions include the Alphabet Workers Union, but firms have engaged in nakedly union-busting activity, from big tech companies like Apple, later-stage startups like Instacart, and supposedly public interest organizations like Code for America. (It’ll be no surprise that Elon Musk’s Tesla was found guilty of illegal union-busting tactics).

Regardless, the industry would gain immensely from unionization — and more and more tech workers agree. It’s not so much about wages as recognition and a say in how these companies are run. Last year, Jane Lytvynenko, senior research fellow on the Tech and Social Change Project at the Shorenstein Center wrote in MIT Technology Review:

[…] Silicon Valley companies don’t see more protests about wages from their white-collar employees—those workers get stock options, good salaries, and free lunch. But such perks do little to address structural discrimination.

My hope is that examples like the WGA’s win will help spread this idea that there should be a counterbalance to corporate power, and that the people who do the work should have influence over how it is organized. If you’ll pardon the pun, tech workers should own the means to push to production. Allworkers should have a say in how their companies function. And I believe — still crossing my fingers, because there’s a lot of work to do and a lot of gains still to be made — that this future is coming.

In the meantime, congratulations to the WGA! Nice work. Solidarity.

· Posts · Share this post

 

Parenting in the age of the internet

A toddler using an iPhone on the floor

I learned to read and write on computers.

Our first home computer, the Sinclair ZX81, had BASIC shortcuts built into the keyboards: you could hit a key combination and words like RUN, THEN, and ELSE would spit out onto the screen. I wrote a lot of early stories using those building blocks.

Our second, the Atari 130XE, had similar BASIC instructions, but also had a much stronger software ecosystem. In one, you would type a rudimentary story, and 8-bit stick figure characters would act it out on screen. “The man walks to the woman”; “The wumpus eats the man.”

We never had a games console in the house, much to my chagrin, although the Atari could take games cartridges, and I once got so far in Joust that the score wrapped back around to 0. But mostly, I used our computers to write stories and play around a little bit with simple computer programming (my mother taught me a little BASIC when I was five).

We walk our son to daycare via the local elementary school. This morning, as we wheeled his empty stroller back past the building, a school bus pulled up outside and a stream of eight-year-olds came tumbling out in front of us. As we stood there and watched them walk one by one into the building, I saw iPhone after iPhone after iPhone clutched in chubby little hands. Instagram; YouTube; texting.

It’s obvious that he’ll get into computers early: he’s the son of someone who learned to write code at the same time as writing English and a cognitive scientist who does research for a big FAANG company. Give him half a chance and he’ll already grab someone’s phone or laptop and find modes none of us knew existed — and he’s barely a year old. The only question is how he’ll get into computers.

I’m adamant with him, as my parents were with me, that he should see a computer as a creation, not a consumption device. At their best, computers are tools that allow children to create things themselves, and learn about the world in the process. At their worst, they’re little more than televisions, albeit with a near-infinite number of channels, that needlessly limit your horizons. For many kids, social media is such a huge part of of their life that being an influencer is their most hoped-for job. No thank you: not for my kid.

But, of course, if we can steer away from streaming media and Instagram’s hollow expectations, there’s a ton of fun to be had. This is one area where I think generative AI could be genuinely joyful: the fun that I had writing stories for those 8-bit stick figures, transposed to a whole universe of visual possibilities. That is, of course, unless using those tools prevents him from learning to draw himself.

He’s entering a very different cultural landscape where computers occupy a very different space. Those early 8-bit machines were, by necessity, all about creation: you often had to type in a BASIC script before you could use any software at all. In contrast, today’s devices are optimized to keep you consuming, and to capture your engagement at all costs. Those iPhones those kids were holding are designed to be addiction machines.

Correspondingly, our role as parents is to teach responsible use. If we are to be good teachers, that also means we have to demonstrate responsible use: something I am notoriously bad at with my own phone. I’ve got every social network installed. I sometimes lose time to TikTok. I’m a slave to my tiny hand-computer in every way I possibly can be. I tell myself that I need to know how it all works because of what I do for a living, but the real truth is, I love it. I don’t need to be on social media; I don’t need to be a part of the iPhone Upgrade Program. I just am.

I think responsible use means dialing up the ratio of creation to consumption for me, too. If I’m to convey that it’s better to be an active part of shaping the world than just being a passive consumer of it, that’s what I have to do. This is true in all things — a core, important lesson is that there isn’t one way to do things, and life is richer if you don’t follow the life templates that are set out for us — but in some ways I feel it most acutely in our relationship to technology.

There will certainly be peer pressure. His friends will have iPhones. I don’t think withholding technology is the right thing to do: consider those kids whose parents never let them have junk food, who then go out and have as much junk food as possible as soon as they can. Instead, if he has an iPhone, he will learn how to make simple iPhone apps. You’d better believe that he’ll learn how to make websites early on (what kind of indieweb advocate would I be otherwise?). He will be writing stories and editing videos and making music. And, sure, he’ll be consuming as part of that — but, in part, as a way to get inspired about making his own things.

These days, creating also means participating in online conversations. As he gets older, we’ll need to have careful discussions about the ideas he encounters. I’m already imagining that first conversation about why Black Lives Matter is an important movement and how to think about right-wing content that seeks to minimize other people. I don’t want our kid to be a lurker who thinks people should be happy with what they get; I want him to feel like the world is his oyster, and that he can help change it for the better. Our devices can be a gateway to bigger ideas, or they can be a path to a constrained walled garden of parochial thought. It all requires guidance and trust.

The computer revolution happened between my birth and his. Realizing so makes me feel as old as dust, but more importantly, it opens up a new set of parental responsibilities. I want to help him be someone who creates and affects the world, not someone who lets the world happen to him. And there’s so much world to see.

· Posts · Share this post

 

Trying out Kagi

As an experiment, I’m trying out Kagi as my default search engine, switching from Google (no link required; they’re probably behind you right now).

I like the idea of an ad-free experience: a paid-for engine puts the incentives in the right place. But it’s got to be about more than ideology. Because search is such an important part of my working life (and most knowledge workers’ working lives) it’s important that the results are actually better than Google’s.

For a while, I tried to use DuckDuckGo, which uses Bing’s search engine behind the scenes. It was just fine for most things and markedly worse for a few, so I had to switch back, even though I love its privacy focus.

Kagi uses a mix of third-party engines and its own to provide its results. So far, they seem pretty good, but the proof will be in intensive use.

What I already know I love: they have a StumbleUpon-like site for discovering small websites, and surface blogs in some search results when they’re relevant. That’s something I want from every search engine.

We will see! I’ll give it a month and then report back.

· Posts · Share this post

 

Subscribing to the blogs of people I follow on Mastodon

It’s no surprise to anyone that I prefer reading peoples’ long-form thoughts to tweets or pithy social media posts. Microblogging is interesting for quick, in-the-now status updates, but I find myself craving more nuance and depth.

Luckily, Blogging is enjoying a resurgence off the back of movements like the Indieweb (at one end of the spectrum) and platforms like Substack (at the other), and far more people are writing in public on their own sites than they were ten years ago. Hooray! This is great for me, but how do I find all those sites to read?

I figured that the people I’m connected to on Mastodon would probably be the most likely to be writing on their own sites, so I wondered if it was possible to subscribe to all the blogs of the people I followed.

I had a few criteria:

  1. I only wanted to subscribe to blogs. (No feeds of updates from GitHub, for example, or posts in forums.)
  2. I didn’t want to have to authenticate with the Mastodon API to get this done. This felt like a job for a scraper — and Mastodon’s API is designed in such a way that you need to make several API calls to figure out each user’s profile links, which I didn’t want to do.
  3. I wanted to write it in an hour or two on Sunday morning. This wasn’t going to be a sophisticated project. I was going to take my son to the children’s museum in the afternoon, which was a far more important task.

On Mastodon, people can list a small number of external links as part of their profile, with any label they choose. Some people are kind enough to use the label blog, which is fairly determinative, but lots don’t. So I decided that I wanted to take a look at every link people I follow on Mastodon added to their profiles, figure out if it’s a blog I can subscribe to or not, and then add the reasonably-bloggy sites to an OPML file that I could then add to an RSS reader.

Here’s the very quick-and-dirty command line tool I wrote yesterday.

Mastodon helpfully produces a CSV file that lists all the accounts you follow. I decided to use that as an index rather than crawling my instance.

Then it converts those account usernames to URLs and downloads the HTML for each profile. While Mastodon has latterly started using JavaScript to render its UI — which means the actual profile links aren’t there in the HTML to parse — it turns out that it includes profile links as rel=“me” metatags in the page header, so my script finds end extracts those using the indieweb link-rel parser to create the list of websites to crawl.

Once it has the list of websites, it excludes any that don’t look like they’re probably blogs, using some imperfect-but-probably-good-enough heuristics that include:

  1. Known silo URLs (Facebook, Soundcloud, etc) are excluded.
  2. If the URL contains /article, /product, and so on, it’s probably a link to an individual page rather than a blog.
  3. Long links are probably articles or resources, not blogs.
  4. Pages with long URL query strings are probably search results, not blogs.
  5. Links to other Mastodon profiles (or Pixelfed, Firefish, and so on) disappear.

The script goes through the remaining list and attempts to find the feed for each page. If it doesn’t find a feed I can subscribe to, it just moves on. Any feeds that look like feeds of comments are discarded. Then, because the first feed listed is usually the best one, the script chooses the first remaining feed in the list for the page.

Once it’s gone through every website, it spits out a CSV and an OPML file.

After a few runs, I pushed the OPML file into Newsblur, my feed reader of choice. It was able to subscribe to a little over a thousand new feeds. Given that I’d written the script in a little over an hour and that it was using some questionable tactics, I wasn’t sure how high-quality the sites would be, so I organized them all into a new “Mastodon follows” folder that I could unsubscribe to quickly if I needed to.

But actually, it was pretty great! A few erroneous feeds did make it through: a few regional newspapers (I follow a lot of journalists), some updates to self-hosted Git repositories, and some Lemmy feeds. I learned quickly that I don’t care for most Tumblr content — which is usually reposted images — and I found myself wishing I’d excluded it. Finally, I removed some non-English feeds that I simply couldn’t read (although I wish my feed reader had an auto-translate function so that I could).

The upshot is that I’ve got a lot more blogs to read from people I’ve already expressed interest in. Is the script anything close to perfect? Absolutely not. It it shippable? Not really. But it did what I needed it to, and I’m perfectly happy.

· Posts · Share this post

 

Long-term blogging

Tracy Durnell celebrates 20 years of blogging:

A blog is a much nicer place to publish than social media, sparking fewer but more meaningful interactions. Blogging allows writers a more forgiving pace with slower conversation. On their blog, people can be themselves instead of playing to an audience and feeling judged — a place to escape the pressures of one-upmanship and signaling, the noise of the ever-demanding attention economy, and the stress of hustle culture.

It’s a huge achievement, to be sure, and I couldn’t agree more with Tracy’s sentiment here. Congratulations, Tracy!

I’m a little jealous that she can pinpoint an anniversary date. For me, it depends on how you judge: I had a hand-rolled blog of sorts when I went to university in 1998, but was it really a blog? I definitely had a public Livejournal in 2001, but was that a blog? How about blog I used to keep on Elgg dot net (now a domain squatter, may it rest in peace)? My old domain, benwerd.com, dates back to 2006, and my current one, werd.io, only goes back to 2013. It’s a bit of a messy history, with stops and false starts.

On the other hand, I know people who have posted to the same domain for almost as long as they’ve been online. I don’t know if I can match that sort of dedication - or a commitment to even having a continuous identity for all that time. Am I the same person I was 20+ years ago? A little bit yes, but mostly not really. The idea of joining up my life online on a long-term basis is actually quite daunting.

Tracy links to Mandy Brown’s piece on writers vs talkers, which also deeply resonates: I’m a writer. I hate being drawn into making decisions in ad hoc meetings. I want to write my thoughts down, structure them, and then come to a conclusion after getting feedback and iterating. Perhaps that’s why blogging early appeals to me so much: I can put out ideas and very quickly engage in conversations about them that pushes my thinking along.

Blogging might seem like a solitary activity, but it’s very, very social. Even the name — a pun derived from weblog = we blog — is about community. Writing for 20 years also means building community for that long.

Here’s to the next 20!

· Posts · Share this post

 

The worst part of writing is writing

I’ve been neck-deep in a long-form first draft for months; at this point I’m many tens of thousands of words in. Every time I look back at my writing from tens of thousands of words ago, it’s a horrible mistake that opens up floodgates of self-questioning. How could I possibly have thought that I could do this? Who on earth would want to read this? Amateur! Go back to whatever it is you do for your day job. (Do you even know? I thought you wrote software? When was the last time you actually wrote software, you hack?)

But I’m determined. The only thing I can say for sure is that, eventually, I will have a manuscript. I have professional mentors who will read and critique it once I’ve iterated on it a few times. Beyond that, I can’t say. Perhaps, if I’m lucky, someone will like it. But perhaps it really is doomed to sit on my hard drive, unloved.

The deeper I get into it, the more I’m comfortable with the idea of failure. I think I started with the idea that I might be intentionally writing something that a lot of people might enjoy, but at this point it’s for me. The more I pour in of myself, and the ideas I have about the world (and the future of technology, because that’s the kind of book this is), the more I feel comfortable with it. Even if nobody loves it, it’ll be representative of me: a genuine work of self-expression hooked onto a plot that I continue to think is really interesting. And the feedback I get will help me learn to write the next one.

It turns out that the thing which most motivates me to write is my sense of humor. If it’s too self-serious, I stall. (Honestly, I expect readers would, too.) On the other hand, if I’m amusing myself, undercutting my serious points with irony or adding notes about things from the real world that I think are ridiculous, I can go forever. That’s probably something worth knowing about myself: I thrive on irreverence. I cut my teeth on Douglas Adams, Terry Pratchett, and Charlie Brooker’s early stuff, so that’s probably not surprising. I could probably use more of that here, too.

Anyway. It’s like pulling teeth, but joyously. A gleeful festival of unpleasant monotony wherein I make myself laugh while disgusting myself with my own ineptitude. And maybe, if I’m really, really lucky, something will even come of it.

· Posts · Share this post

 

AI is not a paradigm shift. But it could be useful

A light painting of the word

It’s been interesting to watch all of the articles celebrating the death of NFTs lately. For years, they were the harbinger of the next big thing, hawked by A-list celebrities. Behind the scenes, some of the biggest tech companies in the world spawned NFT strategies, even as critics noted that valuations were partially driven by money laundering and wash trading.

Cut to 2023, and surprise, surprise: 95% of NFTs are now completely worthless.

If you missed the craze: while most digital data can be infinitely replicated for almost no cost, Non Fungible Tokens, or NFTs, were a way to ensure there was only one of an item using blockchains. NFTs were often attached to digital art — for example, these hideous apes — and because they were both scarce and tradable, for a while each one was going for the equivalent of thousands of dollars. Of course, it couldn’t last, and NFTs turned out to be the digital equivalent of investing in Beanie Babies or tulips(pick your proverbial market collapse).

It’s now controversial to say that crypto isn’t completely useless, but if you look beyond the brazen grift, international crimes, and planet-destroying environmental impact, I do think there are a few things to celebrate about the trend. The crypto community deployed the most widely-used ever implementation of identity in the browser, for one: people who installed software like Metamask could choose to identify themselves to a website with a single click. In some countries, digital currencies also gave citizens an accessible safe haven when their own local currency tumbled. And finally, it introduced a much wider audience to the concept of decentralization, where a large-scale internet system is run co-operatively by all of its users instead of a giant megacorp.

Although the rampant speculation and wildly inflated prices are gone, there are some technical outcomes that will likely be with us for some time. And some of those are positive and useful.

This is exactly how the hype cycle works. A technology breakthrough kicks things off and gets people all excited. The market works itself into a frenzy over the technology, and lots of people imagine that it can do all kinds of amazing things. Those inevitably don’t actually pan out, and people lose hope and interest. But it turns out that the technology is useful for something, and eventually, it finds a mainstream use.

The Gartner Hype Cycle

Crypto is very much in the trough of disillusionment right now; eventually some aspects of the technology (maybe identity in the browser, maybe something else) will find a use.

Meanwhile, AI? AI is right at the top of that hype curve.

There are people out there who believe we’re building a new kind of higher consciousness, and that our goal as humans should be to support and spread that consciousness to the stars. A galaxy full of stochastic parrots is an inherently funny, Douglas Adams-esque idea, but naturally, they’re serious, partially because they feel this idea absolves them of dealing with the truth that there are actual human beings living on a dying planet who need help and assistance right now. In erasing the needs of vulnerable communities, AI supremacy (officially called effective accelerationism) is the new white supremacy (sitting comfortably alongside the old white supremacy, which is still going strong).

There are also people who think AI will replace poets, artists, neurosurgeons, and political leaders. AI systems will farm for us, tend to our children, and imagine whole new societies that we wouldn’t otherwise be capable of envisioning. They will write great literature and invent wholly new, never-ending dramatic entertainment for us to sit and consume.

It’s horseshit. The technology can’t do any of those things well. It’s best thought of us a really advanced version of auto-complete, and everyone who claims it’s something more is trying to sell you something.

Which isn’t to say it’s not useful. I’ve certainly used it as a utility in my writing — not to do the writing itself (it produces mediocre porridge-writing), but to prompt for different angles or approaches. I’ve used it to suggest ways to code a function. And I’ve certainly used it, again and again, as a quick way to autocomplete a line of code or an English sentence.

What’s going to happen is this: in a few years, AI will come crashing down as everyone realizes it’s not going to be an evolution of human consciousness, and some other new technology will take its place. Valuations of AI companies will fall and some will go out of business. Then, some of the actual uses of the technology will become apparent and it’ll be a mainstream, but not dominant, part of the technology landscape.

The hype cycle is well-understood. What surprises me, again and again, is how thoroughly people follow it. Across industries, CEOs are right now thinking, “holy shit, if we don’t jump on AI, we’re going to be completely left behind. This is a paradigm shift.” It’s kind of the equivalent of a bunch of soccer players chasing the ball — It’s over here! No, it’s over here! Let’s run towards it! — which is how three-year-olds play soccer. A more strategic approach (let’s call it thinking for yourself) will be more productive for most businesses.

There will absolutely be uses for AI tools. The important thing is to take a step back and think: what are my needs? What are the needs of my customers or my community? Given the actual demonstrated capabilities of the software, does it help me meet any of them in a reliable way? If I do use it, am I holding true to my values and keeping my customers and community safe? If the answer is yes to all of these things, then great! Otherwise it might be worth taking a step back and letting the dust settle.

Keep me honest: if AI doesn’t enter a trough of disillusionment and just keeps growing and growing exponentially, call me on it. But I think it’s a pretty safe bet that it won’t.

· Posts · Share this post

 

I'm going to keep using Zapier for my link blog

The way my link blog works is like this:

I save an article, website, or book I thought was interesting to a database in Notion using the web clipper, together with a description and a high-level category. (These are Technology, Society, Democracy, and so on.) I also have a checkbox that designates whether the link is something I’d consider business-friendly.

Zapier watches for new links. When it finds one, it publishes it to my website using the micropub protocol. (My website then tries to send a webmention to that site to let it know I’ve linked to them.)

Then, it publishes the link to my Mastodon profile using the top-level category as a hashtag. If the link is to a book, it also adds the bookstodon hashtag.

Following that, it publishes to all my other social networks via Buffer, without the hashtag. (The exception is my Bluesky profile, which I had to write some custom API code for). If the business-friendly box was checked, that includes publishing to my LinkedIn profile.

If I’m feeling particularly motivated, I’ll copy and paste the link to my Threads profile, but because there’s no API, it’s a fully manual process. Which means I usually don’t.

Very occasionally, Zapier will pick up a link before the Notion entry has fully saved, which means that links post without a description or a category. Then I either shrug my shoulders and accept that I have some weird posts on my timeline, or I go back and edit or repost each and every one.

Because of this bug, I’ve thought about writing my own code to do all of the above on my server. It would work the exact way I want it to be. It would be cheaper, too: I pay for Zapier every month, and the cost adds up.

But while I could do this, and the up-front cost would certainly be lower, what if something goes wrong? Let’s say LinkedIn changes the way their API works. If I wrote the connection myself, I would need to keep my code up to date every time this happened — and, in turn, stay on top of codebase changes for every single social media platform I used.

And the truth is: I’m tired, friends. I want to be really careful about the amount of code I set myself up to maintain. It might seem like a simple script now, but over time I build up more and more simple scripts and, cumulatively, I end up buried in code.

As I get older, I find myself optimizing that cost more and more. I’d much rather pay something up-front that saves me a ton of time and cognitive overhead, because both of these things are at such an enormous premium for me.

I could also just not post to those social media accounts, or do it fully-manually, but there’s something really satisfying about publishing once and syndicating everywhere I’m connected to people. I could save my links straight to something like Buffer, but I also like having my categorized database of everything I’ve shared. And Notion makes it easy to save links across my devices (I’m sometimes on my phone, sometimes on my laptop, sometimes on my desktop).

So I’m keeping Zapier, at least for now. I like keeping my links, and I like sharing them. And, more than anything else, I like not having to maintain the code that does it.

· Posts · Share this post

 

An update on Sup, the ActivityPub API

An abstract network

A little while back I shared an idea about an API service that would make it easy to build on top of the fediverse. People went wild about it on Mastodon and Bluesky, and I got lots of positive feedback.

My startup experience tells me that it’s important to validate your idea and understand your customers before you start building a product, lest you spend months or years building the wrong thing. So that’s exactly what I did.

I put out a simple survey that was really just an opener to find people who would be interested in having a conversation with me about it. I bought each person who replied a book certificate (except for one participant who refused it), and listened to why they had been interested enough to answer my questions. If they asked, I told them a little more about my idea.

The people I spoke with ran the gamut from the CEOs of well-funded tech companies to individuals building something in the context of cash-strapped non-profits. I also spoke with a handful of venture capitalists at various firms who had proactively reached out.

A shout-out to Evan Prodromou, one of the fathers of the fediverse, here: he very kindly spent a bunch of time with me keeping me honest and helping to move the project along.

What I discovered was that the people who wanted me to build my full idea were people who really cared about the fediverse, but were not going to be customers. The people who were going to be customers wanted two specific things:

A fast way to make informational bots. Twitter used to be full of informational, automated accounts. Consider accounts containing local weather updates, earthquake reports, and so on. That’s been much harder for people to build on the fediverse.

Statistics about trends and usage. Aggregate information about how the fediverse is behaving, including about how accounts are responding to individual links and domains.

While these signals were very clear, I couldn’t yet validate the core thing I’d proposed to build, which was a full API service with libraries that let people build fully-featured fediverse-compatible software. I also couldn’t yet validate the idea that existing startups would use a service like this to add fediverse compatibility to their products.

But I believe, to reference a way-overused cliché, that this is where the puck is going.

I strongly believe that the fediverse is how new social networks over the next decade will be built. I also have conviction that more people will be interested in building fully-featured fediverse services once Threads federates and Tumblr joins. It’s likely that another large network will also start supporting these protocols.

However, someone financially backing the project would be doing so on the basis of my conviction alone. I couldn’t yet find strong customers for this use case.

I think that’s okay! In the shorter term, I’m very interested in helping people build those bots in particular — it’s a great place to start and a good example of building the smallest, simplest, thing.

The original name I came up with, Sup, was taken by another fediverse project. So for now, this idea is called Feddy.

Anyway, I wanted to report back on what I’d found and how I was thinking about the project today. As always, I’d love your feedback and ideas! You can always email me at ben@werd.io.

· Posts · Share this post

 

Bush's legacy

The contemporary New York City skyline

Twenty-two years ago, I sat in the office — actually the bottom two floors of a Victorian home with creaking, carpeted floorboards and an overstuffed kitchen — at Daily Information, the local paper where I worked in Oxford. It was mid-afternoon, and I probably had Dreamweaver open; I can’t remember exactly now. I’d taken a year’s break from my computer science degree because my as-yet-undiagnosed anxiety had gotten the better of me in the wake of the death of a close friend. It was the first job I’d ever had that paid for lunch, and the remains of a wholewheat bread slice with spicy red bean paté sat on a plate beside me. Between that and the array of laser printers, the room smelled of toast and ozone.

My dad showed up and told me what had happened: the twin air strikes of September 11, 2001, the details of which are now part of our indelible cultural consciousness. For the rest of the afternoon, we tried to learn what we could, refreshing website after website on the overloaded ISBN connection. One by one, every news website went down for us under the strain of unprecedented traffic, with the exception of The Guardian. I alternated between that and a fast-moving MetaFilter thread until it was time to go home. I vividly remember sitting at the bus stop, watching the faces of all the people in the cars that drove past, thinking that the world would likely change in ways that we didn’t understand yet.

George W Bush was President of the United States: a man who previously had presided over more executions than any other Governor of the State of Texas in history (roughly one every two weeks). While the attacks themselves were obviously an atrocity, he was, in my eyes, unmistakably an evil, untrustworthy leader, and it wasn’t clear that he wouldn’t start a terrible war in response. That was the fear expressed by most of my friends in England at the time: not who was behind the attacks and why?, but what will America do? I was the only American in my friend group, but I shared the same fear.

Of course, now we all know the story of the next two decades. We invaded Iraq under false pretenses, established a major erosion of civil liberties ironically called the PATRIOT Act which granted unprecedented authorities that live on to this day, and racist anti-Muslim rhetoric cranked up to eleven. All in the name of 2,753 people who didn’t ask for any of it. Even the first responders, much lauded at the time, struggle to get the support they need.

In 2002, my parents moved back to California to look after my Oma, and I joined them for a few months. I had the whole row on my transatlantic flight to myself, which seemed strange until I remembered, mid-flight, that it was September 11, 2002 (in retrospect probably the safest day to fly in history). When I arrived, I saw that the freeways were littered with tiny American flags that had fallen off the cars they had presumably been waving from over the last year. As a metaphor, discarded disposable American flags bought to illustrate a kind of temporary superficial patriotism seemed a little on the nose.

While the roads were littered with flags, the air was still thick with fear. My parents had moved to Turlock, a small town outside of Modesto where the radio stations mostly played country music and almond dust polluted the air. There was still a feeling that the next attack could happen at any time, and if it did, why wouldn’t it be here? The dissonance between the significance of the World Trade Center in New York City and the Save-Mart in Turlock seemed to be lost on them. It could happen anywhere. It was the perfect environment for manufacturing consent for war. What did it matter that Saddam Hussein had precisely nothing to do with the attacks and that the purported weapons of mass destruction were obviously fictional? He was brown too, wasn’t he? And, boy, we needed to get revenge.

Even now, I wonder if I should be writing these opinions. In a way, September 11 has become a sacred event. And, seriously, what gives me the right to be talking about it to begin with?

But the tragedy of that day has touched all of us, everywhere. It has also been used as a cover for harms that continue to this day. The deaths of those innocent people are still used to justify erosions of civil liberties; they are still used to justify racism; they are still used to justify mass surveillance domestically and drone strikes internationally; they are still used to justify draconian foreign policies. If any lessons at all were learned from September 11, I think they were the wrong ones.

There’s an alternate universe where America as a population decided that funding and arming covert operations in foreign nations to support American aims was a bad idea. The late Robin Cook, MP, the former British Foreign Secretary, wrote in the wake of the July 7 bombings in London:

‌In the absence of anyone else owning up to yesterday's crimes, we will be subjected to a spate of articles analysing the threat of militant Islam. Ironically they will fall in the same week that we recall the tenth anniversary of the massacre at Srebrenica, when the powerful nations of Europe failed to protect 8,000 Muslims from being annihilated in the worst terrorist act in Europe of the past generation.

[…] Bin Laden was, though, a product of a monumental miscalculation by western security agencies. Throughout the 80s he was armed by the CIA and funded by the Saudis to wage jihad against the Russian occupation of Afghanistan. Al-Qaida, literally "the database", was originally the computer file of the thousands of mujahideen who were recruited and trained with help from the CIA to defeat the Russians. Inexplicably, and with disastrous consequences, it never appears to have occurred to Washington that once Russia was out of the way, Bin Laden's organisation would turn its attention to the west.

The CIA, for the record, denies this. But there’s no denying the effect of American foreign policies overall, from Chile (whose US-aided coup was 50 years ago today) to Iran, let alone the disastrous wars in Iraq and Afghanistan. It’s still a mystery to some Americans why the rest of the world isn’t particularly fond of us, but it really shouldn’t be. (And it’s not, as some particularly tone deaf commentators have suggested, jealousy.)

I remember visiting Ground Zero for the first time. By that time, reconstruction was underway, but the holes were clearly visible: conspicuous voids shot through a bustling, diverse city. I think New York City is one of the most amazing places I’ve ever been to: all kinds of people living on top of each other in relative harmony. It’s alive in a way that many places aren’t. Every time I visit I feel enriched by the humanity around me. One of the reasons I live where I do now is to be closer to it.

I think New York City itself is a demonstration of the lesson we should have learned: one that’s more about cross-border co-operation and humanity than isolation and dominance. To put it another way, a lesson that’s more about love than fear. Some conservative politicians talk derisively about “New York values”, but man — if those values were actually shared by the whole nation, America would be a far better place. That was obvious in the way the city came together that day, and it’s been obvious in the way it’s held itself together since.

In contrast, I think the way America as a whole responded to the September 11 attacks directly paved the way to Trump. It enriched a right-wing populist leader and his party; it created divisive foreign policy based on a supremacist foundation; it once again marked people with a certain skin tone and a different religion as being second-class citizens; it promoted nationalism and exceptionalism; it eroded hard-won freedoms for everyone. We can thank Bush for stoking those fires.

True progress towards peace looks like a collaborative world where we consider ourselves to have kinship with everyone of all religions, skin tones, and nationalities, and where every human being’s life has inherent value. It looks like building foreign policy for the benefit of all people, not the people of one nation. It looks like true, vibrant democracy. It doesn’t look like performative flag-waving, drone strikes, religious intolerance, homogeneity, or surveillance campaigns.

Saying so shouldn’t dishonor the memories of everyone who died on that day, or everyone who died as a result of everything that followed. It also doesn’t besmirch our values. One of the greatest things about America is our freedom to hold it to account. That’s what democracy and free expression are all about. And those values — collaboration, inclusion, freedom, representation, multiculturalism, democracy, and most of all, peace — are what we should be working towards.

· Posts · Share this post

 

An AI capitalism primer

A clenched robot fist

Claire Anderson (hi Claire!) asked me to break down the economics of AI. How is it going to make money, and for whom?

In this post I’m not going to talk too much about how the technology works, and the claims of its vendors vs the actual limitations of the products. Baljur Bjarnason has written extensively about that, while Simon Willison writes about building tools with AI and I recommend both of their posts.

The important thing is that when we talk about AI today, we are mostly talking about generative AI. These are products that are capable of generating content: this could be text (for example, ChatGPT), images (eg Midjourney), music, video, and so on.

Usually they do so in response to a simple text prompt. For example, in response to the prompt ‌Write a short limerick about Ben Werdmuller asking ChatGPT to write a short limerick about Ben Werdmuller, ChatGPT instantly produced:

Ben Werdmuller pondered with glee,
“What would ChatGPT write about me?”
So he posed the request,
In a jest quite obsessed,
And chuckled at layers, level three!

Honestly, it’s pretty clever.

While a limerick isn’t particularly economically useful, you can ask these technologies to write code for you, find hidden patterns in data, highlight potential mistakes in boilerplate legal documents, and so on. (I’m personally aware of companies using it to do each of these things.)

Each of these AI products is powered by a large foundation model: deep learning neural networks that are trained on vast amounts of data. In essence, the neural network is a piece of software that ingests a huge amount of source material and finds patterns in it. Based on those patterns and the sheer amount of data involved, it can statistically decide what the outcome of a prompt should be. Each word of the limerick above is what the model has decided is the most probably likely next piece of the output in response to my prompt.

The models are what have been called stochastic parrots: their output is entirely probabilistic. This kind of AI isn’t intelligence and these models have no understanding of what they’re saying. It’s a bit like a magic trick that’s really only possible because of the sheer amount of data that’s wrapped up in the training set.

And here’s the rub: the training set is a not insignificant percentage of everything that’s ever been published by a human. A huge portion of the web is there; it’s also been shown that entire libraries of pirated books have been involved. No royalties or license agreements have been paid for this content. The vast majority of it seems to have been simply scraped. Scraping publicly accessible content is not illegal(and nor should it be); incorporating pirated books and licensed media clearly is.

Clearly if you’re sucking up everything people have published, you’re also sucking up the prejudices and systemic biases that are a part of modern life. Some vendors, like OpenAI, claim to be trying to reduce those biases in their training sets. Others, like Elon Musk’s X.AI, claim that reducing those biases is tantamount to training your model to lie. He claims to be building an “anti-woke” model in response to OpenAI’s “politically correct” bias mitigation, which is pretty on-brand for Musk.

In other words, vendors are competing on the quality, characteristics, and sometimes ideological slant of their models. They’re often closed-source, giving the vendor control over how the model is generated, tweaked, and used.

These models all require a lot of computing power both to be trained and to produce their output. It’s difficult to provide a service that offers generative AI to large numbers of people due to this need: it’s expensive and it draws a lot of power (and correspondingly has a large environmental footprint).

The San Francisco skyline, bathed in murky red light.

Between the closed nature of the models, and the computing power required to run them, it’s not easy to get started in AI without paying an existing vendor. If a tech company wants to add AI to a product, or if a new startup wants to offer an AI-powered product, it’s much more cost effective to piggyback on another vendor’s existing model than to develop or host one of their own. Even Microsoft decided to invest billions of dollars into OpenAI and build a tight partnership with the company rather than build its own capability.

The models learn from their users, so as more people have conversations with ChatGPT, for example, the model gets better and better. These are commonly called network effects: the more people that use the products, the better they get. The result is that they have even more of a moat between themselves and any competitors over time. This is also true if a product just uses a model behind the scenes. So if OpenAI’s technology is built into Microsoft Office — and it is! — its models get better every time someone uses them while they write a document or edit a spreadsheet. Each of those uses sends data straight back to OpenAI’s servers and is paid for through Microsoft’s partnership.

What’s been created is an odd situation where the models are trained on content we’ve all published, and improved with our questions and new content, and then it’s all wrapped up to us as a product and sold back to us. There’s certainly some proprietary invention and value in the training methodology and APIs that make it all work, but the underlying data being learned from belongs to us, not them. It wouldn’t work — at all — without our labor.

There’s a second valuable data source in the queries and information we send to the model. Vendors can learn what we want and need, and deep data about our businesses and personal lives, through what we share with AI models. It’s all information that can be used by third parties to sell to us more effectively.

Google’s version of generative AI allows it to answer direct questions from its search engine without pointing you to any external web pages in the process. Whereas we used to permit Google to scrape and index our published work because it would provide us with new audiences, it now continues to scrape our work in order to provide a generated answer to user queries. Websites are still presented underneath, but it’s expected that most users won’t click through. Why would you, when you already have your answer? This is the same dynamic as OpenAI’s ChatGPT: answers are provided without credit or access to the underlying sources.

Some independent publishers are fighting back by de-listing their content from Google entirely. As the blogger and storyteller Tracy Darnell wrote:

I didn’t sign up for Google to own the whole Internet. This isn’t a reasonable thing to put in a privacy policy, nor is it a reasonable thing for a company to do. I am not ok with this.

CodePen co-founder Chris Coyier was blunt:

Google is a portal to the web. Google is an amazing tool for finding relevant websites to go to. That was useful when it was made, and it’s nothing but grown in usefulness. Google should be encouraging and fighting for the open web. But now they’re like, actually we’re just going to suck up your website, put it in a blender with all other websites, and spit out word smoothies for people instead of sending them to your website. Instead.

For small publishers, the model is intolerably extractive. Technical writer Tom Johnson remarked:

With AI, where’s the reward for content creation? What will motivate individual content creators if they no longer are read, but rather feed their content into a massive AI machine?

Larger publishers agree. The New York Times recently banned the use of its content to train AI models. It had previously dropped out of a coalition led by IAC that was trying to jointly negotiate scraping terms with AI vendors, preferring to arrange its own deals on a case-by-case basis. A month earlier, the Associated Press had made its own deal to license its content to OpenAI, giving it a purported first-mover advantage. The terms of the deal are not public.

Questions about copyright — and specifically the unlicensed use of copyrighted material to produce a commercial product — persist. The Authors Guild has written an open letter asking them to license its members’ copyrighted work, which is perhaps a quixotic move: rigid licensing and legal action is likely closer to what’s needed to achieve their hoped-for outcome. Perhaps sensing the business risks inherent in using tools that depend on processing copyrighted work to function, Microsoft has promised to legally defend its customers from copyright claims arising from their use of its AI-powered tools.

Meanwhile, a federal court ruled that AI-generated content cannot, itself, be copyrighted. The US Copyright Office is soliciting comments as it re-evaluates relevant law, presumably encompassing the output of AI models and the processes involved in training them. It remains to be seen whether legislation will change to protect publishers or further enable AI vendors.

The ChatGPT homepage

So. Who’s making money from AI? It’s mostly the large vendors who have the ability to create giant models and provide API services around them. Those vendors are either backed by venture capital investment firms who hope to see an exponential return on their investment (OpenAI, Midjourney) or publicly-traded multinational tech companies (Google, Microsoft). OpenAI is actually very far from profitability — it lost $540M last year. To break even, the company will need to gain many more customers for its services while spending comparatively little on content to train its models with.

In the face of criticism, some venture capitalists and AI founders have latterly embraced an ideology called effective accelerationism, or e/acc, which advocates for technical and capitalistic progress at all costs, almost on a religious basis:

Technocapital can usher in the next evolution of consciousness, creating unthinkable next-generation lifeforms and silicon-based awareness.

In part, it espouses the idea that we’re on the fringe of building an “artificial general intelligence” that’s as powerful as the human brain — and that we should, because allowing different kinds of consciousness to flourish is a general good. It’s a kooky, extreme idea that serves as marketing for existing AI products. In reality, remember, they are not actually intelligence, and have no ability to reason. But if we’re serving some higher ideal of furthering consciousness on earth and beyond, matters like copyright law and the impact on the environment seem more trivial. It’s a way of re-framing the conversation away from author rights and considering societal impacts on vulnerable communities.

Which brings us to the question of who’s not making money from AI. The answer is people who publish the content and create the information that allow these models to function. Indeed, value is being extracted from these publishers — and the downstream users whose data is being fed into these machines — more than ever before. This, of course, disproportionately affects smaller publishers and underrepresented voices, who need their platforms, audiences, and revenues more than most to survive.

On the internet, the old adage is that if you’re not the customer, you’re the product being sold. When it comes to AI models, we’re all both the customer and the product being sold. We’re providing the raw ingredients and we’re paying for it to be returned to us, laundered for our convenience.

· Posts · Share this post

 

Some newsletter changes

I’m making some experimental updates to my newsletter:

Starting next week, this newsletter will come in several flavors:

Technology, Media, and Society: technology and its impact on the way we live, work, learn, and vote.

Late Stage: personal reflections on living and surviving in the 21st century.

The Outmap: new speculative and contemporary fiction.

Most of Technology, Media, and Society will continue to be posted on this website. I am experimenting with publishing more personal posts and fiction over there.

Prefer to subscribe via RSS? Here’s the feed URL for those posts.

· Posts · Share this post

 

In defense of being unfocused

Literally an unfocused photo of a sunset. Yes, I know it's a little on the nose. Work with me here.

I spent a little time updating my resumé, which is a process that basically sits at the top of all the things I least like to do in the world. This time around I tried to have an eye towards focus: what about the work I do might other organizations find valuable? Or to put it another way: what am I?

I grew up and went to school in the UK. At the time, the A-level system of high school credentials required you to pick a narrow number of subjects to take at 16. In contrast to the US, where university applications are more universal and you don’t pick a degree major until you’ve actually taken courses for a while, British applicants applied for a major at a particular institution. The majors available to you were a function of the A-level subjects you chose to take. In effect, 16 year olds were asked to pick their career track for the rest of their lives.

I now know that I take a kind of liberal arts approach to product and technology leadership. My interests are in how things work, for sure, but more so who they work for. I care about the mechanics of the internet, but I care more about storytelling. I’m at least as interested in how to build an empathetic, inclusive team as I am in any new technology that comes along. The internet, to me, is made of people, and the thing that excites me more than anything else is connecting and empowering them. I’ll do any work necessary to meet their needs - whether it’s programming, storytelling, research, design, team-building, fundraising, or cleaning the kitchen.

Which means, when I picked my A-levels in 1995, and when I applied for universities two years later, that it was hard to put me in a box.

My high school didn’t even offer computing as a subject, so I arranged to take it as an extra subject in my own time. The standardized tests were so archaic that they included tape drives and punchcards. Meanwhile, my interest in storytelling and literature meant that I studied theater alongside more traditional STEM subjects: something that most British universities rejected outright as being too unfocused.

I have an honors degree in computer science but I don’t consider myself to be a computer scientist. I’ve been a senior engineer in multiple companies, but my skillset is more of a technical generalist: technology is one of the things I bring together in service of a human-centered strategy. I like to bring my whole self to work, which also includes a lot of writing, generative brainstorming, and thinking about who we’re helping and how best to go about it.

Even the term human-centered feels opaque. It just means that I describe my goals and the work I do in terms of its impact on people, and like to figure out who those people are. It’s hard to help people if you don’t know who you’re helping. People who say “this is for everyone!” tend to be inventing solutions for problems and people that they only imagine exist. But there’s no cleanly concise way of saying that without using something that sounds like a buzzword.

So when I’m putting together a resumé, I don’t know exactly what to say that ties together who I am and the way I approach my work in a way that someone else can consume. Am I an entrepreneur? I have been, and loved it; I like to bring that energy to organizations I join. A product lead or an engineering manager or a design thinker? Yes, and I’ve done all those jobs. I think those lines are blurry, though, and a really good product lead has a strong insight into both engineering and design. I’ve also worked on digital transformation for media organizations and invested in startups at an accelerator — two of my favorite things I’ve ever done — and where do I put that?

In the end, I wrote:

I’m a technology and product leader with a focus on mission-driven organizations.

I’ve designed and built software that has been used by social movements, non-profits, and Fortune 500 companies. As part of this work, I’ve built strong technology and product team cultures and worked on overall business strategy as a key part of the C-suite. I’ve taught the fundamentals of building a strong organizational culture, design thinking, product design, and strategy to organizations around the world.

I’m excited to work on meaningful projects that make the world better.

I’ve yet to get feedback on this intro — I guess that’s what this post is, in part — but it feels close in a way that isn’t completely obtuse to someone who’s basing their search on a simple job description. It will still turn off a bunch of people who want someone with a more precise career focus than I’ve had, but perhaps those roles are also not a good fit for me.

Perhaps I should be running my own thing again. I promised myself that I would give myself a third run at a startup, and it’s possible that this is the only thing that really fits. At the same time, right now I’m doing contracts, and I love the people and organization I’m working with right now.

If I think of my various hats as an a la carte menu that people can pick from rather than an all-in-one take-it-or-leave-it deal, this kind of work becomes less daunting. Either way, I do think it’s a strength: even if I’m working as one particular facet officially, the others inform the work I’m doing. As I mentioned, I think it’s helpful for an engineering lead to have a product brain, and vice versa. It’s not a bad thing for either to understand design. And every lead needs to understand how to build a strong culture.

But how to wrap all of that neatly up in a bow? I’m still working on it.

· Posts · Share this post

 

Press Forward brings much-needed support for local news

A man speaking into a number of microphones.

I was pleased to see this announcement from the MacArthur Foundation:

A coalition of 22 donors today announced Press Forward, a national initiative to strengthen communities and democracy by supporting local news and information with an infusion of more than a half-billion dollars over the next five years. Press Forward will enhance local journalism at an unprecedented level to re-center local news as a force for community cohesion; support new models and solutions that are ready to scale; and close longstanding inequities in journalism coverage and practice.

I think this is huge. As I wrote the other day, I think building a commons of tightly-focused newsrooms is absolutely key:

A wide news commons, comprised of many smaller newsrooms with specific areas of focus, as well as the perspectives of individuals in the community, would improve our democracy at the local level. In doing so, it would make a big difference to how the whole country works. I’d love to see us collectively make it happen.

The new initiative has a few key areas:

Strengthen Local Newsrooms That Have Trust in Local Communities: the announcement suggests they will provide direct philanthropic funding to exactly the kinds of newsrooms I’ve been talking about.

Accelerate the Enabling Environment for News Production and Dissemination: Providing shared infrastructure of all kinds is going to be really important. As a rule, I believe newsrooms should be spending their time and resources on things that make them uniquely viable. The various commodity resources that every newsroom must build — technical tools, legal assistance, revenue experiments, help with people operations, assistance with reaching audiences — should be shared so that everyone can take advantage of improvements an discoveries, in a way that keeps costs low for all.

Close Longstanding Inequalities in Journalism Coverage and Practice: ensuring “the availability of accurate and responsive news and information in historically underserved communities and economically challenged news deserts” is vital here. Again, as I mentioned: direct subscriptions don’t work in communities were few can afford to pay. Philanthropic support can help ensure peoples’ stories are told — and when they are, local corruption measurably decreases.

Advance Public Policies That Expand Access to Local News and Civic Information:‌ supporting public policies that will protect journalists and improve support for newsrooms.

My hope is that most of the money will go directly to newsrooms, and to the sorts of shared infrastructure that every newsroom needs. I also hope that this shared infrastructure will be open sourced as much as possible, so that any public interest organization can take advantage — thereby increasing the impact of these donations. While public policy support is important, communities need coverage now, particularly in the run-up to the 2024 election.

· Posts · Share this post

 

· Posts · Share this post

 

My technology thesis

Lights as seen from space, showing networks of communities

My technology thesis is this:

All technology should punch up, not down.

It should connect the disconnected. It should empower the disempowered. It should inform the uninformed.

It should empower communities that historically have been underrepresented and underserved.

It should help people with similar needs and values to organize together in order to improve their quality of life.

It should help spread wealth and power more equitably. It should allow people to make things they could not previously make. It should let people find audiences they could not previously reach.

It should help us make a positive impact on each other, and on the planet.

It should not replace the social frameworks and safety nets that are required to ensure everybody has the ability to have a good life. It should not be a tool for wealth hoarding or rent seeking. It should not privatize the parts of peoples’ lives that were previously their own.

It should not fortify the power of people who already have it. It should not maintain the status quo.

It should not make the world less democratic. It should not make the world more dangerous. It should further mutual learning and empathy, not hate and parochialism.

It should amplify and accelerate the power of communities and of individuals, of collaboration, and of mutual aid.

It should never simply be a vehicle for power. It should never simply be a vehicle for wealth. It should always be for people.

Because that’s all we are. That’s all we can build. Networks of people, all inexorably connected.

We can all rise together, or we can all sink together. Those are our choices.

 

Also see: more concrete thoughts on how to build technology.

· Posts · Share this post

 

The racist would-be CEO king

You’ve all heard about Elon Musk blaming the Anti-Defamation League for the erosion of Twitter/X’s value over the last year. Advertising revenue is down by 60%.

This, of course, has nothing at all to do with the precipitous rise in hate speech on the platform since Musk took it over, to which he has responded by threatening legal action instead of doing something about the problem.

Musk’s typically bull-in-a-china-shop approach has been winning fans. Lately we’ve taken to ending that preceding sentence with “on the right”, but these people go far beyond the tax policies and routine crustiness of store brand conservatism. His approach has been winning fans, to be clear, in communities that support literal white supremacy.

Andrew Torba, the fully-racist founder of Gab, posted on X:

In under five years we went from having every single one of our guys banned from the big tech platforms to the richest man in the world noticing, naming, and waging total war on our largest enemy while running one of those platforms. Let that sink in. Keep the faith. We are winning.

This all further cements Musk’s purchase as being part of the backlash to gains on diversity and inclusion that took place during the pandemic. I consider them all to be the last echoes of the 20th century. I will not support them, obviously.

The thing about these communities, though, is that they make everything they touch toxic. The former Twitter was famously problematic at the time, even if it was at a level that seems quaint now. They didn’t block white supremacists because they had a good, ethical heart; they were capitalists just like everyone else who runs a big platform. They knew, though, that the racists diminish the quality of the product for everyone else, and that usage and advertising revenue would be adversely affected. Which is pretty much what Musk is finding out right now (although his personal goals don’t appear to be to create a viable business, hence his lack of motivation to change).

I’d rather spend my time and energy with platforms and people that are part of building an inclusive, equal future for everyone. I don’t know that this is a widely-shared value, although I wish it was. But I also want to spend my time on platforms that are fun to use — and I’m pretty sure that’s true of most people who use social media.

I don’t want to diminish the danger — we must always be vigilant of these toxic ideals — but I’m also confident they’ll lose. These assholes will fade into obscurity soon enough.

· Posts · Share this post

 

Building a wide news commons

An array of newspapers on sale

Doc Searls writes about what he calls wide news:

Local and regional papers covered politics, government, crises, disasters, sports, fashion, travel, business, religion, births, deaths, schools, and happenings of all kinds. They had reporters assigned across all their sections. No other medium could go as wide.

Doc’s argument is that a local commons of publications can, together, create a wide news ecosystem that fulfills the same role (with potentially deeper content). I agree.

I started my career building the first website and BBS for a local paper in Oxford that carried classified ads as well as event listings, notices, and that sort of thing: all the community stuff that the internet took over from newspapers. (As it happens, it’s still around, but most are not.) It was a real community hub, to the extent that anyone could come to the office to do some word-processing or get their photocopying done: a co-working space in the midst of the paper’s offices, long before anyone knew what co-working was.

Social media — and early on, blogging in particular — has played this role of reporting widely around a community. You could click through to local blogs, or Twitter, and learn about things that happened around your town. I found this particularly useful when I visited somewhere I didn’t live: for example, on my regular visits to San Francisco, I’d check out the blogs and Upcoming to see where I should be going.

But as social media has consolidated, many of those venues have gone away. (Some were replaced by Twitter bots, which have now also gone away.) It’s also easier to discover some types of voices than others: for someone to post regularly to the internet, they need to have a certain level of free time, technical prowess, equipment at their disposal, and so on. And it’s easier for someone in an “in” crowd to be linked to and re-shared than an entirely new voice with an underrepresented perspective, who might not have the same level of systemic support.

While newspapers are obsolete technology, local newsrooms are vitally important: journalists have a remit to tell stories that might not otherwise be told. That might be a story about corruption in the local police or government (which occurs a surprising amount of the time), but it could just as easily be a story about an immigrant starting a new business, or a trend piece centered on a less-affluent part of town.

I really love what Tiny News Collective is doing here. From its website:

Have you struggled to find stories relevant to you and your community in existing media? Do you worry that lack of information keeps people from being involved in important local issues? Has it been your dream to see your community represented accurately and thoughtfully in the media?

The organization then provides the funding, technology, training, and network to get those newsrooms started. Consider Austin Vida, which reports with a Latinidad perspective from Austin, or Ang Diaryo, which reports with a lens centered on working class Filipino communities in Los Angeles.

Locally to me, Kensington Voice reports on a neighborhood in Philadelphia that normally is the subject of stories about addiction (including a New York Times piece that called it the “Walmart of Heroin”). The real, three-dimensional human beings who lived there were woefully underserved, until journalists began to report on their actual stories. These weren’t holes that social media could adequately fill; nor were they covered by the local newspaper of record.

With a skeleton team, Open Vallejo reports on the North Bay town that happens to have one of the most corrupt police forces and local governments in the country. It’s rightly won awards for its work.

I’ve fallen in love with these kinds of small, tightly-focused, non-profit newsrooms. One thing that’s missing is a way to find all the newsrooms that cover a geographic area, or a particular demographic, or other focus area. I want to be able to discover stories from newsrooms I’ve never heard of, based on their characteristics. I might have never heard of a particular newsroom in Philadelphia, but I have heard of Philadelphia, and happen to live there. It would be great if I could click through a website to find everyone publishing about communities in the city. The front page of that site — which, in essence, would be a specialized feed reader — would be better by far than any large newspaper.

To make that really viable, there needs to be a flowering of tightly-focused newsrooms — and people to fund their reporting. Crowdfunding, direct sales, or subscriptions are not always the right approach, because, for example, people in a less-affluent community are typically less able to pay, while their stories are no less valuable. Just as bloggers are less likely to write from the perspective of communities that can less afford computers, broadband internet connections, and the time to write, funding dollars must sometimes be found elsewhere. Tiny News Collective is wonderful, as I mentioned. The Brown Institute for Media Innovation’s Local News Lab does great work, albeit around business innovation rather than direct funding. And there is some wider foundational support. Still, making more support available for these newsrooms would go a long way.

All of this is not to say that personal sharing and social media aren’t valuable. They clearly are, and a lot of people are reporting their perspectives and sharing what they love about their communities. We need more of this, too! One of my favorite local social media accounts is Caffs not cafes, which reports on small, independent, usually low-budget restaurants and cafes in London. Sure, it’s not breaking news, but who else would do this kind of local coverage of diverse businesses, many of which have immigrant owners?

A wide news commons, comprised of many smaller newsrooms with specific areas of focus, as well as the perspectives of individuals in the community, would improve our democracy at the local level. In doing so, it would make a big difference to how the whole country works. I’d love to see us collectively make it happen.

· Posts · Share this post

 

I don't want my software to kill people

A screen full of JavaScript code

Dave Winer poses:

If you think of yourself as an "open source developer" please ask yourself this question. Are you as committed to freedom for people who use your software as you are to freedom for developers? Not just freedom to modify the source code, but freedom to do anything they like with the stuff they create. If not, why, and where do you draw the line?

I’m not sure if I do consider myself an open source developer these days. I don’t have the time or bandwidth to write software for myself on a regular basis in the way that I used to. I have the software I help with in my work (which is, these days, more about team dynamics and process rather than writing code); that’s about all I have time for outside of my family. I am having a lot of trouble making any time at all for my own projects.

But I used to write a lot of open source code (Elgg, Known, more contributions elsewhere). And I’ve spent a lot of time thinking about this subject.

I think we have to consider that the principles of the free software movement, revolutionary though they genuinely were, were also set in the same mindset that latterly saw its founder Richard Stallman spectacularly fall from grace. They are principles that deal in software development and licensing in strict isolation, outside of the social context of their use. They are code-centered, not human-centered.

Dave’s question has two angles that I’d like to discuss: one briefly, and the other at more length (because it’s more controversial in open source circles).

The first is: how easy is open source software to use, anyway? Can users do anything they like with the stuff they create? Doesn’t a commitment to user freedom also necessitate a commitment to ease of use? I think yes, but open source projects rarely have capacity for design or user experience research, and even when people with those skillsets want to contribute, projects quite often don’t know what to do with them. The tools (from GitHub on down), the culture, the mindsets are all code-first. There is no good way to open source user research or the empathy work that is a core part of software development. A code-centric approach takes the humanity out of software, and work has to be done to put people back in the center.

The second, more complicated one, is: I don’t want my software to be used to cause harm.

You could couch that in liability. Many software licenses disallow use in a nuclear facility, for example. But I want to go further. I don’t want anything I built to be used to kill people; nor to discriminate against them; nor to commit hate crimes; nor to intentionally organize or facilitate any act of violence or assault.

I think many software developers would feel the same way. But any license that incorporated clauses to this effect would fail to be recognized by the Free Software Foundation or the Open Source Initiative.

My blunt take on that is that I don’t care: clearly the principle of not causing harm is more important than recognition by some foundations (and particularly not foundations like the FSF whose leaders have been found to be so lacking in empathy). If the idea of not causing harm is outside the realm of the existing open source movement, then we need a new movement.

The word “free” in free software is famously overloaded. It’s “free as in speech, not free as in beer”. But there are many kinds of free speech, and even in America, where it’s the First Amendment to the Constitution, there are limits to it.

It’s worth considering whose freedom we value. Do we value the freedom of the people who use software, or do we also value the freedom of the people the software is used on? While the latter group doesn’t always exist, when they do, how we consider them says a lot about us and our priorities.

Take a drone used in warfare out in the field which incorporates an open source library that had originally been developed for some other purpose. The author released it under a license that dictated how it could be modified and shared. Shouldn’t they also have a right to say that you can’t use it in a bombing campaign? Open source principles say no.

Consider a police AI system that is used to pre-emptively target people who might commit a crime. Because of underlying biases both in the corpus of data the model was trained on and in the police force itself, and because of a fundamental disconnect between the Minority Report promise of this technology and what it can actually deliver, they tend to be wildly discriminatory and are essentially a new cover for racial profiling. Shouldn’t a software library author be able to opt out from being a part of this kind of system? Open source principles, once again, say no.

Or, closer to home for me, take an open source community platform that is used by neo-Nazis to publish propaganda about Jewish people, or to organize acts against specific people or organizations. The authors might have designed it for use with aid workers or in education, but open source licenses make no restriction on other uses.

Code does tend to find other uses. I once co-organized a demo day when I was at Matter Ventures, and had the privilege of chatting with Chelsea Manning, who was in attendance. I asked her what she thought; she was glowing about some ventures, but then went through a point-by-point list of which platforms on show could be used for military and surveillance purposes in the hands of the wrong investors or acquirers. It was one of the most eye-opening conversations of my life.

When an author releases code to the open source commons, they invite others to enter into a relationship with them. Those third parties can incorporate the code into their own projects under some restrictions, and modify and re-share it under others. The exact nature of how open source code may be incorporated, modified, and re-shared varies from license to license. But other restrictions are not a stretch. The author is giving their work away for free; this is not work for hire. They should have the right to restrict its use. They should not have to simply accept that someone could use their work to kill people, commit hate, perpetuate systemic injustices, or otherwise harm. There is nothing good and principled about that idea.

There is also no need for the FSF or OSI to be the sole arbiters of what is free or open source software. The only thing that really matters is how authors want to release their work, how downstream users might incorporate it, and how the rights and well-being of people it is used on are affected.

This isn’t just about warfare, systemic discrimination, or hate crimes (although those all should be enough). There are questions here about the rights of software authors, and the role of software in a just and equitable society. To limit our considerations to code is to say we don’t care about the people affected by our work. And to do good work, we must care.

· Posts · Share this post

 

The notable list: September 2023

A woman wearing a VR headset under an LED lace curtain

This is my monthly roundup of the links, books, and media I found interesting. Do you have suggestions? Let me know!

Apps + Websites

Productivity

iA Presenter. I’ve been really enjoying this. It does have the unfortunate effect of reducing the time you spend faffing with slide design and font choices, which means you actually have to write the substance of your presentation. Curses! Still, despite its attempts to thwart my procrastination, it’s beautifully designed and perfect for the way I think.

Media

404 Media. A new, independent, worker-owned venture by ex-Motherboard journalists. I’m a subscriber.

Center for News, Technology & Innovation. I would love to contribute to something like this.

Technology

Educational Sensational Inspirational Foundational. A really great list of foundational and/or influential writing about how to build the web, starting with Tim Berners-Lee’s Cool URIs Don’t Change post from 1998.

Datasette Cloud. Simon Willison’s Datasette now has a SaaS version that saves you having to install or set anything up. This is perfect for smaller newsrooms and orgs that are technically stretched but want to analyze data. I’m excited to see where he goes with it.

StreetPass for Mastodon. Genuinely brilliant. StreetPass finds the Mastodon accounts of people whose websites you browse, allowing you to check out their accounts and follow if you’re interested. I love it.

Books

Fiction

Yellowface, by R. F. Kuang. This tale from a deeply unreliable, envy-driven narrator is more of a sharp satire of liberal racism than its publishing industry setting. It’s at its least compelling when discussing Twitter drama, but there’s ample snark just underneath each turn of phrase, and more than enough ratcheting tension to have kept me turning the pages.

Foundry, by Eliot Peper. A knockabout spy adventure that takes a few unexpected turns and sticks a landing that had me cheering. Truly a lot of fun - I inhaled it in one sitting. As always, it’s deeply researched, but the detail only ever adds to the entertainment. (Without spoiling anything, I’m very familiar with some of the settings and cultural overtones, and they rang completely true.) There are knowing callbacks to some of Eliot’s earlier work, but this stands alone - and could be the start of a new series that I would gladly read the hell out of.

Nonfiction

Reading for Our Lives: A Literacy Action Plan from Birth to Six, by Maya Payne Smart. In turns reassuring and helpful, this was a great primer on what to do to provide a foundation to help my child eventually learn to read. If only all parenting books could be as human and equity-minded as this one is.

Notable Articles

AI

US Copyright Office wants to hear what people think about AI and copyright. I certainly have some thoughts that I will share. Imagine if you could allow an AI agent to create copyrighted works at scale with no human involvement. It would allow for an incredible intellectual property land grab.

The A.I. Surveillance Tool DHS Uses to Detect ‘Sentiment and Emotion’. Customs and Border Protection is using sentiment analysis on inbound and outbound travelers who “may threaten public safety, national security, or lawful trade and travel”. That’s dystopian enough in itself, but there’s no way they could limit the trawl to those people, and claims made about what the software can do are dubious at best.

This AI Watches Millions Of Cars And Tells Cops If You’re Driving Like A Criminal. A good rule of thumb is that if technology makes something feasible, someone will do it regardless of the ethics. Here, AI makes it easy to perform warrantless surveillance at scale - so someone has turned it into a product and police are buying it.

New York Times considers legal action against OpenAI as copyright tensions swirl. Whether this comes to fruition with the NYT vs OpenAI or another publisher vs another LLM vendor, there will be a court case like this, and it will set important precedent for the industry. My money’s on the publishers.

School district uses ChatGPT to help remove library books. Probably inevitable, but it nonetheless made my jaw drop. What an incredibly wrong-headed use of an LLM.

New York Times: Don't use our content to train AI systems. The NYT’s new terms disallow use of its content to develop any new software application, including machine learning and AI systems. It’s a shame that this has to be explicit, rather than a blanket right afforded to publishers by default, but it’s a sensible clause that many more will be including.

We need a Weizenbaum test for AI. “Weizenbaum’s questions, though they seem simple—Is it good? Do we need it?—are difficult ones for computer science to answer. They could be asked of any proposed technology, but the speed, scope, and stakes of innovation in AI make their consideration more urgent.”

AI social media videos depict missing, dead children narrating their stories. Utterly ghoulish.

Google says AI systems should be able to mine publishers’ work unless companies opt out. I strongly disagree with this stance. Allowing your work to be mined by AI models should be opt-in only - otherwise there is no possible way for a publisher or author to apply a license or grant rights.

AI language models are rife with political biases. Different AI models have different political biases. Google’s tend to be more socially conservative - possibly in part because they were trained on books rather than the wider internet. Regardless of the cause, this is proof, again, that AI models are not objective.

In every reported case where police mistakenly arrested someone using facial recognition, that person has been Black. Black faces are overrepresented in databases used to train AI for law enforcement - and some facial recognition software used in this context fails 96% of the time. This practice is an accelerant for already deeply harmful inequities. Time to ban it.

Catching up on the weird world of LLMs. This is a really comprehensive history and overview of LLMs. Simon has been bringing the goods, and this talk is no exception.

Climate

Is Big Oil Turning on Big Auto? It makes sense that oil companies would try to frame driving a gas car as freedom. As an EV driver, I can tell you that it is not. I would prefer if we all had great, integrated public transit - but for the moment, at least, it has been an improvement in every way for me as a driver. I’ll never go back.

The true cost of climate pollution? 44% of corporate profits. I’m surprised that mandatory disclosure of carbon emissions isn’t widespread - it does seem like the prerequisite to making any change. And yeah, these companies should pay. And be forced to reduce their emissions. And be fined heavily, and prosecuted, when they don’t.

Neoclassical economists are the last people to listen to on climate change. Interesting commentary on “economic theories that have led to government by markets, fuelling financial and other shocks, and the rise of authoritarian, and even neo-fascist regimes promising citizens ‘protection’ from ‘globalised’ markets.”

Climate change is death by a thousand cuts. “Whenever someone says, “we’ll adapt to climate change,” 100% of the time it’s a rich person. Poor people never say “we’ll adapt” because they know they can’t afford it. For them, adaptation = suffering.” That’s the pull-quote for me: this won’t affect everyone equally. As always, the most vulnerable, the people who are already struggling the most, will suffer the worst of it.

Culture

Being Black in a Small Town. “When popular culture thinks of Blackness, rarely does somebody think of a tiny little town or a mountainside and the Black person who’s there. I want to be a part of revealing that this thread—that Black skin—can be even on the side of a mountain.”

How to Uphold the Status Quo: The Problem With Small Town Witch Romances. I see this as less of a problem in cozy witch fiction - which, I must be clear, I have read zero of - and more of an issue in American fiction as a whole, across all media. These books (probably) aren’t actively laundering racist ideas; they’re perpetuating cultural discrimination that is under the surface everywhere. Still, it’s incumbent on authors to understand and be accountable to the tropes they’re building with.

thoughts on the suicidal mind. This resonated with me a lot. What I’ll say is: I’m glad Winnie is in the world. I know these feelings, intimately. I don’t have much definitive to say about that. I haven’t drawn any conclusions. It’s a journey, daily.

Turn-On Found. None of this looks like it comes from 1969. Although some of the content is outdated today, the style is far more modern - this feels like something straight from the internet era. Fascinating and relentless (I couldn’t watch the whole thing).

Democracy

Supreme Risk: An Interactive Guide to Rights the Supreme Court Could Take Away. “An interactive guide to rights the Supreme Court has established — and could take away.” Published a few months ago, but completely relevant, on-point reporting (served as a fully-static web page).

Americans Rate Dallas and Boston Safest of 16 U.S. Cities. Republicans think cities are much less safe than Democrats do. San Francisco and Philadelphia (my old neighborhood and new one) are notable here: Democrats agree that they’re pretty safe, whereas Republicans seem to think they’re war zones. I think we can solidly blame conservative media propaganda for this.

Just 23% Of Americans Know The U.S. Has Failed To Pass An Internet-Era Privacy Law. Less than a quarter of Americans know they don’t have meaningful privacy protections on the internet. The first step to changing this fact might be to change this number.

The Shocking Voter Purge Crisis of Democracy Revealed. Always a good sign when a democratic movement wants to win through the will of the people rather than through obstructive election fraud.

House GOP adds dozens of anti-LGBTQ+ provisions to must-pass bills. Smuggling naked bigotry through bills that must pass to keep the government working is a deeply underhanded tactic. It’s hard to see the modern Republican Party as anything other than a party of exclusion, catering to the dregs of the twentieth century who desperately don’t want to see the world change around them.

Equality

Most students haven’t learned about LGBTQ+ issues in school, survey shows. Why the internet - as well as more traditional media like books - are a lifeline for kids hungry to learn about queer history. Of course, I’m sure the usual suspects will come for those too.

International Chess Org: Trans Women Have "No Right To Participate" In Women's Chess. This stance by the International Chess Federation is so transparently bigoted that it helps clarify other anti-trans measures happening across competitive sports. There’s nothing here about fairness; it’s all to do with conservative division and hatred.

Henrietta Lacks family to get compensation for use of her cell taken decades ago without consent. Late as it is, it’s good to see this to some kind of resolution. I hope the posthumous recognition Lacks receives includes the story of how it happened in the first place.

Health

'Horribly Unethical': Startup Experimented on Suicidal Teens on Social Media With Chatbot. Taking lean startup research techniques that were developed for basic social networks or, say, 3D avatars and transposing them to real-world domains with real consequences seems to be an ongoing trend. It’s a misunderstanding of the startup playbook that causes real harm. This is obviously unethical; it is nowhere near as “nuanced” as this CEO says it is.

Labor

Remote workers' connection to companies' missions hits record low. Remote workers feel less connected to company missions, but the big message here is that nobody really feels all that connected. There are no superficial answers here: the real differentiators are better company cultures where people feel truly valued, much stronger communication, and better missions.

NLRB Says Companies That Union-Bust Must Recognize Busted Union. A neat rule: union-busters must recognize the unions they’re trying to undermine. The union rebound continues.

Why the Hollywood strike matters to all of us. On the wage threat of AI: “Hollywood is showing us how best to take that stand: by unionizing our workplaces, and fighting for strong contracts. Now’s the time to form a union with your coworkers, and discuss what protections you’ll need to face this moment.”

Negative Space. A perfect piece on where we’re at in time. Personally, I’m not going back to the office, and I applaud greater worker power. We need to move forward.

We're now finding out the damaging results of the mandated return to the office–and it's worse than we thought. Return to Office mandates are counterproductive and destroy morale. They also make your team less productive. They’re worker-hostile and work-hostile. So why do them?

Media

Medium is for human storytelling, not AI-generated writing. Medium has made it clear that it is not a home for AI-driven content. And it’s experiencing record growth now that its recommendation engine has been re-tuned for substance, as decided by humans. This is all great news: for Medium and as an example for everyone on the web.

How We Create Custom Graphics at The Markup. I like this approach to building graphics for journalism. Management of these kinds of static assets feels like a cumulative problem, but lightweight HTML / CSS / JS is pretty portable and sandboxable. And ACF is the hidden hero behind journalism’s WordPress sites.

Society

My Caste. “Allow me to introduce you to one of the largest population groups in India, as recognized by the constitution of India: Other Backward Classes. I belong to OBC Category. […] I was, however, not ready to publicly declare it until I received tenure as it seemed too risky.”

Police departments pull school officers due to Minnesota restraint law. It says a lot that in areas where officers aren’t allowed to put schoolchildren in holds that restrict breathing or their ability to speak, departments take officers out of schools in protest. These laws should be in place everywhere, and police officers should not be in schools.

She Just Had a Baby. Soon She'll Start 7th Grade. There are so many stories like this one. There should never be another. And yet, we’ve rolled back the clock at the behest of religious extremists, so there will be many more. This cannot go on.

FAU Study: Perils of Not Being Attractive or Athletic in Middle School. Hey, sounds like my middle school experience! This is important for me to understand as a parent, and it’s important for schools to adapt to as de facto caregivers. These dynamics should be corrected for, not accepted.

Right-Wing Writer Richard Hanania's Racist Past Exposed. A prominent writer platformed by the New York Times and Washington Post, and championed by major figures in tech, including by Marc Andreessen and the CEO of Substack, turns out to be an actual white supremacist.

A beautiful, broken America: what I learned on a 2,800-mile bus ride from Detroit to LA. I’ve traveled across America four times: three by car and one by train. I’ve never done it by Greyhound, and I probably never will. This country’s infrastructure is falling apart and being eaten by wolves.

Startups

The State of Seed Stage Funding to Underrepresented Founders. “White women founded companies comprise 79% of reported early-stage VC dollars going to underrepresented founders and 64% of investments made into companies with underrepresented founders by deal count. Ecosystem-wide, we need to up our game by investing seed money into a broader spectrum of founders of color.”

letter to a friend who is thinking of starting something new. These are the right questions to ask.

Technology

In Europe, a regulatory vise tightens around big tech. Good overview. I think European tech regulations have been broadly good, establishing the anti-competitive and pro-privacy rules that US legislators have failed to enact. If we could only all be so protected.

Web Scraping for Me, But Not for Thee. Good commentary on the dissonance between vendors like Microsoft banning scraping of their platforms while simultaneously releasing products that depend on scraping other peoples’ data. Some sort of commons agreement would go a long way here, but it won’t happen while platforms can get away with this one-sided relationship.

Introducing the 100-Year Plan: Secure Your Online Legacy for a Century. I’d love to understand what prompted Automattic to offer a hosting plan for $38K. On one level, I love it - it lasts for 100 years! and I love Automattic! - but I can’t justify this, and I’m not quite sure who it’s for? If this is marketing, what are the goals?

Changes to UK Surveillance Regime May Violate International Law. The UK seems to want to break international law to retain its ability to mass surveil by forcing software vendors to break their protections for users everywhere. It’s an anti-democratic approach that puts journalists and vulnerable populations at risk. It also counter-productively undermines the UK’s own technology sector.

'We're Winning': Apple Formally Endorses Right to Repair Legislation After Spending Millions Fighting It.I’m a little bit suspicious that Apple is suddenly into right-to-repair, but broadly this is good. I just wish it was a nationwide law instead of one that is limited to California. Hopefully the idea can expand to the federal level.

The Secret Weapon Hackers Can Use to Dox Nearly Anyone in America for $15. It costs $15 to uncover an American’s personally identifiable information illegally for potentially violent purposes. But also consider the number of entities that have access to this information legally, without any oversight. None of it should be allowable.

RSS Zero isn’t the path to RSS Joy. “RSS is not email. You don’t have to get to inbox zero!” is a correct take, in my opinion; that’s certainly how I approach my feed reader. But also, I’ve got bad news about my email inbox.

Thousands of scientists are cutting back on Twitter, seeding angst and uncertainty. Scientists are fleeing X for Mastodon, citing far-right science denialism - and far-right hate in general. I don’t exactly know what Musk thinks he’s going to be left with after all this.

An Opinionated Guide To Alt-text. A great, short guide to writing alt text to support data visualization from Jasmine Mithani.

Lamborghini teases first fully electric supercar ahead of official reveal August 18th. OK, good for them, but I’m far less excited by an electric Lambo than an electric car for $20K. Or, you know, zero-emissions buses that work as part of a functional integrated public transit system. You’re right, that does sound like science fiction.

Elon Musk's Twitter throttles links to Threads, Blue Sky and New York Times. Really, truly: there is no good reason for any media company or publisher to still be posting on X.

Announcing the Tor University Challenge. This is a worthwhile project, and would be a major win for freedom of expression and freedom from surveillance. I’d love to see more of my higher education friends take part.

How to verify your Threads account using your Mastodon profile. It’s truly beautiful to see Threads begin to embrace indieweb and federated social web protocols. This is a first step; true federation is, I’ve been assured, coming.

Why Sam Altman wants to scan two billion eyes. We’ve seen the United Nations share their biometric registration of Rohingya refugees with the Myanmar government without their consent. A private company that subcontracts services in other countries makes accountability very difficult when there are rights violations.”

Raku: A Language for Gremlins. That’s a giant “nope” from me, but your mileage may vary.

PIE failed. But it’s a failure worth celebrating and learning from. It’s very painful to see accelerators that are also vibrant community hubs shut down because of business dynamics. I’ve lived that. What I can see here is someone who cares about his community. I was never a part of PIE, but I know Rick did it for the right reasons. And I know from Matter that the community continues long after the thing itself has disappeared. The legacy is long lasting. Congratulations, Rick - on to the next thing.

How I make annotated presentations. It’s been a long time since I’ve given any talks (the pandemic put a stop to that) but I really like this approach, and I’ll do something similar in the future.

Just normal web things. Yes to all of this. These are basic functions that the web gives you almost by default. Everything on the web should let you do them.

The open source licensing war is over. I broadly agree with this rallying cry against dogmatism in open source. I think dogmatism is harmful in all parts of tech; divisive and often a kind of gatekeeping. Let authors build and release according to their needs.

How to Search for a Better Deal on Broadband. The broadband situation in America is surprisingly bad - so I love that there’s a new version of the National Broadband Map. The Markup has done a public service by taking us through it.

· Posts · Share this post

 

Noah

We lost my beloved cousin Noah this week.

I don’t have the words yet. But Noah, you were wonderful, and we all loved you so much.

Ma used to say that she liked to think of everyone who was gone having a picnic on the beach and looking down at us. I would like to imagine you there, too, arriving after sailing across the bay. I bet they all greeted you with open arms.

Even as he was getting sick, Noah wanted to find ways to be helpful. If I asked you over the last year about places where a lawyer with technical skills might help with human rights or civil engagement, it was for him.

He made his career helping defendants with hard immigration and criminal justice cases. His former colleagues at Koehler Law have posted a memorial to him. I feel like anything I post here will always be inadequate — there’s no way to cover him adequately. He was a sweet, smart man who was an important part of all of our lives. We had so much in common; so many shared points of reference. I will miss him very much.

His immediate family is asking, in lieu of flowers, to donate to the Capital Area Immigrants’ Rights Coalition. I made a donation; I would love for you to join me.

· Posts · Share this post

 

The Online Journalism Awards and why non-profit news is awesome

I was pleased to attend the Online Journalism Awards on Saturday night. Some winning highlights included:

The 19th won a Breaking News award for its coverage of the Dobbs decision, including some really great data journalism. I’m proud of, and very happy for, my friends there. By the way, you should subscribe to data visuals reporter Jasmine Mithani’s great newsletter, data + feelings.

ProPublica won a few awards for its journalism, including on the proliferation of junk science in the justice system and on how viruses transmit from animals to people.

The Marshall Project won two awards for its work covering the American criminal justice system. Stories included a two-year investigation into abuses by correctional officers in New York State and a three-year story about mitigation specialists who help death penalty defendants by documenting their childhood traumas.

The Markup won an Excellence in Technology Reporting award for its reporting on broadband pricing across the US. I loved this reporting and directly used it to help a family member get a better broadband deal.

Every one of these finalists and winners is worth checking out. This is why I’m finding working in product and technology for non-profit news to be so rewarding: you get to support journalists who are genuinely making the world a better, more democratic place by shedding light on stories we need to know about.

News media in the US gets a lot of flak, and some of it is deserved. But the non-profit news industry in particular is doing incredible work, sometimes reporting stories for years on end, and putting every story out there for the public to read without a paywall in sight. These non-profit organizations deserve our personal and institutional support. They make our democracy better.

· Posts · Share this post

 

Technology isn't something that just happens to your newsroom

I’ve come away from the Online News Association conference with a really familiar feeling: somewhere between unsettled and frustrated. Not at journalists, I hasten to add, who are doing important, democratic work despite shrinking budgets and adverse conditions. But a little bit at the business sides of their organizations, and certainly at the ecosystem of vendors and evangelists that circle them.

Some quick observations:

Work on inclusion in the newsroom has stagnated in most organizations, despite the very real impacts this has on audiences and communities who depend on newsrooms to tell their stories and speak truth to power on their behalf. There is lip service here and there, but not a lot of true equity-sharing.

A few people on stage and elsewhere expressed the opinion that it doesn’t matter if journalists stay on X or not, despite the steep rise in hate on the platform. They might not be comfortable with Elon Musk, but the platform would chug along whether they were participating or not, so they might as well be there if they got something out of it.

AI vendors are out in force, expressing ways in which their software can speed up newsroom tasks, with little time being spent on the functional realities of their products or the issues this can create.

More newsrooms than I would expect are spending time writing and maintaining their own content management systems rather than leveraging existing open source software and collaborating with other organizations.

The feeling it’s left me with is similar to the one I felt when I co-founded Elgg in higher education. At least at the time, there was very little diversity in higher education decision-making; meanwhile, the software tools being deployed made it harder to learn, were inaccessible to many people, locked teaching and learning behind exploitative license agreements, and were being sold for seven figure sums. It didn’t feel right that something as fundamentally important to society as education was being locked down to a narrow demographic of decision-makers and strip-mined for value by rent-seekers. (It must be acknowledged that while accessible open source tools in education are now commonplace, rent-seekers like Blackboard still do a lot of business.)

To briefly return to each of those observations in turn:

You need diverse points of view in a newsroom (both in editorial and management) in order to be able to reflect the communities you’re both covering and trying to reach. A diverse team is more resilient; diverse teams are smarter and do better work.

Journalists have outsize power with regards to a platform like X. They create much of the content that will be shared and discovered on the platform. Their actions matter, and they can effect change in the tech industry. I think this speaks to how disempowered newsrooms have felt at the hands of technology changes over the last decade or two — but it need not be the case.

AI seems like magic but is more like a magic trick. Meredith Broussard’s discussion on recognizing inequalities in artificial intelligence is arguably vital for anyone considering adopting AI. There are genuine use cases for the technology, but her definition of techno-chauvinism — the assumption that technical solutions are better than human ones — rings true in this case.

And development teams should spend most of their time working on projects that add value to their newsroom. Working to maintain commodity technology (as in, maintaining the exact same thing hundreds of other teams are building, like a CMS) more than about 20% of the time is a waste of very scant resources. Generally, development teams should be spending their time building differentiated technology.

Every newsroom needs nuanced technical advice, but not every newsroom can afford to hire a CTO. A few organizations offer platforms, technical and business advice, and fractional technical leadership as a service for newsrooms. They’re a vital part of the ecosystem — and the truth is that some larger newsrooms need something similar. It’s all too easy to fall prey to the hype cycle, and to continue to believe that the internet is something that happens to you rather than something newsrooms can help shape and change according to their needs.

As I’ve written before, I would like to see a kind of tech union for newsrooms that would provide technical advice and commodity technology under an open source license, and then represent newsrooms in technical forums like the W3C. If the internet is a network of people, then journalism is a way for their stories to be told, and for the truth about abuses of power and systemic imbalances to come to light. It should be a virtuous relationship, and I believe it can be. I also believe it is far from this right now.

· Posts · Share this post

 

The mugshot

I think it’s important to prefix this post with the obvious: I am not a fan or apologist for Donald Trump. I think he’s nakedly undermined the workings of democracy, and has used the authoritarian playbook to build a movement that is, at its heart, anti-immigrant, anti-inclusion, and anti-progress, and shares at least some DNA with fascist movements of the past.

But I don’t like the mugshot.

I understand the glee that some people greeted it with. Of course I do; the catharsis is real. But let me lay out my disquiet:

The first is the principle. A mugshot is not a conviction. Many news organizations stopped the practice of publishing mugshots because, even if the person is exonerated, they tend to last online. As the Marshall Project wrote a few years ago:

Publishing mugshots can disproportionately impact people of color by feeding into negative stereotypes and undermining the presumption of innocence, said Johnny Perez, a formerly incarcerated New Yorker who is currently director of U.S. prison programs for the National Religious Campaign Against Torture.

Clearly Trump isn’t a vulnerable person, nor a person of color. But publishing the mugshot normalizes the practice of publishing mugshots, which is in totality more harmful than it is healthy.

The second is that I believe his base will love it. Here is this outsider leader of their movement that the libs hate so much that they’ll try and throw him in jail. The image far eclipses the real, extensive crimes that he’s been accused of. As Jesse Watters from Fox News said: “he looks good and he looks hard.” It’s real collateral for the 2024 election.

And last but not least: I’m just so fucking tired of seeing his face. It just gives him and his movement oxygen. He thrives on attention, like a vampire that sucks on primetime TV audiences. If he is found guilty, as I believe he will be, I would like him to sit out his time in jail without any more eyes, any more attention. I’d love to move on.

· Posts · Share this post