Calculating your potential reach on Mastodon with Google Cloud Workflows orchestrating the Mastodon APIs
With Twitter, you can access your analytics to know how your tweets are doing, how many views you’re getting. Working in developer relations, it’s always interesting to get some insights into those numbers to figure out if what you’re sharing is interesting for your community. But for various (actually good) reasons, Mastodon doesn’t offer such detailed analytics. However, I wanted to see what the Mastodon APIs offered.
How to calculate your potential reach
Your “toots” (ie. your posts on Mastodon) can be “boosted” (equivalent of a retweet on Twitter). Also, each actor on Mastodon has a certain number of followers. So potentially, one of your toots can reach all your followers, as well as all the followers of the actors who reshare your toot.
So the maximum potential reach of one of your posts would correspond to the following equation:
∑ ( boosters[i].followers_count )
Let’s play with the Mastodon APIs to compute your reach
Fortunately, the Mastodon APIs allow you to get those numbers, albeit not with a single API call. Let’s have a look at the interesting endpoints to get the potential reach of my most recent posts.
First of all, I’ll look up my account on the Mastodon instance that hosts me:
I pass my account name as a query parameter to the /accounts/lookup endpoint.
In return, I get a JSON document that contains various details about my account and me (I’ll just show some of the interesting fields, not the whole payload):
I get two important pieces of information here: the followers_count gives me, you guessed it, the number of followers my account has, thus the potential number of persons that can see my toots. Also the id of my account, which I’ll need for some further API calls further down.
To get the most recent statuses I’ve posted, I’ll indeed need that account id for crafting the new URL I’ll call:
This call will return a list of statuses (again, snipped less interesting part of the payload):
In each status object, you can see the number of replies, the number of times the post was reshared or favorited, or whether it’s a reshared toot itself. So what’s interesting here is the reblogs_count number.
However, you don’t get more details about who reshared your toot. So we’ll need some extra calls to figure that out!
So for each of your posts, you’ll have to call the following endpoint to know more about those “reblogs”:
This time, you’ll get a list of all the persons who reshared your post:
And as you can see the details of those persons also have the followers_count field, that tells the number of people that follow them.
So now, we have all the numbers we need to calculate the potential reach of our toots: your own number of followers, and the number of followers of all those who reshared! It doesn’t mean that your toots will actually be viewed that many times, as one doesn’t necessarily read each and every toots on their timelines, but at least, that’s an approximation of the maximum reach you can get.
Automating the potential reach calculation with Web API orchestration
Initially I played with both cURL and a little Apache Groovy script to better understand the Mastodon APIs to figure out how to chain them to get to the expected result. Then I decided to automate that series of Web API calls using an API orchestrator: Google Cloud Workflows.
To recap, we need to:
Get the details of your account
Get the recent posts for that account
Get all the followers count for each person who reshared each post
Let’s have a look at this piece by piece:
First, the workflow takes an account and server arguments, in my case that is glaforge and uwyn.net. And I’m defining a variable with the base path of the Mastodon API, and a dictionary to hold the data for each toot.
Above, I’m doing an account lookup, to get the id of the account, but also the followers count.
We get the list of most recent toots.
Now things get more interesting, as we need to iterate over all the statuses. We’ll do so in parallel, to save some time:
To parallelize the per-status calls, we just need to state it’s parallel, and that the variable we’ll keep our data in is a shared variable that needs to be accessed in parallel. Next, we define the steps for each parallel iteration:
Above, we get the list of people who reshared our post. And for each of these accounts, we’re incrementing our impact counter with the number of their followers. It’s another loop, but that doesn’t need to be done in parallel, as we’re not calling any API:
And we finish the workflow by returning the data:
This will return an output similar to this:
With this little workflow, I can check how my toots are doing on this new social media! As next steps, you might want to check out how to get started with API orchestration with Google Cloud Workflows, in the cloud console, or from the command-line. And to go further, potentially, it might be interesting to schedule a workflow execution with Cloud Scheduler. We could also imagine storing those stats in a database (perhaps BigQuery for some analytics, or simply Firestore or CloudSQL), to see how your impact evolves over time.