Feedbin is the Best Way to Read Twitter

by Ben Ubois

March, 2023 Update: Feedbin is no longer the best way to read Twitter. They have revoked Feedbin’s access and the access of every other third-party client. You can no longer subscribe to Twitter content in Feedbin.


You can now subscribe to Twitter content in Feedbin.

Tweets have become media rich, with support for multiple photos, videos and links. However, traditional Twitter clients are limited to showing tiny thumbnails and plain links. They make it too easy to mindlessly scroll through endless inane thoughts.

Feedbin treats tweets differently. The idea of the feature is to fully unpack the tweet. If a tweet links to an article, Feedbin will attempt to load the full article and display it alongside the tweet. Feedbin will also include full-size images, videos and gifs with native YouTube, Vimeo and Instagram embeds.

You can start adding Twitter content to Feedbin the same way you would subscribe to a feed. Feedbin will recognize any Twitter URL that contains tweets. It also supports shortcuts for subscribing directly to twitter @usernames as well as #hashtags. For example:

To achieve the best possible experience, I have a few recommendations:

  1. The best stuff on Twitter exists in the form of media attached to tweets like links and images. Feedbin includes a built-in filter that will only show you these tweets. The filter is on by default, but when you subscribe you’ll be able to choose to see all tweets instead.

  2. Follow fewer accounts in Feedbin. Rather than following your entire home timeline, try creating a Twitter list that only includes a few of your favorite accounts.

Twitter is deeply integrated with Feedbin and tweets include a number of new searchable fields. Using these fields you can easily find and filter tweets:

  • twitter_screen_name:"@feedbin"
  • twitter_name:"Feedbin"
  • twitter_retweet:true|false
  • twitter_media:true|false (link or image)
  • twitter_image:true|false
  • twitter_link:true|false

I’d be interested to hear your feedback on this feature. Get in touch!

Feedbin Notifier 1.1

by Ben Ubois

There’s a big update to Feedbin Notifier for iOS and Apple Watch available now.

Both the iOS app and the watch app now feature images in the notifications. This leverages the images found by Feedbin’s image finder so you’ll see the best image for a post when available.

The Apple Watch app has been re-written to make it work without your iPhone present. It will work over LTE or WiFi.

Full release notes:

iPhone & iPad

  • NEW: App Icon
  • NEW: Notifications now include images (when available)
  • NEW: Password AutoFill
  • IMPROVED: Better performance loading articles
  • FIXED: Various iOS 11 compatibility issues

Apple Watch

  • NEW: App Icon
  • NEW: Notifications now include images (when available)
  • NEW: Works over WiFi and LTE, no phone required
  • NEW: Syncs in the background so your content is always up-to-date
  • IMPROVED: Much better performance loading articles from notifications
  • FIXED: Haptic feedback once-again plays when notifications are received
  • FIXED: Marking as read from a notification

Add to Home Screen Fixed and Improved

by Ben Ubois

Feedbin once-again supports the Add to Home Screen option on iOS.

The experience of using Feedbin in Mobile Safari has been improved too. You can now swipe horizontally to navigate between the panels.

Feedbin used to support Add to Home Screen, but an update was made to home screen web apps that prevented you from navigating between pages. This meant the feature really only worked for single-page sites, because any attempt to login would kick you out to Safari.

This all changed when I came across a simple work-around. If you hijack a link click with JavaScript and then use window.location to direct the the page to the original location then you don’t get kicked to Safari. For example:

<a
  href="/login"
  onclick="window.location=this.href; return false;"
>
  Login
</a>

Podcast Playback Improvements

by Ben Ubois

There’s a few great improvements for playing and managing podcasts in Feedbin.

  1. An all new player. The player has both a mini-view and expanded-view, which features the podcasts’ artwork. You can click on the player to toggle between the views.
  2. The player is persistent. You can continue to use Feedbin while the podcast is playing.
  3. Feedbin remembers your progress for all podcasts so you can resume where you left off.
  4. A new Recently Played section so you can quickly find anything you were in the middle of listening to.

View Links in Feedbin

by Ben Ubois

Articles often link to other websites and blogs. I’ll usually open these links in a new tab as I go, to read what the links contain. However, I like to do all my reading in Feedbin because it’s a pleasant and consistent reading environment.

This feature adds the ability to view the contents of a link, all without leaving Feedbin. Only the article contents are displayed, so anything loaded this way is optimized for reading.

Feedbin Supports JSON Feed!

by Ben Ubois

Feedbin now supports subscribing to JSON Feeds.

JSON Feed is an alternative to the RSS/Atom formats. The great thing about JSON Feed is that it encodes the content as JSON instead of XML. This is good because parsing and writing XML feeds is hard.

The specification has a small surface area and is a great piece of technical writing. You should check it out. If you publish a website, consider offering a JSON Feed alongside your RSS feed.

One of the criticisms I’ve seen of JSON Feed is that there’s no incentive for feed readers to support JSON Feed. This is not true. One of the largest-by-volume support questions I get is along the lines of “Why does this random feed not work?” And, 95% of the time, it’s because the feed is broken in some subtle way. JSON Feed will help alleviate these problems, because it’s easier to get right.

I also want JSON Feed to succeed because I remember how daunting RSS/Atom parsing were when building Feedbin. If JSON Feed was the dominant format back then, it would have been a non-issue.

Testing Huge PostgreSQL Backups for 37 Cents a Day

by Ben Ubois

I spent some time revisiting and improving Feedbin’s backup strategy recently.

I was prompted by a couple of high-profile database incidents including GitLab and Instapaper.

I already felt pretty good about the backup system I had in place. I had tested backup recovery manually and felt the tools were solid.

However, if for some reason the database backup was failing or my restore procedure had a mistake, I would not know until I actually had to recover the database.

Feedbin uses WAL-E from Heroku to continually back up the database. WAL-E, and the PostgreSQL architecture that enables it, gives you offsite backups that can be recovered to any specified point in time.

WAL-E is set to back up the full database every day via Cron.

#!/bin/bash
wal-e backup-push /var/lib/postgresql/9.2/main

This command pushes the database to S3. This functions as a base backup that when combined with the WAL archives, that are continuously uploaded, can be restored to any point after the base backup started.

WAL-E offers a counterpart command, backup-fetch, to actually restore the data from a backup-push. To test the backups I needed to build in an automated way to restore the database.

Feedbin already uses Digital Ocean for a few things, so my first thought was to use Digital Ocean for this. I wrote a script to provision a Digital Ocean server, restore the backup to it and then delete the server after the backup completed.

This turned out to be too expensive. Sending data to S3 is free, but reading it back out will cost you. For Feedbin’s database this worked out to be about $40 every time I restored the database. I wanted to test backups daily but the data transfer cost would quickly add up to about $1,200/month.

While I was looking at S3 pricing, I found out that reading from S3 is free when it is read by an EC2 instance in the same region as your S3 bucket. It was also possible to save money on the EC2 instance itself by using Spot Instances instead of on-demand. With Spot Instances you bid on your instance and AWS tells you if you can have it for that price or not.

Critically, no matter what your bid is, you never pay more than the spot price which is “The current market price of a Spot instance per hour.” With this in mind you don’t have to guess what to bid and since your bid matches the on-demand price, your instance will never be terminated early due to the price exceeding your bid.

The instance I want costs $0.78/hr so I bid $0.78/hr, but only end up paying the spot price of ~$0.18/hr.

Price of database restore with On-Demand vs. Spot Instances

I was new to the AWS CLI, but once I figured out the right data to send, it turned out to be a fairly simple script.

#!/bin/bash

export AWS_SHARED_CREDENTIALS_FILE=/etc/aws.d/credentials.ini

user_data=$(base64 --wrap 0 /etc/aws.d/pg_restore)
launch_specification=$(cat <<_EOF_
{
    "UserData": "${user_data}",
    "ImageId": "ami-6edd3078",
    "SecurityGroupIds": ["sg-XYXYXY"],
    "InstanceType": "c4.4xlarge",
    "SubnetId": "subnet-XYXYXY",
    "EbsOptimized": true,
    "BlockDeviceMappings": [
        {
            "DeviceName": "/dev/sda1",
            "Ebs": {
                "DeleteOnTermination": true,
                "SnapshotId": "snap-0a7be01f9503d7bee",
                "VolumeSize": 800,
                "VolumeType": "gp2"
            }
        }
    ],
    "IamInstanceProfile": {
        "Name": "database-restore"
    }
}
_EOF_
)

/usr/local/bin/aws --region us-east-1 \
ec2 request-spot-instances \
--spot-price "0.78" \
--instance-count 1 \
--type "one-time" \
--launch-specification "${launch_specification}"

This puts in a request to launch a c4.4xlarge instance with 800GB of storage. It also specifies UserData which is executed by the instance after it boots up, which is a perfect fit for the script that configures the machine to run postgres and restore the database backup.

#!/bin/bash

# pg_restore

# Add postgresql.org offical releases as an apt source
sh -c 'echo "deb http://apt.postgresql.org/pub/repos/apt/ $(lsb_release -cs)-pgdg main" > /etc/apt/sources.list.d/pgdg.list'
apt-get install -y wget ca-certificates
wget --quiet -O - https://www.postgresql.org/media/keys/ACCC4CF8.asc | sudo apt-key add -

# Install dependencies
apt-get -y update
apt-get install -y postgresql-9.2 postgresql-contrib-9.2 postgresql-server-dev-9.2 \
build-essential python3-dev python3-pip libevent-dev daemontools lzop pv ssl-cert

# Make the postgres user a sudoer so it can shut down the machine later
echo "postgres ALL=(ALL) NOPASSWD:ALL" > /etc/sudoers.d/postgres
chmod 440 /etc/sudoers.d/postgres

# Install and configure WAL-E
python3 -m pip install wal-e[aws]
mkdir -p /etc/wal-e.d/env
echo "secret-key"   > /etc/wal-e.d/env/AWS_SECRET_ACCESS_KEY
echo "access-key"   > /etc/wal-e.d/env/AWS_ACCESS_KEY_ID
echo 's3://bucket/' > /etc/wal-e.d/env/WALE_S3_PREFIX
chown -R root:postgres /etc/wal-e.d

# Download the latest backup
service postgresql stop
rm -Rf /var/lib/postgresql/9.2/main
envdir /etc/wal-e.d/env /usr/local/bin/wal-e backup-fetch --pool-size=16 /var/lib/postgresql/9.2/main LATEST

# set the postgres recovery settings
sudo -u postgres bash -c "cat > /var/lib/postgresql/9.2/main/recovery.conf <<- _EOF_
restore_command = 'envdir /etc/wal-e.d/env /usr/local/bin/wal-e wal-fetch --prefetch=16 \"%f\" \"%p\"'
recovery_end_command = 'mail -s \"Database Restore Complete\" email@example.com && sudo shutdown -h now'
_EOF_
"

service postgresql start

This is all that is needed to stand-up a fully functioning PostgreSQL server and restore the database. No Chef, Ansible or any other provisioning tools required.

The important part here is that postgres lets you specify a command to run once recovery is complete, the recovery_end_command.

recovery_end_command = 'mail -s "Database Restore Complete" email@example.com && sudo shutdown -h now'

Here I have it send me an email and shut down the server, which terminates the EC2 instance so it’s no longer incurring cost.

If the email goes missing, then I know the restore never completed and I can go figure out what went wrong. AWS helps you out here too. The results of the UserData script are automatically logged to /var/log/cloud-init-output.log So you can seen exactly where the restore went wrong.

I would be interested in hearing any questions or comments about this.

Subscribe Improvements

by Ben Ubois

There’s a great new interface for subscribing to feeds. The process of adding a new feed has been streamlined by giving you the ability to rename and tag feeds while subscribing.

Before, subscribing, renaming and tagging were separate operations. You had to click around a lot to get your new feed in the right place with the right name.

Also, you can now see a favicon and the feed URL for what you are about to subscribe to. And the whole thing is easy to navigate using just the keyboard.

Regarding Readability

by Ben Ubois

Readability is shutting down and that means making a few changes to Feedbin.

Readability offered two services both named Readability.

  1. A read-it-later/Instapaper type of service. Feedbin offered an integration to let you easily add links to your Readability account. This has been removed from Feedbin.
  2. A parser API. Feedbin used this service to provide the full content of partial-content feeds. This functionality will continue to exist in Feedbin, but powered by Diffbot.

For now I’ve chosen Diffbot to fill in for the functionality that Readability’s parser API provided. Diffbot’s data is great. The whole company is focused on offering a suite of products that extract useful information from webpages and has a simple subscription business model so I’m optimistic that it will only improve.

I looked at a few alternatives to Diffbot, including some open-source projects and Mercury. Ultimately Diffbot’s solid data and presence of a business model made it the easy choice.

I’m planning on leaving the Readability icon in-place. Readability’s parser functionality is tough to convey in an icon, and I think that taking advantage of the brand recognition of Readability makes sense for now. Also, I like to think of it as a small homage.

Readability was a great product that Feedbin relied on for years. It will be missed!

Update

It turned out that DiffBot did not offer the performance necessary for this feature. It’s been updated to use Mercury Web Parser. Also the icon has been updated to:

Newer
Older