Using AFFiNE as your own blog - technical

Using AFFiNE as your own blog. How we did it.
All posts
Christopher Smolak
Peng Xiao
Updated: 06/28/2024

AFFiNE is great and we think it is ready for use in the wild. So in order to align our actions with our beliefs, we have decided to migrate our existing blog to AFFiNE.

In this article, we will share with you the reasons behind our decision to switch to our own product and the benefits we hope to achieve, as well as the challenges we faced along the way. We hope that our experience will inspire and guide other users who are looking to integrate and utilise AFFiNE in their own workflows.

Switching to our own product

Aside from explicit testing we are always looking for more ways we can expand how we use AFFiNE in-house. Not only do we want to better understand and relate to our users but we also are looking to promote a positive work culture.

The more we use AFFiNE the more we can embrace and appreciate the feature ideas and/or issues raised by our users. We take community feedback very seriously and we give serious thought to everything that is brought upd And by using the product internally, we will create more shared experiences among employees, which will foster our own sense of community and collaboration. We want everyone to feel proud of the work they have done, and the product that has been produced.

It will also make the process of writing on our blog much easier. Currently our blog lives together with the whole official website. So any updates may impact the whole AFFiNE homepage in their shared repo. So this also brings a rather large learning curve to any non-technical members who have to learn Markdown, Git, GitHub and how to deploy articles. AFFiNE offers the perfect solution to make blogging an easy and accessible for everyone.

Setting up the new blog

The old blog was a standard Next.js blog and post files are static markdown files that are managed in a GitHub repo: the file directory & names generate the file tree and routes; post metadata is in form of frontmatter and is placed at the beginning of a post, and content is written in markdown. Each of the files will be parsed and pre-rendered as static Next pages.

There are lots of various great community-driven projects utilising lots of different software, that allow the user to render their content managed. Whether that is using databases to other types of media, such as Next.js-powered blogs. We wondered if it would be possible to write our blog using AFFiNE.

Although the AFFiNE project is relatively new, we think it would be cool to do the same. However, we have no official public APIs that make it easy to do this yet. AFFiNE Downhills introduces the public workspace, which means your workspace content can be shared with anyone else without the need for authentication. This gives us an opportunity to read the workspace documents from the public AFFiNE API. For example, you can get the workspace data by callinghttps://app.affine.pro/api/public/doc/{workspaceId}/ .

Getting the markdown files from public workspace API

To build a Next.js site, it is essential to have a list of available files and the file contents. For real-world usage, we will also put some metadata information for the file contents, such as tags, authors, whether it is a draft or not (publish status), and so on.

Theoretically, if we can convert a yDoc to a list of markdown files, then our job is mostly done. However, the only reader of the workspace yDoc is blocksuite, but it doesn't allow us to convert yDoc to Markdown files yet. Therefore, our new AFFiNE team member, Peng , wrote a library called "affine-reader " to parse yDoc binary into markdown files with some simple functions.

Technically, affine-reader does the following:

  1. Reads the page list from yDoc
  2. For each page (which is a list of blocks), get Quill delta of each block
  3. Converts delta to markdown (using a fork of https://github.com/frysztak/quill-delta-to-markdown)
  4. Concatenates block-level markdowns into a single markdown page If you are familiar with Next.js, we can use getStaticPaths to generate the static page routes and getStaticProps to get each page’s markdown content. In Next.13 App Layout, you can even render the post list and a random post like this pretty easily:
// [workspaceId]/page.tsx
import { getBlocksuiteReader } from "blocksuite-reader";

export default async function Post(params: { workspaceId: string, pageId: string }) {
  const reader = getBlocksuiteReader({
    params.workspaceId,
  });
  const pages = await reader.getWorkspacePages());
  return (
    <div>
       <ul>{/* ... render page list */}</ul>
    </div>
  );
}

// [workspaceId]/[pageId].tsx
import { getBlocksuiteReader } from "blocksuite-reader";

import { mdToHTML } from "./md-to-html";

export default async function Post(params: { workspaceId: string, pageId: string }) {
  const reader = getBlocksuiteReader({
    params.workspaceId,
  });
  const page = await reader.getWorkspacePage(params.pageId));
  return (
    <div>
      <article dangerouslySetInnerHTML={{ __html: mdToHTML(page.md) }} />
    </div>
  );
}

The primary issue is resolved. To be more creative, we can include frontmatter at the beginning of the page to define some metadata for the current page, such as authors, tags, whether or not to publish this page, etc. Additionally, we can use MDX to enrich both the default and custom elements of the page by embedding link previews and other interactive components.

By implementing Next.js infrastructure and enabling ISR, we eliminate the need to rebuild and deploy the blog each time content is updated. Moving forward, we can focus on using AFFiNE as our CMS to write posts for the blog. This will allow us to easily manage and publish our content with minimal hassle.

Migrating the old blog

At the time of writing there were around 50 articles that needed migrating over, that would be a fairly boring manual task to perform so why not make it a little more interesting and utilise some code to help us do that a little quicker

Challenge 1: Getting our files

Our old blog lived inside the general AFFiNE homepage repo and followed a structure of /blog/{YEAR}/{MONTH}/{SLUG}/ . An example would be blog/2023/02/moving-to-app-affine-pro and then inside this folder there is an index.md file which is the contents of the article, supported by an images folder with all of the relevant images for the article.

So first things first, let's take a copy of that blog folder. It would make things easier if everything was all in one folder, not in different sub directories and that's where our first challenge comes.

What do we want to do?

  1. We want to find every file)
  2. The structure is /blog/{YEAR}/{MONTH}/{SLUG}/index.md
  3. We want to move these files to a shared directory
  4. So we cannot leave them all with the index.md name

How can we do this?

Let's just move all our files //blog/ we just need to match the pattern /blog/*/*/*/index.md so that deals with the first task fairly easily. And for thje let's use the slug (the last part of the url of the article) as that's going to give a unique name without special characters. We can get the slug from checking the last folder name /blog/{YEAR}/{MONTH}/{SLUG}/.

So how shall we go about doing this? Well lets just put together a simple little bash script.

#!/bin/bash

for file in /blog/*/*/*/index.md
do
  directory=$(dirname "$file")
  article_title=$(basename "$directory")
  mv "$file" "$directory/../../../${article_title}.md"
done
  1. So first we just declare this is a bash file #!/bin/bash
  2. Next we get every file that matches our pattern /blog/*/*/*/index.md
  3. One by one we take the matched file and then perform the do operations
  4. Let's get the current directory directory=$(dirname "$file"
  5. Let's get the article title by checking the last directory article_title=$(basename "$directory")
  6. And finally let's move all these files - take our original $file and move it to up 3 folders ../../../ to the /blog folder giving it the new article_title name. Great all our files are in the blog folder with unique names and we can just upload them now? No, not quite yet...

Challenge 2: Updating the metadata

There are some key things that our blog shows that are not strictly part of the article - these items are our metadata that contains lots of configurable options to be displayed on the article but are not part of the main content.

---
title: Moving to app.affine.pro
author: Christopher Smolak
tags: User Guide
cover: ./images/cover.jpg
description: A big step forward for AFFiNE
created: 2023-02-20
updated: 2023-02-20
layout: blog
---

Now this doesn't really prevent too many issues - our new blog system builds on the old setup and can understand this information and know to exclude it from the main article content. However, with this update we added two new features that are crucial to being shown and available on the blog.

What do we need to update?

  1. We can remove cover as it's no longer relevant - we can upload images as blobs to AFFiNE and can simply use the first image in the article as the cover image
  2. We need to add the new slug option to make our blog URLs beautiful and to avoid a mess of random characters. Good job we set our filename to our slug earlier
  3. We need to add the new publish option and set it to true

How can we do this?

Let's again go with a simple little bash script.

#!/bin/bash

for file in $(ls -t -r /blog/*.md)
do
  slug=$(basename "$file" | sed 's/\.md//' | tr '[:upper:]' '[:lower:]')
  sed -i '' '5d' "$file"
  sed -i '' "4a\ 
slug: ${slug}\

" "$file"
  sed -i '' "9a\ 
publish: true\

" "$file"
done
  1. As usual lets acknowledge this is a bash file #!/bin/bash
  2. So again we are going to do things to each file in our blog folder - for all of the .md files
  3. We want to keep the order of files (from oldest to newest) to import them in the write order hence we use ls -t -r /blog/*.md
  4. Firstly let's get the file name - which I have called slug.
  5. Get the filename basename "$file"."
  6. Remove the .md extension with sed 's/\.md//'
  7. Ensure the string is all lowercase tr '[:upper:]' '[:lower:]'
  8. Next we remove line 5 which is the cover option sed -i '' '5d' "$file
  9. Then at the end of line 4 we start a new line and define our slug option with sed -i '' "4a\ slug: ${slug}\ " "$file" - the new lines force the new line to be entered in the doc
  10. Similarly, but starting on line 5, we start a new line and we set publish: true with sed -i '' "9a\ publish: true\ " "$file"- again the formatting is matched in the file (new lines) Great, all done? Well we still need to upload these files to AFFiNE right?

Challenge 3: Uploading to AFFiNE

AFFiNE doesn't yet, at the time of writing, offer an API. That makes this task impossible right? Not quite... Again let's add some automation and see what we can do.

And poof! A neat little product that can automatically import your data into AFFiNE. For its initial purpose it has been setup to work with .md files but could be re-configured to import any content files - such as .doc files and .html files.

ShortCipher5/AFFiNE-Importer (github.com)

It's a much longer script so maybe another article if you are interested in this little code project specifically. And it's open-source for you to go ahead, check it out and have a play. Looking forward to your own projects and PRs are welcome.

To summarise this script uses Microsoft Playwright to automate Chromium and perform the actions of copy and paste which you'd otherwise have to do manually.

Even with this app it didn't solve everything. There was still the process of having to manually adding images (this also could be automated with more time). And in terms of the automation code itself it relies on copy-paste of data which means we have to force the code to wait sometimes to allow for the copy and paste actions to complete - with an arbitrary wait time - which no doubt can also be improved on.

The main part is that this helped significantly in lowering the manual cost and time needed to migrate these articles over.

Let's take a quick look at it in action:

hLo595V8PtMkLAz63KPasmu7KGHRtxgrwPpBCCqvLO4=

Conclusion

We hope you enjoyed this article and found some of the points and discussions useful. Be sure to check out the two GitHub projects mentioned, we welcome you to help expand, develop and even utilise these in your own projects.

Let us know if you enjoyed this article, and what type of technical (or non-technical) articles you'd like to see next.

Contact Us