+++ /dev/null
-import Dot from 'dot';
-import { cp, mkdir, readdir, readFile, writeFile } from 'fs/promises';
-import { debuglog } from 'util';
-import { join } from 'path';
-import { rmIfExists } from '../utils.js';
-
-const internals = {
- kIndexName: 'index.gmi',
- kGeminiRe: /\.gmi$/i,
-
- debuglog: debuglog('blog'),
-
- buildUrl(id, slug) {
-
- return `./${id}/${slug}`;
- },
-
- buildTitle(id, slug) {
-
- const date = new Date(Number(id));
- const shortDate = date.toISOString().split('T')[0];
- const title = slug.split('-').join(' ');
- return `${shortDate} ${title}`;
- },
-
- buildLink(id, slug) {
-
- return `=> ${internals.buildUrl(id,slug)} ${internals.buildTitle(id,slug)}`;
- }
-};
-
-export default async function (templateDirectory, source, target) {
-
- internals.debuglog(`Reading archive ${source}`);
- const postIds = (await readdir(source))
- .sort((a, b) => Number(b) - Number(a));
- const posts = [];
- for (const id of postIds) {
- const postDirectory = join(source, id);
- const slug = (await readdir(postDirectory))
- .filter((entry) => internals.kGeminiRe.test(entry))[0];
-
- posts.push({ id, slug });
- }
-
- internals.debuglog(`Read ${posts.length} posts`);
-
- internals.debuglog('Generating Archive Index');
- const indexLocation = join(templateDirectory, internals.kIndexName);
-
- internals.debuglog(`Reading ${indexLocation}`);
- const indexTemplate = await readFile(indexLocation, { encoding: 'utf8' });
-
- internals.debuglog('Writing Archive Index');
- const index = Dot.template(indexTemplate, {
- ...Dot.templateSettings,
- strip: false
- })({
- posts: posts.map((post) => internals.buildLink(post.id, post.slug)).join('\n')
- });
-
- try {
- internals.debuglog('Removing index');
- await rmIfExists(target);
- }
- finally {
- internals.debuglog('Creating index');
- await mkdir(target);
- const indexFile = join(target, internals.kIndexName);
- await writeFile(indexFile, index);
-
- internals.debuglog('Copying posts to archive');
- await cp(source, target, { recursive: true });
- }
-}
+++ /dev/null
-import { access, cp, readdir, readFile, writeFile } from 'fs/promises';
-import { exec } from 'child_process';
-import { basename, join } from 'path';
-import ParseGemini from 'gemini-to-html/parse.js';
-import RenderGemini from 'gemini-to-html/render.js';
-import { debuglog, promisify } from 'util';
-import { ensureDirectoryExists, rmIfExists } from './utils.js';
-import { kFileNotFoundError } from './constants.js';
-
-// Generators for the Blog
-
-import StaticGenerator from './generators/static.js';
-import HTMLGenerator from './generators/html.js';
-import RSSGenerator from './generators/rss.js';
-import TXTGenerator from './generators/txt.js';
-
-// Archiving Methods
-
-import GemlogArchiver from './archivers/gemlog.js';
-
-// Remote Handler
-
-import Remote from './remote.js';
-
-const internals = {
-
- // Promisified functions
- exec: promisify(exec),
-
- debuglog: debuglog('blog'),
-
- // constants
-
- kGeminiRe: /\.gmi$/i,
- kMetadataFilename: 'metadata.json',
-
- // Strings
-
- strings: {
- geminiNotFound: 'Gemini file was not found in blog directory. Please update.'
- }
-};
-
-/**
- * The Blog class is the blog generator, it's in charge of adding and
- * updating posts, and handling the publishing.
- *
- * @class Blog
- * @param {Blog.tConfiguration} config the initialization options to
- * extend the instance
- */
-export default class Blog {
-
- constructor(config) {
-
- Object.assign(this, config);
- }
-
- /**
- * Shifts the blog posts, adds the passed file to slot 0, and
- * generates files.
- *
- * @function add
- * @memberof Blog
- * @param {string} postLocation the path to the blog post file
- * @return {Promise<undefined>} empty promise, returns no value
- * @instance
- */
- async add(postLocation) {
-
- await ensureDirectoryExists(this.postsDirectory);
- try {
- await this.syncDown();
- }
- catch {}
-
- await this._shift();
- const firstDirectory = join(this.postsDirectory, '0');
- await rmIfExists(firstDirectory);
- await ensureDirectoryExists(firstDirectory);
- await this._update(postLocation);
- }
-
- /**
- * Update slot 0 with the passed gmi file, and generates files.
- *
- * @function update
- * @memberof Blog
- * @param {string} postLocation the path to the blog post file
- * @return {Promise<undefined>} empty promise, returns no value
- * @instance
- */
- async update(postLocation) {
-
- try {
- await this.syncDown();
- }
- catch {}
-
- await this._update(postLocation);
- }
-
- /**
- * Publishes the files to a static host.
- *
- * @function publish
- * @memberof Blog
- * @return {Promise<undefined>} empty promise, returns no value
- * @instance
- */
- async publish(host) {
-
- internals.debuglog(`Publishing to ${host}`);
- try {
- await internals.exec('which rsync');
- }
- catch (err) {
- console.error('Please install and configure rsync to publish.');
- }
-
- try {
- internals.debuglog(`Copying ephemeral blog from ${this.blogOutputDirectory}`);
- await internals.exec(`rsync -r ${this.blogOutputDirectory}/ ${host}`);
- }
- catch (err) {
- console.error('Failed to publish');
- console.error(err.stderr);
- }
-
- internals.debuglog('Finished publishing');
- }
-
- /**
- * Publishes the archive to a host using rsync. Currently assumes
- * gemlog archive.
- *
- * @function publishArchive
- * @memberof Blog
- * @return {Promise<undefined>} empty promise, returns no value
- * @instance
- */
- async publishArchive(host) {
-
- internals.debuglog(`Publishing archive to ${host}`);
- try {
- await internals.exec('which rsync');
- }
- catch (err) {
- console.error('Please install rsync to publish the archive.');
- }
-
- try {
- internals.debuglog(`Copying archive from ${this.archiveOutputDirectory}`);
- await internals.exec(`rsync -r ${this.archiveOutputDirectory}/ ${host}`);
- }
- catch (err) {
- console.error('Failed to publish archive');
- console.error(err.stderr);
- }
-
- internals.debuglog('Finished publishing');
- }
-
- /**
- * Adds a remote
- *
- * @function addRemote
- * @memberof Blog
- * @return {Promise<undefined>} empty promise, returns no value
- * @instance
- */
- async addRemote(remote) {
-
- await ensureDirectoryExists(this.configDirectory);
- await Remote.add(this.remoteConfig, remote);
- }
-
- /**
- * Removes a remote
- *
- * @function removeRemote
- * @memberof Blog
- * @return {Promise<undefined>} empty promise, returns no value
- * @instance
- */
- async removeRemote() {
-
- await Remote.remove(this.remoteConfig);
- }
-
-
- /**
- * Pulls the posts and archive from the remote
- *
- * @function syncDown
- * @memberof Blog
- * @return {Promise<undefined>} empty promise, returns no value
- * @instance
- */
- async syncDown() {
-
- internals.debuglog('Pulling remote state');
- await ensureDirectoryExists(this.dataDirectory);
- await Remote.syncDown(this.remoteConfig, this.dataDirectory);
- internals.debuglog('Pulled remote state');
- }
-
- /**
- * Pushes the posts and archive to the remote
- *
- * @function syncUp
- * @memberof Blog
- * @return {Promise<undefined>} empty promise, returns no value
- * @instance
- */
- async syncUp() {
-
- internals.debuglog('Pushing remote state');
- await ensureDirectoryExists(this.dataDirectory);
- await Remote.syncUp(this.remoteConfig, this.dataDirectory);
- internals.debuglog('Pushed remote state');
- }
-
- // Adds the passed path to slot 0, and generates files.
-
- async _update(postLocation) {
-
- const metadata = await this._getMetadata();
- await ensureDirectoryExists(this.postsDirectory);
- await this._copyPost(postLocation);
- await this._writeMetadata(metadata);
-
- await this._archive(postLocation);
-
- await this.generate();
- try {
- await this.syncUp();
- }
- catch {}
- }
-
-
- // Parses Gemini for each page, copies assets and generates index.
-
- async generate() {
-
- internals.debuglog('Generating output');
-
- const posts = await this._readPosts();
-
- // Start from a clean slate.
- await rmIfExists(this.blogOutputDirectory);
- await ensureDirectoryExists(this.blogOutputDirectory);
-
- // Run each generator
- await StaticGenerator(this.staticDirectory, this.blogOutputDirectory, posts);
- await HTMLGenerator(await this._templateDirectoryFor('index.html'), this.blogOutputDirectory, posts);
- await RSSGenerator(await this._templateDirectoryFor('feed.xml'), this.blogOutputDirectory, posts);
- await TXTGenerator(await this._templateDirectoryFor('index.txt'), this.blogOutputDirectory, posts);
-
- // Start from a clean slate.
- await rmIfExists(this.archiveOutputDirectory);
- await ensureDirectoryExists(this.archiveOutputDirectory);
- await ensureDirectoryExists(this.archiveDirectory);
-
- // Run each archiver
- await GemlogArchiver(await this._templateDirectoryFor('index.gmi'), this.archiveDirectory, this.archiveOutputDirectory);
- // TODO: GopherArchiver
- }
-
- // Reads the posts into an array
-
- async _readPosts() {
-
- internals.debuglog('Reading posts');
- const posts = [];
-
- for (let i = 0; i < this.maxPosts; ++i) {
- try {
- posts.push(await this._readPost(i));
- }
- catch (error) {
- if (error.code === kFileNotFoundError) {
- internals.debuglog(`Skipping ${i}`);
- continue;
- }
-
- throw error;
- }
- }
-
- return posts;
- }
-
- // Reads an individual post
-
- async _readPost(index = 0) {
-
- const postSourcePath = join(this.postsDirectory, `${index}`);
-
- internals.debuglog(`Reading ${postSourcePath}`);
-
- await access(postSourcePath);
-
- const metadata = await this._getMetadata(index);
-
- const postContentPath = await this._findBlogContent(postSourcePath);
- internals.debuglog(`Reading ${postContentPath}`);
- const postContent = await readFile(postContentPath, { encoding: 'utf8' });
-
- internals.debuglog('Parsing Gemini');
- return {
- ...metadata,
- location: postSourcePath,
- index,
- html: RenderGemini(ParseGemini(postContent)),
- raw: postContent
- };
- }
-
- // Shift the posts, delete any remainder.
-
- async _shift() {
-
-
- for (let i = this.maxPosts - 1; i >= 1; --i) {
- const targetPath = join(this.postsDirectory, `${i}`);
- const sourcePath = join(this.postsDirectory, `${i - 1}`);
-
- try {
- internals.debuglog(`Archiving ${targetPath}`);
- await rmIfExists(targetPath);
- await access(sourcePath); // check the source path
-
- internals.debuglog(`Shifting blog post ${sourcePath} to ${targetPath}`);
- await cp(sourcePath, targetPath, { recursive: true });
- }
- catch (error) {
- if (error.code === kFileNotFoundError) {
- internals.debuglog(`Skipping ${sourcePath}: Does not exist.`);
- continue;
- }
-
- throw error;
- }
- }
- }
-
- // Moves older posts to the archive
-
- async _archive() {
-
- internals.debuglog('Archiving post');
- const post = await this._readPost(0);
- await ensureDirectoryExists(this.archiveDirectory);
-
- const targetPath = join(this.archiveDirectory, post.id);
-
- internals.debuglog(`Removing ${targetPath}`);
- await rmIfExists(targetPath);
- internals.debuglog(`Adding ${post.location} to ${targetPath}`);
- await ensureDirectoryExists(targetPath);
- await cp(post.location, targetPath, { recursive: true });
- internals.debuglog(`Added ${post.location} to ${targetPath}`);
- }
-
- // Attempts to read existing metadata. Otherwise generates new set.
-
- async _getMetadata(index = 0) {
-
- const metadataTarget = join(this.postsDirectory, String(index), internals.kMetadataFilename);
-
- try {
- internals.debuglog(`Looking for metadata at ${metadataTarget}`);
- return JSON.parse(await readFile(metadataTarget, { encoding: 'utf8' }));
- }
- catch (e) {
- internals.debuglog(`Metadata not found or unreadable. Generating new set.`);
- const createdOn = Date.now();
- const metadata = {
- id: String(createdOn),
- createdOn
- };
-
- return metadata;
- }
- }
-
- // Writes metadata. Assumes post 0 since it only gets written
- // on create
-
- async _writeMetadata(metadata) {
-
- const metadataTarget = join(this.postsDirectory, '0', internals.kMetadataFilename);
- internals.debuglog(`Writing ${metadataTarget}`);
- await writeFile(metadataTarget, JSON.stringify(metadata, null, 2));
- }
-
- // Copies a post file to the latest slot.
-
- async _copyPost(postLocation) {
-
- internals.debuglog(`Copying ${postLocation}`);
- const targetPath = join(this.postsDirectory, '0');
- const postName = basename(postLocation);
- const targetPost = join(targetPath, postName);
-
- await rmIfExists(targetPath);
- await ensureDirectoryExists(targetPath);
- await cp(postLocation, targetPost, { recursive: true });
- internals.debuglog(`Added ${postLocation} to ${targetPath}`);
- }
-
- // Looks for a `.gmi` file in the blog directory, and returns the path
-
- async _findBlogContent(directory) {
-
- const entries = await readdir(directory);
-
- const geminiEntries = entries
- .filter((entry) => internals.kGeminiRe.test(entry))
- .map((entry) => join(directory, entry));
-
- if (geminiEntries.length > 0) {
- internals.debuglog(`Found gemini file: ${geminiEntries[0]}`);
- return geminiEntries[0];
- }
-
- throw new Error(internals.strings.geminiNotFound);
- }
-
- // Gets the template directory for a given template.
- async _templateDirectoryFor(template) {
-
- try {
- await access(join(this.templatesDirectory, template));
- return this.templatesDirectory;
- }
- catch (error) {
- if (error.code === kFileNotFoundError) {
- internals.debuglog(`No custom template for ${template}`);
- return this.defaultTemplatesDirectory;
- }
-
- throw error;
- }
- }
-}
+++ /dev/null
-import Dot from 'dot';
-import { readFile, writeFile } from 'fs/promises';
-import { join } from 'path';
-import { debuglog } from 'util';
-
-const internals = {
- debuglog: debuglog('blog'),
-
- kIndexName: 'index.html'
-};
-
-/**
- * Generates the blog index page
- *
- * @name HTMLGenerator
- * @param {string} source the source directory
- * @param {string} target the target directory
- * @param {Array.<Blog.tPost>} posts the list of posts
- */
-export default async function HTMLGenerator(source, target, posts) {
-
- internals.debuglog('Generating HTML');
- const indexTarget = join(target, internals.kIndexName);
- const indexLocation = join(source, internals.kIndexName);
-
- internals.debuglog(`Reading ${indexLocation}`);
- const indexTemplate = await readFile(indexLocation, { encoding: 'utf8' });
-
- internals.debuglog('Writing HTML');
- const indexHtml = Dot.template(indexTemplate, {
- ...Dot.templateSettings,
- strip: false
- })({ posts });
- await writeFile(indexTarget, indexHtml);
-}
+++ /dev/null
-import Dot from 'dot';
-import { encodeXML } from 'entities';
-import { readFile, writeFile } from 'fs/promises';
-import { join } from 'path';
-import { debuglog } from 'util';
-
-const internals = {
- debuglog: debuglog('blog'),
-
- kFeedName: 'feed.xml',
-
- extractTitle(postText) {
-
- return postText.trim()
- .split('\n')[0]
- .replace('#', '')
- .replace(/&/g, '&')
- .trim();
- }
-};
-
-/**
- * Generates an RSS feed XML file
- *
- * @name RSSGenerator
- * @param {string} source the source directory
- * @param {string} target the target directory
- * @param {Array.<Blog.tPost>} posts the list of posts
- */
-export default async function RSSGenerator(source, target, posts) {
-
- internals.debuglog('Generating RSS');
- const feedTarget = join(target, internals.kFeedName);
- const feedLocation = join(source, internals.kFeedName);
-
- internals.debuglog(`Reading ${feedLocation}`);
- const feedTemplate = await readFile(feedLocation, { encoding: 'utf8' });
-
- internals.debuglog('Writing RSS');
- posts = posts.map((post) => ({
- ...post,
- createdOn: (new Date(post.createdOn)).toUTCString(),
- title: internals.extractTitle(post.raw),
- html: encodeXML(post.html)
- }));
- const feedXml = Dot.template(feedTemplate, {
- ...Dot.templateSettings,
- strip: false
- })({ posts });
- await writeFile(feedTarget, feedXml);
-}
+++ /dev/null
-import { access, cp, readdir } from 'fs/promises';
-import { constants } from 'fs';
-import { join } from 'path';
-import { debuglog } from 'util';
-import { kFileNotFoundError } from '../constants.js';
-
-const internals = {
- debuglog: debuglog('blog'),
- kAssetsDirectoryName: 'assets'
-};
-
-/**
- * Generates the static assets required for the blog
- *
- * @name StaticGenerator
- * @param {string} source the source directory
- * @param {string} target the target directory
- * @param {Array.<Blog.tPost>} posts the list of posts
- */
-export default async function StaticGenerator(source, target, _) {
-
- try {
- await access(source, constants.R_OK);
-
- const entries = await readdir(source, { withFileTypes: true });
- for (const entry of entries) {
- const sourceAsset = join(source, entry.name);
- const targetAsset = join(target, entry.name);
-
- internals.debuglog(`Copying ${sourceAsset} to ${targetAsset}`);
-
- if (entry.isDirectory()) {
- await cp(sourceAsset, targetAsset, { recursive: true });
- }
- else {
- await cp(sourceAsset, targetAsset);
- }
- }
- }
- catch (error) {
- if (error.code === kFileNotFoundError) {
- internals.debuglog(`No static directory found in ${source}`);
- return;
- }
-
- throw error;
- }
-}
+++ /dev/null
-import Dot from 'dot';
-import { readFile, writeFile } from 'fs/promises';
-import { join } from 'path';
-import { debuglog } from 'util';
-
-const internals = {
- debuglog: debuglog('blog'),
-
- kTextName: 'index.txt'
-};
-
-/**
- * Generates a TXT version of the blog
- *
- * @name TXTGenerator
- * @param {string} source the source directory
- * @param {string} target the target directory
- * @param {Array.<Blog.tPost>} posts the list of posts
- */
-export default async function TXTGenerator(source, target, posts) {
-
- internals.debuglog('Generating TXT');
- const textTarget = join(target, internals.kTextName);
- const textLocation = join(source, internals.kTextName);
-
- internals.debuglog(`Reading ${textLocation}`);
- const textTemplate = await readFile(textLocation, { encoding: 'utf8' });
-
- internals.debuglog('Writing TXT');
- const text = Dot.template(textTemplate, {
- ...Dot.templateSettings,
- strip: false
- })({ posts });
- await writeFile(textTarget, text);
-}
+++ /dev/null
-import { readFile, writeFile } from 'fs/promises';
-import { rmIfExists } from './utils.js';
-
-import GitStrategy from './remotes/git.js';
-
-const internals = {
- strings: {
- configurationNotFound: 'Remote configuration not set, consult help for more info.'
- },
- strategies: [
- GitStrategy
- ]
-};
-
-export default {
- async add(remoteConfig, remote) {
-
- await writeFile(remoteConfig, remote);
- },
-
- async remove(remoteConfig) {
-
- await rmIfExists(remoteConfig);
- },
-
- async syncUp(remoteConfig, blogDirectory) {
-
- await this._executeMethodOnStrategy(remoteConfig, 'syncUp', blogDirectory);
- },
-
- async syncDown(remoteConfig, blogDirectory) {
-
- await this._executeMethodOnStrategy(remoteConfig, 'syncDown', blogDirectory);
- },
-
- async _executeMethodOnStrategy(remoteConfig, method, blogDirectory) {
-
- const remote = await this._ensureConfiguration(remoteConfig);
-
- for (const strategy of internals.strategies) {
- if (strategy.canHandle(remote)) {
- await strategy[method](remote, blogDirectory);
- }
- }
- },
-
- async _ensureConfiguration(remoteConfig) {
-
- try {
- const configuration = await readFile(remoteConfig, { encoding: 'utf8' });
- return configuration;
- }
- catch {
- throw new Error(internals.strings.configurationNotFound);
- }
- }
-};
+++ /dev/null
-import { exec } from 'child_process';
-import { promisify } from 'util';
-
-const internals = {
- // Promisified functions
- exec: promisify(exec)
-};
-
-export default {
- canHandle() {
-
- // For the future: actually check if it's a valid git url
- return true;
- },
-
- async syncUp(remote, blogDirectory) {
-
- await internals.exec(`cd ${blogDirectory} && git init -b main`);
- await internals.exec(`cd ${blogDirectory} && git add .`);
- await internals.exec(`cd ${blogDirectory} && git commit --allow-empty -m blog-sync-up-${Date.now()}`);
- await internals.exec(`cd ${blogDirectory} && git push ${remote} main --force`);
- },
-
- async syncDown(remote, blogDirectory) {
-
- await internals.exec(`cd ${blogDirectory} && git init -b main`);
- try {
- await internals.exec(`cd ${blogDirectory} && git checkout .`);
- }
- catch {}
-
- await internals.exec(`cd ${blogDirectory} && git clean . -f`);
- await internals.exec(`cd ${blogDirectory} && git pull ${remote} main`);
- }
-};
+++ /dev/null
-import { access, constants, mkdir, rm } from 'fs/promises';
-import { kFileNotFoundError } from './constants.js';
-
-// File system utilities
-
-export const rmIfExists = async function rmIfExists(location) {
-
- try {
- await access(location, constants.F_OK);
- await rm(location, { recursive: true });
- }
- catch (error) {
- if (error.code === kFileNotFoundError) {
- return;
- }
-
- throw error;
- }
-};
-
-export const ensureDirectoryExists = async function ensureDirectoryExists(directory) {
-
- try {
- await access(directory);
- }
- catch (error) {
- if (error.code === kFileNotFoundError) {
- await mkdir(directory, { recursive: true });
- return;
- }
-
- throw error;
- }
-};
-use std::fs::{create_dir_all, write};
use std::io::Result;
use crate::configuration::Configuration;
+use crate::remote::add;
pub struct AddRemote;
}
fn execute(&self, input: Option<&String>, configuration: &Configuration, _: &String) -> Result<()> {
- create_dir_all(&configuration.config_directory)?;
let input = input.expect("You must provide a location for the remote.");
- write(&configuration.remote_config, input)?;
- return Ok(())
+ add(&configuration.config_directory, &configuration.remote_config, input)
}
fn after_dependencies(&self) -> Vec<Box<dyn super::Command>> {
-use std::fs::remove_file;
use std::io::Result;
use crate::configuration::Configuration;
+use crate::remote::remove;
pub struct RemoveRemote;
}
fn execute(&self, _: Option<&String>, configuration: &Configuration, _: &String) -> Result<()> {
- if configuration.remote_config.exists() {
- remove_file(&configuration.remote_config)?;
- }
- return Ok(())
+ remove(&configuration.remote_config)
}
fn after_dependencies(&self) -> Vec<Box<dyn super::Command>> {
-use std::fs::create_dir_all;
-use std::io::{Result, Error};
+use std::io::{Result};
use crate::configuration::Configuration;
+use crate::remote::sync_down;
pub struct SyncDown;
}
fn execute(&self, _: Option<&String>, configuration: &Configuration, command: &String) -> Result<()> {
- match create_dir_all(&configuration.data_directory) {
- Ok(_) => {
- // We only care to show these warnings if this is the primary command.
- if command == self.command() {
- println!("WARNING: Sync Down Not yet implemented");
- }
- return Ok(())
- },
- Err(e) => Err(Error::new(e.kind(), format!("Could not create data directory")))
+ if command == self.command() {
+ return sync_down(&configuration.data_directory, &configuration.remote_config);
}
+ return Ok(())
}
fn after_dependencies(&self) -> Vec<Box<dyn super::Command>> {
use std::io::Result;
use crate::configuration::Configuration;
+use crate::remote::sync_up;
pub struct SyncUp;
vec![]
}
- fn execute(&self, input: Option<&String>, _: &Configuration, _: &String) -> Result<()> {
- println!("Sync Up: {:?}!", input);
- return Ok(())
+ fn execute(&self, _: Option<&String>, configuration: &Configuration, _: &String) -> Result<()> {
+ sync_up(&configuration.data_directory, &configuration.remote_config)
}
fn after_dependencies(&self) -> Vec<Box<dyn super::Command>> {
mod post;
mod template;
mod utils;
+mod remote;
use std::iter::once;
use std::env::args;
--- /dev/null
+use std::io::Result;
+use std::path::PathBuf;
+use std::process::Command;
+use std::time::{SystemTime, UNIX_EPOCH};
+
+pub struct Git;
+
+impl Git {
+ pub fn new() -> Self {
+ Git
+ }
+}
+
+impl super::Remote for Git {
+ fn can_handle(&self, _: &str) -> bool {
+ // If we ever implement another remote we will want a check strategy
+ true
+ }
+
+ fn sync_up(&self, remote: &str, directory: &PathBuf) -> Result<()> {
+ let timestamp = SystemTime::now().duration_since(UNIX_EPOCH)
+ .expect("Invalid time")
+ .as_millis();
+
+ let commands = vec![
+ format!("cd {} && git init -b main", directory.display()),
+ format!("cd {} && git add .", directory.display()),
+ format!("cd {} && git commit --allow-empty -m blog-sync-up-{}", directory.display(), timestamp),
+ format!("cd {} && git push {} main --force", directory.display(), remote),
+ ];
+
+ for command in commands {
+ Command::new("sh")
+ .arg("-c")
+ .arg(&command)
+ .status()
+ .expect("Failed while performing sync up with git");
+ }
+
+ Ok(())
+ }
+
+ fn sync_down(&self, remote: &str, directory: &PathBuf) -> Result<()> {
+ let commands = vec![
+ format!("cd {} && git init -b main", directory.display()),
+ format!("cd {} && git checkout .", directory.display()),
+ format!("cd {} && git clean . -f", directory.display()),
+ format!("cd {} && git pull {} main", directory.display(), remote),
+ ];
+
+ for command in commands {
+ Command::new("sh")
+ .arg("-c")
+ .arg(&command)
+ .status()
+ .expect("Failed while performing sync down with git");
+ }
+ Ok(())
+ }
+}
+mod git;
+
+use std::fs::{create_dir_all, remove_file, write, File};
+use std::io::{Error, ErrorKind::Other, Read, Result};
+use std::path::PathBuf;
+
+use git::Git;
+
+pub trait Remote {
+ fn can_handle(&self, remote: &str) -> bool;
+ fn sync_up(&self, remote: &str, directory: &PathBuf) -> Result<()>;
+ fn sync_down(&self, remote: &str, directory: &PathBuf) -> Result<()>;
+}
+
+pub fn add(config_directory: &PathBuf, remote_config: &PathBuf, remote: &String) -> Result<()> {
+ create_dir_all(config_directory)?;
+ write(remote_config, remote)?;
+ Ok(())
+}
+
+pub fn remove(remote_config: &PathBuf) -> Result<()> {
+ if remote_config.exists() {
+ remove_file(remote_config)?
+ }
+ Ok(())
+}
+
+pub fn sync_up(data_directory: &PathBuf, remote_config: &PathBuf) -> Result<()> {
+ let remote_address = read_remote(remote_config)
+ .expect("No remote is configured");
+ create_dir_all(data_directory)?;
+ let remotes = available_remotes();
+ for remote in remotes {
+ if remote.can_handle(&remote_address) {
+ return remote.sync_up(&remote_address, data_directory);
+ }
+ }
+ Err(Error::new(Other, "No valid strategies found for your configured remote."))
+}
+
+pub fn sync_down(data_directory: &PathBuf, remote_config: &PathBuf) -> Result<()> {
+ let remote_address = read_remote(remote_config)
+ .expect("No remote is configured");
+ create_dir_all(data_directory)?;
+ let remotes = available_remotes();
+ for remote in remotes {
+ if remote.can_handle(&remote_address) {
+ return remote.sync_down(&remote_address, data_directory);
+ }
+ }
+ Err(Error::new(Other, "No valid strategies found for your configured remote."))
+}
+
+fn available_remotes() -> Vec<Box<dyn Remote>> {
+ vec![
+ Box::new(Git::new())
+ ]
+}
+
+fn read_remote(file_path: &PathBuf) -> Option<String> {
+ let mut file = File::open(file_path).ok()?;
+ let mut contents = String::new();
+ file.read_to_string(&mut contents).ok()?;
+ Some(contents)
+}