Skip to content
This repository has been archived by the owner on Feb 8, 2023. It is now read-only.
/ website-honestly Public archive

🦄 The Red Badger website. Honestly.

Notifications You must be signed in to change notification settings

redbadger/website-honestly

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Website. Honestly. 🦄

CircleCI Maintainability Test Coverage

Red Badger Website Episode 6: Return of the Jedi.

Usage

Codespaces

Github Codespaces is enabled for this repository to get your development environment up and running as quickly as possible. To get started,

  • Click the Code drop-down menu, and then in the Codespaces tab, click New codespace.
  • A pop up should appear to select the size of the VM, select a 4 core VM or higher.
  • Then select whether you would like to use VSCode in the browser (preferably a chromium based browser) or the native app on your local machine.
  • Follow instructions on how to run the website locally.

Note: All the necessary credentials for local development should already be embedded within the codespaces environment on the startup automatically.

Local Setup

You will require the correct version of Node.js installed - check it in the .node-version file. On OSX you can use nvm and avn to install and manage versions (avn will read the .node-version file).

First set up your dev environment and install global dependencies:

# Install the package manager
npm install --global yarn

# Clone the repo
git clone https://github.com/redbadger/website-honestly.git

# Install the deps
yarn

# Set up the environment variables. Follow the instructions of this command:
make keyrings

# Go to https://github.com/redbadger/blackbox-secrets/blob/master/README.md
# Follow instructions to
# - "Getting Blackbox"
# - "Get access to existing blackbox secrets"

# Make sure you have git-secrets installed by running
brew install git-secrets

# Set up git secrets to scan for secrets pre-commit. Follow the instructions of this command:
make scanner

# Go to https://github.com/redbadger/secrets-scanner/blob/master/README.md
# Follow instructions to
# - "Getting Git Secrets"
# - "Setting up to scan"

# Fetch and stash dynamic data for development. You might need to do this
# every time you want to get updated data from staging for local development.
make fetch

Running the website

Now start the website:

make                   # Print the help
make dev               # Run the dev server (url - localhost:8080)
make clean dev-static  # Compile the site locally (url - localhost:8000)

# Deploy to dev lambda environment
make clean services-deploy

# Invoke dev lambdas
make services-invoke-publish

Technical Overview

This site is a static site hosted on AWS S3 behind a proxy that provides HTTPS. You can find the proxy here: https://github.com/redbadger/website-next-proxy

The code used to build the site's HTML and CSS can be found in the site/ directory. It's mostly React components and CSS modules.

The same React app in the site/ directory is used on the front end for interactive features such as the "Sign Up" form. The entry point for the front end Javascript can be found in the client/ directory.

The content for the site is (will be) pulled from the Badger Brain GraphQL service before compilation. The code for fetching and preprocessing this data can be found in the state/ directory.

Generation of the site is done on AWS Lambda, so there are no production servers to look after. This lambda function is to be called any time there is a content update on the Prismic CMS. This and other lambda functions can be found in the services/ directory.

The lambda functions are managed and deployed using Serverless framework, and AWS CloudFormation is used with Serverless to provision the rest of the new site infrastructure.

The dev/ directory contains two entry points to the application. The first is browser-app/, which is the site mounted as a regular React app in the browser.

Test data is loaded from assets/state.json during development and can be re-fetched by running make fetch.

When make dev is run this app will be served with webpack-dev-server, allowing for a fast feedback loop development. This is likely to be where you will spend most your time while working on this application.

The other entry point in dev/ is static/ which is similar to the site generator on lambda, but it runs locally and writes the pages to the local filesystem instead of to AWS S3. This is useful for checking functionality that might not work with the front end dev app.

Deployment

We use CircleCI for running tests, building the app, and deploying the app.

When a branch is pushed to CircleCI will compile the app, copy the CSS and client-side JS to a folder in the staging S3 bucket and deploy the lambda to a staging environment. The staging lambda is invoked, writing the HTML pages to the same folder in the staging bucket. Lastly a "View deployment" link is added to any pull request open on GitHub for that branch so it can be tested and reviewed.

When the master branch is updated, CircleCI will do the same deploy + invoke process, writing to the root of the staging bucket. The staging site always reflects the current state of master.

There is only one staging lambda function instance, so there is no guarantee that the correct version will be currently deployed when viewing a branch preview. This matters if you are testing the "sign up" form, for example.

Deploys to the live prod environment are done by starting a parameterised build on CircleCI that does the deploy + invoke process for the production lambda service and S3 bucket. This build can be started via @badgerbot on Slack.

See bin/ci-deploy.sh for more detail on deployment.

Monitoring

We use AWS CloudWatch for monitoring our staging and prod lambdas; all alarms should be picked up by bugsnag and sent to the slack channel #internal-projects.

Assets

SVGs can be just inlined as JSX but they should be optimised first, SVGOMG is a really handy tool for this.

Responsive loader can be used to create an srcSet attribute with multiple sizes of an image (so that appopriately sized images can be served to mobile, tablet and desktop devices for improved performance). See the plugin’s options for more details. Example usage:

import badgers from './badgers.jpeg?min=320&max=1600&steps=6';

const image = (
  <img
    src={badgers}
    srcSet={badgers.srcSet}
    sizes="(min-width: 1200px) 1200px, 100vw"
    alt="Badgers working in a team"
  />
);

Images are automatically optimised by imagemin-webpack-plugin in production.

Tracking

Analytics are tracked through google tag manager. This requires a gtm or container id which should be stored as an environment variable.

We use TagManager supplied by the react-gtm-module package. It is initialised in site/client/index.js with the primary data layer which provides the current domain.

The tag manager is setup for each page during the makeApp step

The rational for using google tag manager is that individual events can be setup and tracked without dev support.

Blog

The blog /blog is hosted on hubspot and can be managed there.