diff --git a/.github/CODE_OF_CONDUCT.md b/.github/CODE_OF_CONDUCT.md
deleted file mode 100644
index 1be855e3fa..0000000000
--- a/.github/CODE_OF_CONDUCT.md
+++ /dev/null
@@ -1,76 +0,0 @@
-# Contributor Covenant Code of Conduct
-
-## Our Pledge
-
-In the interest of fostering an open and welcoming environment, we as
-contributors and maintainers pledge to making participation in our project and
-our community a harassment-free experience for everyone, regardless of age, body
-size, disability, ethnicity, sex characteristics, gender identity and expression,
-level of experience, education, socio-economic status, nationality, personal
-appearance, race, religion, or sexual identity and orientation.
-
-## Our Standards
-
-Examples of behavior that contributes to creating a positive environment
-include:
-
-* Using welcoming and inclusive language
-* Being respectful of differing viewpoints and experiences
-* Gracefully accepting constructive criticism
-* Focusing on what is best for the community
-* Showing empathy towards other community members
-
-Examples of unacceptable behavior by participants include:
-
-* The use of sexualized language or imagery and unwelcome sexual attention or
- advances
-* Trolling, insulting/derogatory comments, and personal or political attacks
-* Public or private harassment
-* Publishing others' private information, such as a physical or electronic
- address, without explicit permission
-* Other conduct which could reasonably be considered inappropriate in a
- professional setting
-
-## Our Responsibilities
-
-Project maintainers are responsible for clarifying the standards of acceptable
-behavior and are expected to take appropriate and fair corrective action in
-response to any instances of unacceptable behavior.
-
-Project maintainers have the right and responsibility to remove, edit, or
-reject comments, commits, code, wiki edits, issues, and other contributions
-that are not aligned to this Code of Conduct, or to ban temporarily or
-permanently any contributor for other behaviors that they deem inappropriate,
-threatening, offensive, or harmful.
-
-## Scope
-
-This Code of Conduct applies both within project spaces and in public spaces
-when an individual is representing the project or its community. Examples of
-representing a project or community include using an official project e-mail
-address, posting via an official social media account, or acting as an appointed
-representative at an online or offline event. Representation of a project may be
-further defined and clarified by project maintainers.
-
-## Enforcement
-
-Instances of abusive, harassing, or otherwise unacceptable behavior may be
-reported by contacting the project team at community@budibase.com. All
-complaints will be reviewed and investigated and will result in a response that
-is deemed necessary and appropriate to the circumstances. The project team is
-obligated to maintain confidentiality with regard to the reporter of an incident.
-Further details of specific enforcement policies may be posted separately.
-
-Project maintainers who do not follow or enforce the Code of Conduct in good
-faith may face temporary or permanent repercussions as determined by other
-members of the project's leadership.
-
-## Attribution
-
-This Code of Conduct is adapted from the [Contributor Covenant][homepage], version 1.4,
-available at https://www.contributor-covenant.org/version/1/4/code-of-conduct.html
-
-[homepage]: https://www.contributor-covenant.org
-
-For answers to common questions about this code of conduct, see
-https://www.contributor-covenant.org/faq
diff --git a/.github/CODE_OF_CONDUCT.md b/.github/CODE_OF_CONDUCT.md
new file mode 120000
index 0000000000..d28e66efdc
--- /dev/null
+++ b/.github/CODE_OF_CONDUCT.md
@@ -0,0 +1 @@
+../docs/CODE_OF_CONDUCT.md
\ No newline at end of file
diff --git a/.github/CONTRIBUTING.md b/.github/CONTRIBUTING.md
deleted file mode 100644
index bd21123709..0000000000
--- a/.github/CONTRIBUTING.md
+++ /dev/null
@@ -1,208 +0,0 @@
-# Contributing
-
-From opening a bug report to creating a pull request: every contribution is appreciated and welcome. If you're planning to implement a new feature or change the api please create an issue first. This way we can ensure that your precious work is not in vain.
-
-## Not Sure Where to Start?
-
-Budibase is a low-code web application builder that creates svelte based web applications.
-
-Budibase is a monorepo managed by [lerna](https://github.com/lerna/lerna). Lerna manages the building and publishing of the budibase packages. At a high level, here are the packages that make up budibase.
-
-- **packages/builder** - contains code for the budibase builder client side svelte application.
-
-- **packages/client** - A module that runs in the browser responsible for reading JSON definition and creating living, breathing web apps from it.
-
-- **packages/server** - The budibase server. This [Koa](https://koajs.com/) app is responsible for serving the JS for the builder and budibase apps, as well as providing the API for interaction with the database and file system.
-
-- **packages/worker** - This [Koa](https://koajs.com/) app is responsible for providing global apis for managing your budibase installation. Authentication, Users, Email, Org and Auth configs are all provided by the worker.
-
-## Contributor License Agreement (CLA)
-
-In order to accept your pull request, we need you to submit a CLA. You only need to do this once. If you are submitting a pull request for the first time, just submit a Pull Request and our CLA Bot will give you instructions on how to sign the CLA before merging your Pull Request.
-
-All contributors must sign an [Individual Contributor License Agreement](https://github.com/budibase/budibase/blob/next/.github/cla/individual-cla.md).
-
-If contributing on behalf of your company, your company must sign a [Corporate Contributor License Agreement](https://github.com/budibase/budibase/blob/next/.github/cla/corporate-cla.md). If so, please contact us via community@budibase.com.
-
-## Glossary of Terms
-
-To understand the budibase API, it can be helpful to understand the top level entities that make up Budibase.
-
-### Client
-
-A client represents a single budibase customer. Each budibase client will have 1 or more budibase servers. Every client is assigned a unique ID.
-
-### App
-
-A client can have one or more budibase applications. Budibase applications would be things like "Developer Inventory Management" or "Goat Herder CRM". Think of a budibase application as a tree.
-
-### Database
-
-An App can have one or more databases. Keeping with our [dendrology](https://en.wikipedia.org/wiki/Dendrology) analogy - think of an database as a branch on the tree. Databases are used to keep data separate for different instances of your app. For example, if you had a CRM app, you may create a database for your US office, and a database for your Australian office. Databases allow us to support [multitenancy](https://www.gartner.com/en/information-technology/glossary/multitenancy) in budibase applications.
-
-### Table
-
-Tables in budibase are almost akin to tables in relational databases. A table may be a "Car" or an "Employee". They are the main building blocks for the creation and management of backend data in budibase.
-
-### View
-
-A View is an advanced feature in budibase that allows you to write a custom query using [MapReduce](https://pouchdb.com/guides/queries.html) queries. Views enable powerful query functionality and calculations, allowing you to do more with your data.
-
-### Page
-
-A page in budibase is actually a single, self contained svelte web app. There are only 2 pages in budibase. The **login** page and the **main** page.
-
-### Screen
-
-A screen is a component within a single page. Generally, screens represent client side routes, and can be switched without refreshing the page.
-
-### Component
-
-A component is the basic frontend building block of a budibase app.
-
-### Component Library
-
-Component libraries are collections of components as well as the definition of their props contained in a file called `components.json`.
-
-## Contributing to Budibase
-
-* Please maintain the existing code style.
-
-* Please try to keep your commits small and focused.
-
-* Please write tests.
-
-* If the project diverges from your branch, please rebase instead of merging. This makes the commit graph easier to read.
-
-* Once your work is completed, please raise a PR against the `develop` branch with some information about what has changed and why.
-
-### Getting Started For Contributors
-#### 1. Prerequisites
-
-NodeJS Version `14.x.x`
-
-*yarn -* `npm install -g yarn`
-
-*jest* - `npm install -g jest`
-
-#### 2. Clone this repository
-
-`git clone https://github.com/Budibase/budibase.git`
-
-then `cd ` into your local copy.
-
-#### 3. Install and Build
-
-| **NOTE**: On Windows, all yarn commands must be executed on a bash shell (e.g. git bash)
-
-To develop the Budibase platform you'll need [Docker](https://www.docker.com/) and [Docker Compose](https://docs.docker.com/compose/) installed.
-
-##### Quick method
-
-`yarn setup` will check that all necessary components are installed and setup the repo for usage.
-
-##### Manual method
-
-The following commands can be executed to manually get Budibase up and running (assuming Docker/Docker Compose has been installed).
-
-`yarn` to install project dependencies
-
-`yarn bootstrap` will install all budibase modules and symlink them together using lerna.
-
-`yarn build` will build all budibase packages.
-
-#### 4. Running
-
-To run the budibase server and builder in dev mode (i.e. with live reloading):
-
-1. Open a new console
-2. `yarn dev` (from root)
-3. Access the builder on http://localhost:10000/builder
-
-This will enable watch mode for both the builder app, server, client library and any component libraries.
-
-#### 5. Debugging using VS Code
-
-To debug the budibase server and worker a VS Code launch configuration has been provided.
-
-Visit the debug window and select `Budibase Server` or `Budibase Worker` to debug the respective component.
-Alternatively to start both components simultaneously select `Start Budibase`.
-
-In addition to the above, the remaining budibase components may be ran in dev mode using: `yarn dev:noserver`.
-
-#### 6. Cleanup
-
-If you wish to delete all the apps created in development and reset the environment then run the following:
-
-1. `yarn nuke:docker` will wipe all the Budibase services
-2. `yarn dev` will restart all the services
-
-### Backend
-
-For the backend we run [Redis](https://redis.io/), [CouchDB](https://couchdb.apache.org/), [MinIO](https://min.io/) and [NGINX](https://www.nginx.com/) in Docker compose. This means that to develop Budibase you will need Docker and Docker compose installed. The backend services are then ran separately as Node services with nodemon so that they can be debugged outside of Docker.
-
-### Data Storage
-
-When you are running locally, budibase stores data on disk using docker volumes. The volumes and the types of data associated with each are:
-
-- `redis_data`
- - Sessions, email tokens
-- `couchdb3_data`
- - Global and app databases
-- `minio_data`
- - App manifest, budibase client, static assets
-
-### Devlopment Modes
-
-A combination of environment variables controls the mode that budibase runs in.
-Yarn commands can be used to mimic the different modes that budibase can be ran in
-
-#### Self Hosted
-The default mode. A single tenant installation with no usage restrictions.
-
-To enable this mode, use:
-```
-yarn mode:self
-```
-
-#### Cloud
-The cloud mode, with account portal turned off.
-
-To enable this mode, use:
-```
-yarn mode:cloud
-```
-#### Cloud & Account
-The cloud mode, with account portal turned on. This is a replica of the mode that runs at https://budibase.app
-
-
-To enable this mode, use:
-```
-yarn mode:account
-```
-### CI
- An overview of the CI pipelines can be found [here](./workflows/README.md)
-### Troubleshooting
-
-Sometimes, things go wrong. This can be due to incompatible updates on the budibase platform. To clear down your development environment and start again follow **Step 6. Cleanup**, then proceed from **Step 3. Install and Build** in the setup guide above. You should have a fresh Budibase installation.
-### Running tests
-
-#### End-to-end Tests
-
-Budibase uses Cypress to run a number of E2E tests. To run the tests execute the following command in the root folder:
-
-```
-yarn test:e2e
-```
-
-Or if you are in the builder you can run `yarn cy:test`.
-
-
-### Other Useful Information
-
-* The contributors are listed in [AUTHORS.md](https://github.com/Budibase/budibase/blob/master/.github/AUTHORS.md) (add yourself).
-
-* This project uses a modified version of the MPLv2 license, see [LICENSE](https://github.com/budibase/server/blob/master/LICENSE).
-
-* We use the [C4 (Collective Code Construction Contract)](https://rfc.zeromq.org/spec:42/C4/) process for contributions.
- Please read this if you are unfamiliar with it.
diff --git a/.github/CONTRIBUTING.md b/.github/CONTRIBUTING.md
new file mode 120000
index 0000000000..c81d5e88a2
--- /dev/null
+++ b/.github/CONTRIBUTING.md
@@ -0,0 +1 @@
+../docs/CONTRIBUTING.md
\ No newline at end of file
diff --git a/.github/workflows/README.md b/.github/workflows/README.md
index d2fcd16bb0..c33665c964 100644
--- a/.github/workflows/README.md
+++ b/.github/workflows/README.md
@@ -6,7 +6,7 @@ Welcome to the budibase CI pipelines directory. This document details what each
## All CI Pipelines
### Note
-- When running workflow dispatch jobs, ensure you always run them off the `master` branch. It defaults to `develop`, so double check before running any jobs.
+- When running workflow dispatch jobs, ensure you always run them off the `master` branch. It defaults to `develop`, so double check before running any jobs. The exception to this case is the `deploy-release` job which requires the develop branch.
### Standard CI Build Job (budibase_ci.yml)
Triggers:
@@ -24,14 +24,14 @@ The standard CI Build job is what runs when you raise a PR to develop or master.
Triggers:
- Push to develop
-The job responsible for building, tagging and pushing docker images out to the test and staging environments.
+The job responsible for building, tagging and pushing docker images out to the test and release environments.
- Installs all dependencies
- builds the project
- run the unit tests
- publish the budibase JS packages under a prerelease tag to NPM
- build, tag and push docker images under the `develop` tag to docker hub
-These images will then be pulled by the test and staging environments, updating the latest automatically. Discord notifications are sent to the #infra channel when this occurs.
+These images will then be pulled by the test and release environments, updating the latest automatically. Discord notifications are sent to the #infra channel when this occurs.
### Release Job (release.yml)
Triggers:
@@ -57,8 +57,33 @@ This job relies on the release job to have run first, so the latest image is pus
- Build and release the budibase helm chart for kubernetes users
- Perform a github release with the latest version. You can see previous releases here (https://github.com/Budibase/budibase/releases)
+### Deploy Release (deploy-release.yml)
+Triggers:
+- Manual Workflow Dispatch Trigger
-### Cloud Deploy (deploy-cloud.yml)
+This job is responsible for deploying to our release, cloud kubernetes environment. You must run the release job first, to ensure that the latest images have been built and pushed to docker hub. After kicking off this job, the following will occur:
+
+- Checks out the release branch
+- Pulls the latest `values.yaml` from budibase infra, a private repo containing budibases infrastructure configuration
+- Gets the latest budibase version from `lerna.json`, if it hasn't been specified in the workflow when you kicked it off
+- Configures AWS Credentials
+- Deploys the helm chart in the budibase repo to our preproduction EKS cluster, injecting the `values.yaml` we pulled from budibase-infra
+- Fires off a discord webhook in the #infra channel to show that the deployment completely successfully.
+
+### Deploy Preprod (deploy-preprod.yml)
+Triggers:
+- Manual Workflow Dispatch Trigger
+
+This job is responsible for deploying to our preprod, cloud kubernetes environment. You must run the release job first, to ensure that the latest images have been built and pushed to docker hub. After kicking off this job, the following will occur:
+
+- Checks out the master branch
+- Pulls the latest `values.yaml` from budibase infra, a private repo containing budibases infrastructure configuration
+- Gets the latest budibase version from `lerna.json`, if it hasn't been specified in the workflow when you kicked it off
+- Configures AWS Credentials
+- Deploys the helm chart in the budibase repo to our preprod EKS cluster, injecting the `values.yaml` we pulled from budibase-infra
+- Fires off a discord webhook in the #infra channel to show that the deployment completely successfully.
+
+### Deploy Production (deploy-cloud.yml)
Triggers:
- Manual Workflow Dispatch Trigger
@@ -90,4 +115,75 @@ This job is responsible for deploying to our production, cloud kubernetes enviro
### Rollback A Bad Cloud Deployment
- Kick off cloud deploy job
- Ensure you are running off master
-- Enter the version number of the last known good version of budibase. For example `1.0.0`
\ No newline at end of file
+- Enter the version number of the last known good version of budibase. For example `1.0.0`
+
+## Pro
+
+### Installing Pro
+
+The pro package is always installed from source in our CI jobs.
+
+This is done to prevent pro needing to be published prior to CI runs in budiabse. This is required for two reasons:
+- To reduce developer need to manually bump versions, i.e:
+ - release pro, bump pro dep in budibase, now ci can run successfully
+- The cyclic dependency on backend-core, i.e:
+ - pro depends on backend-core
+ - server depends on pro
+ - backend-core lives in the monorepo, so it can't be released independently to be used in pro
+ - therefore the only option is to pull pro from source and release it as a part of the monorepo release, as if it were a mono package
+
+The install is performed using the same steps as local development, via the `yarn bootstrap` command, see the [Contributing Guide#Pro](../CONTRIBUTING.md#pro)
+
+The branch to install pro from can vary depending on ref of the commit that triggered the budibase CI job. This is done to enable branches which have changes in both the monorepo and the pro repo to have their CI pass successfully.
+
+This is done using the [pro/install.sh](../../scripts/pro/install.sh) script. The script will:
+- Clone pro to it's default branch (`develop`)
+- Check if the clone worked, on forked versions of budibase this will fail due to no access
+ - This is fine as the `yarn` command will install the version from NPM
+ - Community PRs should never touch pro so this will always work
+- Checkout the `BRANCH` argument, if this fails fallback to `BASE_BRANCH`
+ - This enables the more complex case of a feature branch being merged to another feature branch, e.g.
+ - I am working on a branch `epic/stonks` which exists on budibase and pro.
+ - I want to merge a change to this branch in budibase from `feature/stonks-ui`, which only exists in budibase
+ - The base branch ensures that `epic/stonks` in pro will still be checked out for the CI run, rather than falling back to `develop`
+- Run `yarn setup` to build and install dependencies
+ - `yarn`
+ - `yarn bootstrap`
+ - `yarn build`
+ - The will build .ts files, and also update the `main` and `types` of `package.json` to point to `dist` rather than src
+ - The build command will only ever work in CI, it is prevented in local dev
+
+#### `BRANCH` and `BASE_BRANCH` arguments
+These arguments are supplied by the various budibase build and release pipelines
+- `budibase_ci`
+ - `BRANCH: ${{ github.event.pull_request.head.ref }}` -> The branch being merged
+ - `BASE_BRANCH: ${{ github.event.pull_request.base.ref}}` -> The base branch
+- `release-develop`
+ - `BRANCH: develop` -> always use the `develop` branch in pro
+- `release`
+ - `BRANCH: master` -> always use the `master` branch in pro
+
+
+### Releasing Pro
+After budibase dependencies have been released we will release the new version of pro to match the release version of budibase dependencies. This is to ensure that we are always keeping the version of `backend-core` in sync in the pro package and in budibase packages. Without this we could run into scenarios where different versions are being used when installed via `yarn` inside the docker images, creating very difficult to debug cases.
+
+Pro is released using the [pro/release.sh](../../scripts/pro/release.sh) script. The script will:
+- Inspect the `VERSION` from the `lerna.json` file in budibase
+- Determine whether to use the `latest` or `develop` tag based on the command argument
+- Go to pro directory
+ - install npm creds
+ - update the version of `backend-core` to be `VERSION`, the version just released by lerna
+ - publish to npm. Uses a `lerna publish` command, pro itself is a mono repo.
+ - force the version to be the same as `VERSION` to keep pro and budibase in sync
+ - reverts the changes to `main` and `types` in `package.json` that were made by the build step, to point back to source
+ - commit & push: `Prep next development iteration`
+- Go to budibase
+ - Update to the new version of pro in `server` and `worker` so the latest pro version is used in the docker builds
+ - commit & push: `Update pro version to $VERSION`
+
+
+#### `COMMAND` argument
+This argument is supplied by the existing `release` and `release:develop` budibase commands, which invoke the pro release
+- `release` will supply no command and default to use `latest`
+- `release:develop` will supply `develop`
+
diff --git a/.github/workflows/budibase_ci.yml b/.github/workflows/budibase_ci.yml
index e940e6fa10..1303d5921a 100644
--- a/.github/workflows/budibase_ci.yml
+++ b/.github/workflows/budibase_ci.yml
@@ -7,6 +7,7 @@ on:
branches:
- master
- develop
+ - new-design-ui
pull_request:
branches:
- master
@@ -59,3 +60,19 @@ jobs:
with:
install: false
command: yarn test:e2e:ci
+
+ - name: Configure AWS Credentials
+ uses: aws-actions/configure-aws-credentials@v1
+ with:
+ aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }}
+ aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
+ aws-region: eu-west-1
+
+ - name: Upload to S3
+ if: github.ref == 'refs/heads/new-design-ui'
+ run: |
+ tar -czvf new_ui.tar.gz packages/server/assets packages/server/index.html
+ aws s3 cp new_ui.tar.gz s3://prod-budi-app-assets/beta:design_ui/
+ aws s3 cp packages/client/dist/budibase-client.js s3://prod-budi-app-assets/beta:design_ui/budibase-client.js
+ aws cloudfront create-invalidation --distribution-id E3ELKP4RCEHVLW --paths "/beta:design_ui/*"
+
diff --git a/.github/workflows/deploy-cloud.yaml b/.github/workflows/deploy-cloud.yaml
index a05f97f097..869a88a5b3 100644
--- a/.github/workflows/deploy-cloud.yaml
+++ b/.github/workflows/deploy-cloud.yaml
@@ -1,4 +1,4 @@
-name: Budibase Cloud Deploy
+name: Budibase Deploy Production
on:
workflow_dispatch:
diff --git a/.github/workflows/deploy-preprod.yml b/.github/workflows/deploy-preprod.yml
index ab941100a0..c3f690f568 100644
--- a/.github/workflows/deploy-preprod.yml
+++ b/.github/workflows/deploy-preprod.yml
@@ -1,4 +1,4 @@
-name: Budibase Release Preprod
+name: Budibase Deploy Preprod
on:
workflow_dispatch:
diff --git a/.github/workflows/deploy-release.yml b/.github/workflows/deploy-release.yml
new file mode 100644
index 0000000000..0fb8a5fea0
--- /dev/null
+++ b/.github/workflows/deploy-release.yml
@@ -0,0 +1,77 @@
+name: Budibase Deploy Release
+
+on:
+ workflow_dispatch:
+
+jobs:
+ release:
+ runs-on: ubuntu-latest
+
+ steps:
+ - uses: actions/checkout@v2
+
+ - name: Configure AWS Credentials
+ uses: aws-actions/configure-aws-credentials@v1
+ with:
+ aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }}
+ aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
+ aws-region: eu-west-1
+
+ - name: Fail if branch is not develop
+ if: github.ref != 'refs/heads/develop'
+ run: |
+ echo "Ref is not develop, you must run this job from develop."
+ exit 1
+
+ - name: Get the latest budibase release version
+ id: version
+ run: |
+ release_version=$(cat lerna.json | jq -r '.version')
+ echo "RELEASE_VERSION=$release_version" >> $GITHUB_ENV
+
+ - name: Tag and release Proxy service docker image
+ run: |
+ docker login -u $DOCKER_USER -p $DOCKER_PASSWORD
+ yarn build:docker:proxy:release
+ docker tag proxy-service budibase/proxy:$RELEASE_TAG
+ docker push budibase/proxy:$RELEASE_TAG
+ env:
+ DOCKER_USER: ${{ secrets.DOCKER_USERNAME }}
+ DOCKER_PASSWORD: ${{ secrets.DOCKER_API_KEY }}
+ RELEASE_TAG: k8s-release
+
+ - name: Pull values.yaml from budibase-infra
+ run: |
+ curl -H "Authorization: token ${{ secrets.GH_PERSONAL_TOKEN }}" \
+ -H 'Accept: application/vnd.github.v3.raw' \
+ -o values.release.yaml \
+ -L https://api.github.com/repos/budibase/budibase-infra/contents/kubernetes/budibase-release/values.yaml
+ wc -l values.release.yaml
+
+ - name: Deploy to Release Environment
+ uses: glopezep/helm@v1.7.1
+ with:
+ release: budibase-release
+ namespace: budibase
+ chart: charts/budibase
+ token: ${{ github.token }}
+ helm: helm3
+ values: |
+ globals:
+ appVersion: develop
+ ingress:
+ enabled: true
+ nginx: true
+ value-files: >-
+ [
+ "values.release.yaml"
+ ]
+ env:
+ KUBECONFIG_FILE: '${{ secrets.RELEASE_KUBECONFIG }}'
+
+ - name: Discord Webhook Action
+ uses: tsickert/discord-webhook@v4.0.0
+ with:
+ webhook-url: ${{ secrets.PROD_DEPLOY_WEBHOOK_URL }}
+ content: "Release Env Deployment Complete: ${{ env.RELEASE_VERSION }} deployed to Budibase Release Env."
+ embed-title: ${{ env.RELEASE_VERSION }}
diff --git a/.github/workflows/release-develop.yml b/.github/workflows/release-develop.yml
index 71d487e0a3..631308d945 100644
--- a/.github/workflows/release-develop.yml
+++ b/.github/workflows/release-develop.yml
@@ -1,10 +1,10 @@
-name: Budibase Release Staging
-concurrency: release-develop
+name: Budibase Prerelease
+concurrency: release-prerelease
on:
push:
branches:
- - release
+ - develop
paths:
- '.aws/**'
- '.github/**'
@@ -21,18 +21,19 @@ env:
# Posthog token used by ui at build time
POSTHOG_TOKEN: phc_uDYOfnFt6wAbBAXkC6STjcrTpAFiWIhqgFcsC1UVO5F
INTERCOM_TOKEN: ${{ secrets.INTERCOM_TOKEN }}
- PERSONAL_ACCESS_TOKEN : ${{ secrets.PERSONAL_ACCESS_TOKEN }}
+ PERSONAL_ACCESS_TOKEN: ${{ secrets.PERSONAL_ACCESS_TOKEN }}
+ FEATURE_PREVIEW_URL: https://budirelease.live
jobs:
release:
runs-on: ubuntu-latest
steps:
- # - name: Fail if branch is not develop
- # if: github.ref != 'refs/heads/develop'
- # run: |
- # echo "Ref is not develop, you must run this job from develop."
- # exit 1
+ - name: Fail if branch is not develop
+ if: github.ref != 'refs/heads/develop'
+ run: |
+ echo "Ref is not develop, you must run this job from develop."
+ exit 1
- uses: actions/checkout@v2
- uses: actions/setup-node@v1
with:
@@ -124,4 +125,4 @@ jobs:
with:
webhook-url: ${{ secrets.PROD_DEPLOY_WEBHOOK_URL }}
content: "Release Env Deployment Complete: ${{ env.RELEASE_VERSION }} deployed to Budibase Release Env."
- embed-title: ${{ env.RELEASE_VERSION }}
\ No newline at end of file
+ embed-title: ${{ env.RELEASE_VERSION }}
diff --git a/README.md b/README.md
index 17a3ab1ef2..e8c6475d90 100644
--- a/README.md
+++ b/README.md
@@ -174,6 +174,7 @@ Budibase is dedicated to providing a welcoming, diverse, and harrassment-free ex
## 🙌 Contributing to Budibase
From opening a bug report to creating a pull request: every contribution is appreciated and welcomed. If you're planning to implement a new feature or change the API please create an issue first. This way we can ensure your work is not in vain.
+Environment setup instructions are available for [Debian](https://github.com/Budibase/budibase/tree/HEAD/docs/DEV-SETUP-DEBIAN.md) and [MacOSX](https://github.com/Budibase/budibase/tree/HEAD/docs/DEV-SETUP-MACOSX.md)
### Not Sure Where to Start?
A good place to start contributing, is the [First time issues project](https://github.com/Budibase/budibase/projects/22).
@@ -187,7 +188,7 @@ Budibase is a monorepo managed by lerna. Lerna manages the building and publishi
- [packages/server](https://github.com/Budibase/budibase/tree/HEAD/packages/server) - The budibase server. This Koa app is responsible for serving the JS for the builder and budibase apps, as well as providing the API for interaction with the database and file system.
-For more information, see [CONTRIBUTING.md](https://github.com/Budibase/budibase/blob/HEAD/.github/CONTRIBUTING.md)
+For more information, see [CONTRIBUTING.md](https://github.com/Budibase/budibase/blob/HEAD/docs/CONTRIBUTING.md)
@@ -202,7 +203,7 @@ Budibase is open-source, licensed as [GPL v3](https://www.gnu.org/licenses/gpl-3
[![Stargazers over time](https://starchart.cc/Budibase/budibase.svg)](https://starchart.cc/Budibase/budibase)
-If you are having issues between updates of the builder, please use the guide [here](https://github.com/Budibase/budibase/blob/HEAD/.github/CONTRIBUTING.md#troubleshooting) to clear down your environment.
+If you are having issues between updates of the builder, please use the guide [here](https://github.com/Budibase/budibase/blob/HEAD/docs/CONTRIBUTING.md#troubleshooting) to clear down your environment.
diff --git a/charts/budibase/templates/worker-service-deployment.yaml b/charts/budibase/templates/worker-service-deployment.yaml
index 73bb26dad9..a7f05f3137 100644
--- a/charts/budibase/templates/worker-service-deployment.yaml
+++ b/charts/budibase/templates/worker-service-deployment.yaml
@@ -123,6 +123,8 @@ spec:
value: {{ .Values.globals.google.clientId | quote }}
- name: GOOGLE_CLIENT_SECRET
value: {{ .Values.globals.google.secret | quote }}
+ - name: TENANT_FEATURE_FLAGS
+ value: {{ .Values.globals.tenantFeatureFlags | quote }}
image: budibase/worker:{{ .Values.globals.appVersion }}
imagePullPolicy: Always
livenessProbe:
diff --git a/charts/budibase/values.yaml b/charts/budibase/values.yaml
index 455d3251a8..2734202fff 100644
--- a/charts/budibase/values.yaml
+++ b/charts/budibase/values.yaml
@@ -103,7 +103,7 @@ globals:
google:
clientId: ""
secret: ""
- automationMaxIterations: "500"
+ automationMaxIterations: "200"
createSecrets: true # creates an internal API key, JWT secrets and redis password for you
diff --git a/docs/CODE_OF_CONDUCT.md b/docs/CODE_OF_CONDUCT.md
new file mode 100644
index 0000000000..1be855e3fa
--- /dev/null
+++ b/docs/CODE_OF_CONDUCT.md
@@ -0,0 +1,76 @@
+# Contributor Covenant Code of Conduct
+
+## Our Pledge
+
+In the interest of fostering an open and welcoming environment, we as
+contributors and maintainers pledge to making participation in our project and
+our community a harassment-free experience for everyone, regardless of age, body
+size, disability, ethnicity, sex characteristics, gender identity and expression,
+level of experience, education, socio-economic status, nationality, personal
+appearance, race, religion, or sexual identity and orientation.
+
+## Our Standards
+
+Examples of behavior that contributes to creating a positive environment
+include:
+
+* Using welcoming and inclusive language
+* Being respectful of differing viewpoints and experiences
+* Gracefully accepting constructive criticism
+* Focusing on what is best for the community
+* Showing empathy towards other community members
+
+Examples of unacceptable behavior by participants include:
+
+* The use of sexualized language or imagery and unwelcome sexual attention or
+ advances
+* Trolling, insulting/derogatory comments, and personal or political attacks
+* Public or private harassment
+* Publishing others' private information, such as a physical or electronic
+ address, without explicit permission
+* Other conduct which could reasonably be considered inappropriate in a
+ professional setting
+
+## Our Responsibilities
+
+Project maintainers are responsible for clarifying the standards of acceptable
+behavior and are expected to take appropriate and fair corrective action in
+response to any instances of unacceptable behavior.
+
+Project maintainers have the right and responsibility to remove, edit, or
+reject comments, commits, code, wiki edits, issues, and other contributions
+that are not aligned to this Code of Conduct, or to ban temporarily or
+permanently any contributor for other behaviors that they deem inappropriate,
+threatening, offensive, or harmful.
+
+## Scope
+
+This Code of Conduct applies both within project spaces and in public spaces
+when an individual is representing the project or its community. Examples of
+representing a project or community include using an official project e-mail
+address, posting via an official social media account, or acting as an appointed
+representative at an online or offline event. Representation of a project may be
+further defined and clarified by project maintainers.
+
+## Enforcement
+
+Instances of abusive, harassing, or otherwise unacceptable behavior may be
+reported by contacting the project team at community@budibase.com. All
+complaints will be reviewed and investigated and will result in a response that
+is deemed necessary and appropriate to the circumstances. The project team is
+obligated to maintain confidentiality with regard to the reporter of an incident.
+Further details of specific enforcement policies may be posted separately.
+
+Project maintainers who do not follow or enforce the Code of Conduct in good
+faith may face temporary or permanent repercussions as determined by other
+members of the project's leadership.
+
+## Attribution
+
+This Code of Conduct is adapted from the [Contributor Covenant][homepage], version 1.4,
+available at https://www.contributor-covenant.org/version/1/4/code-of-conduct.html
+
+[homepage]: https://www.contributor-covenant.org
+
+For answers to common questions about this code of conduct, see
+https://www.contributor-covenant.org/faq
diff --git a/docs/CONTRIBUTING.md b/docs/CONTRIBUTING.md
new file mode 100644
index 0000000000..531ed05749
--- /dev/null
+++ b/docs/CONTRIBUTING.md
@@ -0,0 +1,231 @@
+# Contributing
+
+From opening a bug report to creating a pull request: every contribution is appreciated and welcome. If you're planning to implement a new feature or change the api please [create an issue](https://github.com/Budibase/budibase/issues/new/choose) first. This way we can ensure that your precious work is not in vain.
+
+## Table of contents
+
+- [Quick start](#quick-start)
+- [Status](#status)
+- [What's included](#whats-included)
+- [Bugs and feature requests](#bugs-and-feature-requests)
+
+
+## Not Sure Where to Start?
+
+Budibase is a low-code web application builder that creates svelte-based web applications.
+
+Budibase is a monorepo managed by [lerna](https://github.com/lerna/lerna). Lerna manages the building and publishing of the budibase packages. At a high level, here are the packages that make up budibase.
+
+- **packages/builder** - contains code for the budibase builder client side svelte application.
+
+- **packages/client** - A module that runs in the browser responsible for reading JSON definition and creating living, breathing web apps from it.
+
+- **packages/server** - The budibase server. This [Koa](https://koajs.com/) app is responsible for serving the JS for the builder and budibase apps, as well as providing the API for interaction with the database and file system.
+
+- **packages/worker** - This [Koa](https://koajs.com/) app is responsible for providing global apis for managing your budibase installation. Authentication, Users, Email, Org and Auth configs are all provided by the worker.
+
+## Contributor License Agreement (CLA)
+
+In order to accept your pull request, we need you to submit a CLA. You only need to do this once. If you are submitting a pull request for the first time, just submit a Pull Request and our CLA Bot will give you instructions on how to sign the CLA before merging your Pull Request.
+
+All contributors must sign an [Individual Contributor License Agreement](https://github.com/budibase/budibase/blob/next/.github/cla/individual-cla.md).
+
+If contributing on behalf of your company, your company must sign a [Corporate Contributor License Agreement](https://github.com/budibase/budibase/blob/next/.github/cla/corporate-cla.md). If so, please contact us via community@budibase.com.
+
+## Glossary of Terms
+
+To understand the budibase API, it can be helpful to understand the top level entities that make up Budibase.
+
+### Client
+
+A client represents a single budibase customer. Each budibase client will have 1 or more budibase servers. Every client is assigned a unique ID.
+
+### App
+
+A client can have one or more budibase applications. Budibase applications would be things like "Developer Inventory Management" or "Goat Herder CRM". Think of a budibase application as a tree.
+
+### Database
+
+An App can have one or more databases. Keeping with our [dendrology](https://en.wikipedia.org/wiki/Dendrology) analogy - think of an database as a branch on the tree. Databases are used to keep data separate for different instances of your app. For example, if you had a CRM app, you may create a database for your US office, and a database for your Australian office. Databases allow us to support [multitenancy](https://www.gartner.com/en/information-technology/glossary/multitenancy) in budibase applications.
+
+### Table
+
+Tables in budibase are almost akin to tables in relational databases. A table may be a "Car" or an "Employee". They are the main building blocks for the creation and management of backend data in budibase.
+
+### View
+
+A View is an advanced feature in budibase that allows you to write a custom query using [MapReduce](https://pouchdb.com/guides/queries.html) queries. Views enable powerful query functionality and calculations, allowing you to do more with your data.
+
+### Page
+
+A page in budibase is actually a single, self contained svelte web app. There are only 2 pages in budibase. The **login** page and the **main** page.
+
+### Screen
+
+A screen is a component within a single page. Generally, screens represent client side routes, and can be switched without refreshing the page.
+
+### Component
+
+A component is the basic frontend building block of a budibase app.
+
+### Component Library
+
+Component libraries are collections of components as well as the definition of their props contained in a file called `components.json`.
+
+## Contributing to Budibase
+
+* Please maintain the existing code style.
+
+* Please try to keep your commits small and focused.
+
+* Please write tests.
+
+* If the project diverges from your branch, please rebase instead of merging. This makes the commit graph easier to read.
+
+* Once your work is completed, please raise a PR against the `develop` branch with some information about what has changed and why.
+
+### Getting Started For Contributors
+#### 1. Prerequisites
+
+NodeJS Version `14.x.x`
+
+*yarn -* `npm install -g yarn`
+
+*jest* - `npm install -g jest`
+
+#### 2. Clone this repository
+
+`git clone https://github.com/Budibase/budibase.git`
+
+then `cd ` into your local copy.
+
+#### 3. Install and Build
+
+| **NOTE**: On Windows, all yarn commands must be executed on a bash shell (e.g. git bash)
+
+To develop the Budibase platform you'll need [Docker](https://www.docker.com/) and [Docker Compose](https://docs.docker.com/compose/) installed.
+
+##### Quick method
+
+`yarn setup` will check that all necessary components are installed and setup the repo for usage.
+
+##### Manual method
+
+The following commands can be executed to manually get Budibase up and running (assuming Docker/Docker Compose has been installed).
+
+`yarn` to install project dependencies
+
+`yarn bootstrap` will install all budibase modules and symlink them together using lerna.
+
+`yarn build` will build all budibase packages.
+
+#### 4. Running
+
+To run the budibase server and builder in dev mode (i.e. with live reloading):
+
+1. Open a new console
+2. `yarn dev` (from root)
+3. Access the builder on http://localhost:10000/builder
+
+This will enable watch mode for both the builder app, server, client library and any component libraries.
+
+#### 5. Debugging using VS Code
+
+To debug the budibase server and worker a VS Code launch configuration has been provided.
+
+Visit the debug window and select `Budibase Server` or `Budibase Worker` to debug the respective component.
+Alternatively to start both components simultaneously select `Start Budibase`.
+
+In addition to the above, the remaining budibase components may be run in dev mode using: `yarn dev:noserver`.
+
+#### 6. Cleanup
+
+If you wish to delete all the apps created in development and reset the environment then run the following:
+
+1. `yarn nuke:docker` will wipe all the Budibase services
+2. `yarn dev` will restart all the services
+
+### Backend
+
+For the backend we run [Redis](https://redis.io/), [CouchDB](https://couchdb.apache.org/), [MinIO](https://min.io/) and [NGINX](https://www.nginx.com/) in Docker compose. This means that to develop Budibase you will need Docker and Docker compose installed. The backend services are then run separately as Node services with nodemon so that they can be debugged outside of Docker.
+
+### Data Storage
+
+When you are running locally, budibase stores data on disk using docker volumes. The volumes and the types of data associated with each are:
+
+- `redis_data`
+ - Sessions, email tokens
+- `couchdb3_data`
+ - Global and app databases
+- `minio_data`
+ - App manifest, budibase client, static assets
+
+### Development Modes
+
+A combination of environment variables controls the mode budibase runs in.
+Yarn commands can be used to mimic the different modes as described in the sections below:
+
+#### Self Hosted
+The default mode. A single tenant installation with no usage restrictions.
+
+To enable this mode, use:
+```
+yarn mode:self
+```
+
+#### Cloud
+The cloud mode, with account portal turned off.
+
+To enable this mode, use:
+```
+yarn mode:cloud
+```
+#### Cloud & Account
+The cloud mode, with account portal turned on. This is a replica of the mode that runs at https://budibase.app
+
+
+To enable this mode, use:
+```
+yarn mode:account
+```
+### CI
+ An overview of the CI pipelines can be found [here](./workflows/README.md)
+
+### Pro
+
+@budibase/pro is the closed source package that supports licensed features in budibase. By default the package will be pulled from NPM and will not normally need to be touched in local development. If you require to update code inside the pro package it can be cloned to the same root level as budibase, e.g.
+
+```
+.
+|_ budibase
+|_ budibase-pro
+```
+
+Note that only budibase maintainers will be able to access the pro repo.
+
+The `yarn bootstrap` command can be used to replace the NPM supplied dependency with the local source aware version. This is achieved using the `yarn link` command. To see specifically how dependencies are linked see [scripts/link-dependencies.sh](../scripts/link-dependencies.sh). The same link script is used to link dependencies to account-portal in local dev.
+
+### Troubleshooting
+
+Sometimes, things go wrong. This can be due to incompatible updates on the budibase platform. To clear down your development environment and start again follow **Step 6. Cleanup**, then proceed from **Step 3. Install and Build** in the setup guide above to create a fresh Budibase installation.
+### Running tests
+
+#### End-to-end Tests
+
+Budibase uses Cypress to run a number of E2E tests. To run the tests execute the following command in the root folder:
+
+```
+yarn test:e2e
+```
+
+Or if you are in the builder you can run `yarn cy:test`.
+
+
+### Other Useful Information
+
+* The contributors are listed in [AUTHORS.md](https://github.com/Budibase/budibase/blob/master/.github/AUTHORS.md) (add yourself).
+
+* This project uses a modified version of the MPLv2 license, see [LICENSE](https://github.com/budibase/server/blob/master/LICENSE).
+
+* We use the [C4 (Collective Code Construction Contract)](https://rfc.zeromq.org/spec:42/C4/) process for contributions.
+ Please read this if you are unfamiliar with it.
diff --git a/docs/DEV-SETUP-DEBIAN.md b/docs/DEV-SETUP-DEBIAN.md
new file mode 100644
index 0000000000..88a124708c
--- /dev/null
+++ b/docs/DEV-SETUP-DEBIAN.md
@@ -0,0 +1,52 @@
+## Dev Environment on Debian 11
+
+### Install Node
+
+Budibase requires a recent version of node (14+):
+```
+curl -sL https://deb.nodesource.com/setup_16.x | sudo bash -
+apt -y install nodejs
+node -v
+```
+
+### Install npm requirements
+
+```
+npm install -g yarn jest lerna
+```
+### Install Docker and Docker Compose
+
+```
+apt install docker.io
+pip3 install docker-compose
+```
+### Clone the repo
+```
+git clone https://github.com/Budibase/budibase.git
+```
+
+### Check Versions
+
+This setup process was tested on Debian 11 (bullseye) with version numbers show below. Your mileage may vary using anything else.
+
+- Docker: 20.10.5
+- Docker-Compose: 1.29.2
+- Node: v16.15.1
+- Yarn: 1.22.19
+- Lerna: 5.1.4
+
+### Build
+
+```
+cd budibase
+yarn setup
+```
+The yarn setup command runs several build steps i.e.
+```
+node ./hosting/scripts/setup.js && yarn && yarn bootstrap && yarn build && yarn dev
+```
+So this command will actually run the application in dev mode. It creates .env files under `./packages/server` and `./packages/worker` and runs docker containers for each service via docker-compose.
+
+The dev version will be available on port 10000 i.e.
+
+http://127.0.0.1:10000/builder/admin
\ No newline at end of file
diff --git a/docs/DEV-SETUP-MACOSX.md b/docs/DEV-SETUP-MACOSX.md
new file mode 100644
index 0000000000..5606fd0d10
--- /dev/null
+++ b/docs/DEV-SETUP-MACOSX.md
@@ -0,0 +1,54 @@
+## Dev Environment on MAC OSX 12 (Monterey)
+
+### Install Homebrew
+
+Install instructions [here](https://brew.sh/)
+
+### Install Node
+
+Budibase requires a recent version of node (14+):
+```
+brew install node npm
+node -v
+```
+
+### Install npm requirements
+
+```
+npm install -g yarn jest lerna
+```
+### Install Docker and Docker Compose
+
+```
+brew install docker docker-compose
+```
+### Clone the repo
+```
+git clone https://github.com/Budibase/budibase.git
+```
+
+### Check Versions
+
+This setup process was tested on Mac OSX 12 (Monterey) with version numbers shown below. Your mileage may vary using anything else.
+
+- Docker: 20.10.14
+- Docker-Compose: 2.6.0
+- Node: 18.3.0
+- Yarn: 1.22.19
+- Lerna: 5.1.4
+
+### Build
+
+```
+cd budibase
+yarn setup
+```
+The yarn setup command runs several build steps i.e.
+```
+node ./hosting/scripts/setup.js && yarn && yarn bootstrap && yarn build && yarn dev
+```
+So this command will actually run the application in dev mode. It creates .env files under `./packages/server` and `./packages/worker` and runs docker containers for each service via docker-compose.
+
+The dev version will be available on port 10000 i.e.
+
+http://127.0.0.1:10000/builder/admin
\ No newline at end of file
diff --git a/hosting/letsencrypt/certificate-renew.sh b/hosting/letsencrypt/certificate-renew.sh
new file mode 100644
index 0000000000..df88b44322
--- /dev/null
+++ b/hosting/letsencrypt/certificate-renew.sh
@@ -0,0 +1,13 @@
+#!/bin/bash
+CUSTOM_DOMAIN="$1"
+
+if [[ ! -z "${CUSTOM_DOMAIN}" ]]; then
+ certbot certonly --webroot --webroot-path="/var/www/html" \
+ --register-unsafely-without-email \
+ --domains $CUSTOM_DOMAIN \
+ --rsa-key-size 4096 \
+ --agree-tos \
+ --force-renewal
+
+ nginx -s reload
+fi
diff --git a/hosting/letsencrypt/certificate-request.sh b/hosting/letsencrypt/certificate-request.sh
new file mode 100644
index 0000000000..83f314fc88
--- /dev/null
+++ b/hosting/letsencrypt/certificate-request.sh
@@ -0,0 +1,23 @@
+#!/bin/bash
+CUSTOM_DOMAIN="$1"
+# Request from Lets Encrypt
+certbot certonly --webroot --webroot-path="/var/www/html" \
+ --register-unsafely-without-email \
+ --domains $CUSTOM_DOMAIN \
+ --rsa-key-size 4096 \
+ --agree-tos \
+ --force-renewal
+
+if (($? != 0)); then
+ echo "ERROR: certbot request failed for $CUSTOM_DOMAIN use http on port 80 - exiting"
+ exit 1
+else
+ cp /app/letsencrypt/options-ssl-nginx.conf /etc/letsencrypt/options-ssl-nginx.conf
+ cp /app/letsencrypt/ssl-dhparams.pem /etc/letsencrypt/ssl-dhparams.pem
+ cp /app/letsencrypt/nginx-ssl.conf /etc/nginx/sites-available/nginx-ssl.conf
+ sed -i "s/CUSTOM_DOMAIN/$CUSTOM_DOMAIN/g" /etc/nginx/sites-available/nginx-ssl.conf
+ ln -s /etc/nginx/sites-available/nginx-ssl.conf /etc/nginx/sites-enabled/nginx-ssl.conf
+
+ echo "INFO: restart nginx after certbot request"
+ /etc/init.d/nginx restart
+fi
diff --git a/hosting/letsencrypt/nginx-ssl.conf b/hosting/letsencrypt/nginx-ssl.conf
new file mode 100644
index 0000000000..50c5e0198a
--- /dev/null
+++ b/hosting/letsencrypt/nginx-ssl.conf
@@ -0,0 +1,96 @@
+server {
+ listen 443 ssl default_server;
+ listen [::]:443 ssl default_server;
+ server_name _;
+ ssl_certificate /etc/letsencrypt/live/CUSTOM_DOMAIN/fullchain.pem;
+ ssl_certificate_key /etc/letsencrypt/live/CUSTOM_DOMAIN/privkey.pem;
+ include /etc/letsencrypt/options-ssl-nginx.conf;
+ ssl_dhparam /etc/letsencrypt/ssl-dhparams.pem;
+
+ client_max_body_size 1000m;
+ ignore_invalid_headers off;
+ proxy_buffering off;
+ # port_in_redirect off;
+
+ location ^~ /.well-known/acme-challenge/ {
+ default_type "text/plain";
+ root /var/www/html;
+ break;
+ }
+ location = /.well-known/acme-challenge/ {
+ return 404;
+ }
+
+ location /app {
+ proxy_pass http://127.0.0.1:4001;
+ }
+
+ location = / {
+ proxy_pass http://127.0.0.1:4001;
+ }
+
+ location ~ ^/(builder|app_) {
+ proxy_http_version 1.1;
+ proxy_set_header Connection $connection_upgrade;
+ proxy_set_header Upgrade $http_upgrade;
+ proxy_set_header X-Real-IP $remote_addr;
+ proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
+ proxy_pass http://127.0.0.1:4001;
+ }
+
+ location ~ ^/api/(system|admin|global)/ {
+ proxy_pass http://127.0.0.1:4002;
+ }
+
+ location /worker/ {
+ proxy_pass http://127.0.0.1:4002;
+ rewrite ^/worker/(.*)$ /$1 break;
+ }
+
+ location /api/ {
+ # calls to the API are rate limited with bursting
+ limit_req zone=ratelimit burst=20 nodelay;
+
+ # 120s timeout on API requests
+ proxy_read_timeout 120s;
+ proxy_connect_timeout 120s;
+ proxy_send_timeout 120s;
+
+ proxy_http_version 1.1;
+ proxy_set_header Connection $connection_upgrade;
+ proxy_set_header Upgrade $http_upgrade;
+ proxy_set_header X-Real-IP $remote_addr;
+ proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
+
+ proxy_pass http://127.0.0.1:4001;
+ }
+
+ location /db/ {
+ proxy_pass http://127.0.0.1:5984;
+ rewrite ^/db/(.*)$ /$1 break;
+ }
+
+ location / {
+ proxy_set_header X-Real-IP $remote_addr;
+ proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
+ proxy_set_header X-Forwarded-Proto $scheme;
+
+ proxy_connect_timeout 300;
+ proxy_http_version 1.1;
+ proxy_set_header Connection "";
+ chunked_transfer_encoding off;
+ proxy_pass http://127.0.0.1:9000;
+ }
+
+ client_header_timeout 60;
+ client_body_timeout 60;
+ keepalive_timeout 60;
+
+ # gzip
+ gzip on;
+ gzip_vary on;
+ gzip_proxied any;
+ gzip_comp_level 6;
+ gzip_types text/plain text/css text/xml application/json application/javascript application/rss+xml application/atom+xml image/svg+xml;
+
+}
diff --git a/hosting/letsencrypt/options-ssl-nginx.conf b/hosting/letsencrypt/options-ssl-nginx.conf
new file mode 100644
index 0000000000..52fdfde245
--- /dev/null
+++ b/hosting/letsencrypt/options-ssl-nginx.conf
@@ -0,0 +1,13 @@
+# This file contains important security parameters. If you modify this file
+# manually, Certbot will be unable to automatically provide future security
+# updates. Instead, Certbot will print and log an error message with a path to
+# the up-to-date file that you will need to refer to when manually updating
+# this file.
+
+ssl_session_cache shared:le_nginx_SSL:10m;
+ssl_session_timeout 1440m;
+
+ssl_protocols TLSv1.2 TLSv1.3;
+ssl_prefer_server_ciphers off;
+
+ssl_ciphers "ECDHE-ECDSA-AES128-GCM-SHA256:ECDHE-RSA-AES128-GCM-SHA256:ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384:ECDHE-ECDSA-CHACHA20-POLY1305:ECDHE-RSA-CHACHA20-POLY1305:DHE-RSA-AES128-GCM-SHA256:DHE-RSA-AES256-GCM-SHA384";
diff --git a/hosting/letsencrypt/ssl-dhparams.pem b/hosting/letsencrypt/ssl-dhparams.pem
new file mode 100644
index 0000000000..088f9673dc
--- /dev/null
+++ b/hosting/letsencrypt/ssl-dhparams.pem
@@ -0,0 +1,8 @@
+-----BEGIN DH PARAMETERS-----
+MIIBCAKCAQEA//////////+t+FRYortKmq/cViAnPTzx2LnFg84tNpWp4TZBFGQz
++8yTnc4kmz75fS/jY2MMddj2gbICrsRhetPfHtXV/WVhJDP1H18GbtCFY2VVPe0a
+87VXE15/V8k1mE8McODmi3fipona8+/och3xWKE2rec1MKzKT0g6eXq8CrGCsyT7
+YdEIqUuyyOP7uWrat2DX9GgdT0Kj3jlN9K5W7edjcrsZCwenyO4KbXCeAvzhzffi
+7MA0BM0oNC9hkXL+nOmFg/+OTxIy7vKBg8P+OxtMb61zO7X8vC7CIAXFjvGDfRaD
+ssbzSibBsu/6iGtCOGEoXJf//////////wIBAg==
+-----END DH PARAMETERS-----
\ No newline at end of file
diff --git a/hosting/scripts/build-target-paths.sh b/hosting/scripts/build-target-paths.sh
new file mode 100644
index 0000000000..d1c9b5cd05
--- /dev/null
+++ b/hosting/scripts/build-target-paths.sh
@@ -0,0 +1,17 @@
+#!/bin/bash
+
+echo ${TARGETBUILD} > /buildtarget.txt
+if [[ "${TARGETBUILD}" = "aas" ]]; then
+ # Azure AppService uses /home for persisent data & SSH on port 2222
+ mkdir -p /home/budibase/{minio,couchdb}
+ mkdir -p /home/budibase/couchdb/data
+ chown -R couchdb:couchdb /home/budibase/couchdb/
+ apt update
+ apt-get install -y openssh-server
+ sed -i 's#dir=/opt/couchdb/data/search#dir=/home/budibase/couchdb/data/search#' /opt/clouseau/clouseau.ini
+ sed -i 's#/minio/minio server /minio /minio/minio server /home/budibase/minio ' /runner.sh
+ sed -i 's#database_dir = ./data#database_dir = /home/budibase/couchdb/data#' /opt/couchdb/etc/default.ini
+ sed -i 's#view_index_dir = ./data#view_index_dir = /home/budibase/couchdb/data#' /opt/couchdb/etc/default.ini
+ sed -i "s/#Port 22/Port 2222/" /etc/ssh/sshd_config
+ /etc/init.d/ssh restart
+fi
diff --git a/hosting/scripts/healthcheck.sh b/hosting/scripts/healthcheck.sh
new file mode 100644
index 0000000000..80f2ece0b6
--- /dev/null
+++ b/hosting/scripts/healthcheck.sh
@@ -0,0 +1,40 @@
+#!/usr/bin/env bash
+healthy=true
+
+if [[ $(curl -Lfk -s -w "%{http_code}\n" http://localhost/ -o /dev/null) -ne 200 ]]; then
+ echo 'ERROR: Budibase is not running';
+ healthy=false
+fi
+
+if [[ $(curl -s -w "%{http_code}\n" http://localhost:4001/health -o /dev/null) -ne 200 ]]; then
+ echo 'ERROR: Budibase backend is not running';
+ healthy=false
+fi
+
+if [[ $(curl -s -w "%{http_code}\n" http://localhost:4002/health -o /dev/null) -ne 200 ]]; then
+ echo 'ERROR: Budibase worker is not running';
+ healthy=false
+fi
+
+if [[ $(curl -s -w "%{http_code}\n" http://localhost:5984/ -o /dev/null) -ne 200 ]]; then
+ echo 'ERROR: CouchDB is not running';
+ healthy=false
+fi
+if [[ $(redis-cli -a $REDIS_PASSWORD --no-auth-warning ping) != 'PONG' ]]; then
+ echo 'ERROR: Redis is down';
+ healthy=false
+fi
+# mino, clouseau,
+nginx -t -q
+NGINX_STATUS=$?
+
+if [[ $NGINX_STATUS -gt 0 ]]; then
+ echo 'ERROR: Nginx config problem';
+ healthy=false
+fi
+
+if [ $healthy == true ]; then
+ exit 0
+else
+ exit 1
+fi
diff --git a/hosting/single/Dockerfile b/hosting/single/Dockerfile
index 79d84f70c6..5e1b0b1374 100644
--- a/hosting/single/Dockerfile
+++ b/hosting/single/Dockerfile
@@ -1,79 +1,93 @@
-FROM couchdb
+FROM node:14-slim as build
-ENV DEPLOYMENT_ENVIRONMENT=docker
-ENV POSTHOG_TOKEN=phc_fg5I3nDOf6oJVMHSaycEhpPdlgS8rzXG2r6F2IpxCHS
-ENV COUCHDB_PASSWORD=budibase
-ENV COUCHDB_USER=budibase
-ENV COUCH_DB_URL=http://budibase:budibase@localhost:5984
-ENV BUDIBASE_ENVIRONMENT=PRODUCTION
-ENV MINIO_URL=http://localhost:9000
-ENV REDIS_URL=localhost:6379
-ENV WORKER_URL=http://localhost:4002
-ENV INTERNAL_API_KEY=budibase
-ENV JWT_SECRET=testsecret
-ENV MINIO_ACCESS_KEY=budibase
-ENV MINIO_SECRET_KEY=budibase
-ENV SELF_HOSTED=1
-ENV CLUSTER_PORT=10000
-ENV REDIS_PASSWORD=budibase
-ENV ARCHITECTURE=amd
-ENV APP_PORT=4001
-ENV WORKER_PORT=4002
+# install node-gyp dependencies
+RUN apt-get update && apt-get upgrade -y && apt-get install -y --no-install-recommends apt-utils cron g++ make python
-RUN apt-get update
-RUN apt-get install software-properties-common wget nginx -y
-RUN apt-add-repository 'deb http://security.debian.org/debian-security stretch/updates main'
-RUN apt-get update
+# add pin script
+WORKDIR /
+ADD scripts/pinVersions.js scripts/cleanup.sh ./
+RUN chmod +x /cleanup.sh
+
+# build server
+WORKDIR /app
+ADD packages/server .
+RUN node /pinVersions.js && yarn && yarn build && /cleanup.sh
+
+# build worker
+WORKDIR /worker
+ADD packages/worker .
+RUN node /pinVersions.js && yarn && yarn build && /cleanup.sh
+
+FROM couchdb:3.2.1
+#Â TARGETARCH can be amd64 or arm e.g. docker build --build-arg TARGETARCH=amd64
+ARG TARGETARCH amd64
+#TARGETBUILD can be set to single (for single docker image) or aas (for azure app service)
+#Â e.g. docker build --build-arg TARGETBUILD=aas ....
+ARG TARGETBUILD single
+ENV TARGETBUILD $TARGETBUILD
+
+COPY --from=build /app /app
+COPY --from=build /worker /worker
+
+ENV \
+ APP_PORT=4001 \
+ ARCHITECTURE=amd \
+ BUDIBASE_ENVIRONMENT=PRODUCTION \
+ CLUSTER_PORT=80 \
+ COUCHDB_PASSWORD=budibase \
+ COUCHDB_USER=budibase \
+ COUCH_DB_URL=http://budibase:budibase@localhost:5984 \
+ # CUSTOM_DOMAIN=budi001.custom.com \
+ DEPLOYMENT_ENVIRONMENT=docker \
+ INTERNAL_API_KEY=budibase \
+ JWT_SECRET=testsecret \
+ MINIO_ACCESS_KEY=budibase \
+ MINIO_SECRET_KEY=budibase \
+ MINIO_URL=http://localhost:9000 \
+ POSTHOG_TOKEN=phc_fg5I3nDOf6oJVMHSaycEhpPdlgS8rzXG2r6F2IpxCHS \
+ REDIS_PASSWORD=budibase \
+ REDIS_URL=localhost:6379 \
+ SELF_HOSTED=1 \
+ TARGETBUILD=$TARGETBUILD \
+ WORKER_PORT=4002 \
+ WORKER_URL=http://localhost:4002
+
+# install base dependencies
+RUN apt-get update && \
+ apt-get install -y software-properties-common wget nginx && \
+ apt-add-repository 'deb http://security.debian.org/debian-security stretch/updates main' && \
+ apt-get update
+
+# install other dependencies, nodejs, oracle requirements, jdk8, redis, nginx
+WORKDIR /nodejs
+RUN curl -sL https://deb.nodesource.com/setup_16.x -o /tmp/nodesource_setup.sh && \
+ bash /tmp/nodesource_setup.sh && \
+ apt-get install -y libaio1 nodejs nginx openjdk-8-jdk redis-server unzip && \
+ npm install --global yarn pm2
# setup nginx
ADD hosting/single/nginx.conf /etc/nginx
-RUN mkdir /etc/nginx/logs
-RUN useradd www
-RUN touch /etc/nginx/logs/error.log
-RUN touch /etc/nginx/logs/nginx.pid
+ADD hosting/single/nginx-default-site.conf /etc/nginx/sites-enabled/default
+RUN mkdir -p /var/log/nginx && \
+ touch /var/log/nginx/error.log && \
+ touch /var/run/nginx.pid
-# install java
-RUN apt-get install openjdk-8-jdk -y
-
-# setup nodejs
-WORKDIR /nodejs
-RUN curl -sL https://deb.nodesource.com/setup_16.x -o /tmp/nodesource_setup.sh
-RUN bash /tmp/nodesource_setup.sh
-RUN apt-get install nodejs
-RUN npm install --global yarn
-RUN npm install --global pm2
-
-# setup redis
-RUN apt install redis-server -y
-
-# setup server
-WORKDIR /app
-ADD packages/server .
-RUN ls -al
-RUN yarn
-RUN yarn build
-# Install client for oracle datasource
-RUN apt-get install unzip libaio1
-RUN /bin/bash -e scripts/integrations/oracle/instantclient/linux/x86-64/install.sh
-
-# setup worker
-WORKDIR /worker
-ADD packages/worker .
-RUN yarn
-RUN yarn build
+WORKDIR /
+RUN mkdir -p scripts/integrations/oracle
+ADD packages/server/scripts/integrations/oracle scripts/integrations/oracle
+RUN /bin/bash -e ./scripts/integrations/oracle/instantclient/linux/install.sh
# setup clouseau
WORKDIR /
-RUN wget https://github.com/cloudant-labs/clouseau/releases/download/2.21.0/clouseau-2.21.0-dist.zip
-RUN unzip clouseau-2.21.0-dist.zip
-RUN mv clouseau-2.21.0 /opt/clouseau
-RUN rm clouseau-2.21.0-dist.zip
+RUN wget https://github.com/cloudant-labs/clouseau/releases/download/2.21.0/clouseau-2.21.0-dist.zip && \
+ unzip clouseau-2.21.0-dist.zip && \
+ mv clouseau-2.21.0 /opt/clouseau && \
+ rm clouseau-2.21.0-dist.zip
WORKDIR /opt/clouseau
RUN mkdir ./bin
ADD hosting/single/clouseau ./bin/
-ADD hosting/single/log4j.properties .
-ADD hosting/single/clouseau.ini .
+ADD hosting/single/log4j.properties hosting/single/clouseau.ini ./
RUN chmod +x ./bin/clouseau
# setup CouchDB
@@ -82,18 +96,49 @@ ADD hosting/single/vm.args ./etc/
# setup minio
WORKDIR /minio
-RUN wget https://dl.min.io/server/minio/release/linux-${ARCHITECTURE}64/minio
-RUN chmod +x minio
+ADD scripts/install-minio.sh ./install.sh
+RUN chmod +x install.sh && ./install.sh
# setup runner file
WORKDIR /
ADD hosting/single/runner.sh .
RUN chmod +x ./runner.sh
+ADD hosting/scripts/healthcheck.sh .
+RUN chmod +x ./healthcheck.sh
-EXPOSE 10000
+ADD hosting/scripts/build-target-paths.sh .
+RUN chmod +x ./build-target-paths.sh
+
+# For Azure App Service install SSH & point data locations to /home
+RUN /build-target-paths.sh
+
+# cleanup cache
+RUN yarn cache clean -f
+
+EXPOSE 80
+EXPOSE 443
VOLUME /opt/couchdb/data
VOLUME /minio
+# setup letsencrypt certificate
+RUN apt-get install -y certbot python3-certbot-nginx
+ADD hosting/letsencrypt /app/letsencrypt
+RUN chmod +x /app/letsencrypt/certificate-request.sh /app/letsencrypt/certificate-renew.sh
+# Remove cached files
+RUN rm -rf \
+ /root/.cache \
+ /root/.npm \
+ /root/.pip \
+ /usr/local/share/doc \
+ /usr/share/doc \
+ /usr/share/man \
+ /var/lib/apt/lists/* \
+ /tmp/*
+
+HEALTHCHECK --interval=15s --timeout=15s --start-period=45s CMD "/healthcheck.sh"
+
# must set this just before running
ENV NODE_ENV=production
+WORKDIR /
+
CMD ["./runner.sh"]
diff --git a/hosting/single/README.md b/hosting/single/README.md
new file mode 100644
index 0000000000..1147d55c89
--- /dev/null
+++ b/hosting/single/README.md
@@ -0,0 +1,112 @@
+# Docker Single Image for Budibase
+
+## Overview
+As an alternative to running several docker containers via docker-compose, the files under ./hosting/single can be used to build a docker image containing all of the Budibase components (minio, couch, clouseau etc).
+We call this the 'single image' container as the Dockerfile adds all the components to a single docker image.
+
+## Usage
+
+- Amend Environment Variables
+- Build Requirements
+- Build the Image
+- Run the Container
+
+### Amend Environment Variables
+
+Edit the Dockerfile in this directory amending the environment variables to suit your usage. Pay particular attention to changing passwords.
+The CUSTOM_DOMAIN variable will be used to request a certificate from LetsEncrypt and if successful you can point traffic to port 443. If you choose to use the CUSTOM_DOMAIN variable ensure that the DNS for your custom domain points to the public IP address where you are running Budibase - otherwise the certificate issuance will fail.
+If you have other arrangements for a proxy in front of the single image container you can omit the CUSTOM_DOMAIN environment variable and the request to LetsEncrypt will be skipped. You can then point traffic to port 80.
+
+### Build Requirements
+We would suggest building the image with 6GB of RAM and 20GB of free disk space for build artifacts. The resulting image size will use approx 2GB of disk space.
+
+### Build the Image
+The guidance below is based on building the Budibase single image on Debian 11 and AlmaLinux 8. If you use another distro or OS you will need to amend the commands to suit.
+#### Install Node
+Budibase requires a more recent version of node (14+) than is available in the base Debian repos so:
+
+```
+curl -sL https://deb.nodesource.com/setup_16.x | sudo bash -
+apt install -y nodejs
+node -v
+```
+Install yarn and lerna:
+```
+npm install -g yarn jest lerna
+```
+#### Install Docker
+
+```
+apt install -y docker.io
+```
+
+Check the versions of each installed version. This process was tested with the version numbers below so YMMV using anything else:
+
+- Docker: 20.10.5
+- node: 16.15.1
+- yarn: 1.22.19
+- lerna: 5.1.4
+
+#### Get the Code
+Clone the Budibase repo
+```
+git clone https://github.com/Budibase/budibase.git
+cd budibase
+```
+#### Setup Node
+Node setup:
+```
+node ./hosting/scripts/setup.js
+yarn
+yarn bootstrap
+yarn build
+```
+#### Build Image
+The following yarn command does some prep and then runs the docker build command:
+```
+yarn build:docker:single
+```
+If the docker build step fails try running that step again manually with:
+```
+docker build --build-arg TARGETARCH=amd --no-cache -t budibase:latest -f ./hosting/single/Dockerfile .
+```
+
+#### Azure App Services
+Azure have some specific requirements for running a container in their App Service. Specifically, installation of SSH to port 2222 and data storage under /home. If you would like to build a budibase container for Azure App Service add the build argument shown below setting it to 'aas'. You can remove the CUSTOM_DOMAIN env variable from the Dockerfile too as Azure terminate SSL before requests reach the container.
+```
+docker build --build-arg TARGETARCH=amd --build-arg TARGETBUILD=aas -t budibase:latest -f ./hosting/single/Dockerfile .
+```
+
+### Run the Container
+```
+docker run -d -p 80:80 -p 443:443 --name budibase budibase:latest
+```
+Where:
+- -d runs the container in detached mode
+- -p forwards ports from your host to the ports inside the container. If you are already using port 80 on your host for something else you can try running with an alternative port e.g. `-p 8080:80`
+- --name is the name for the container as shown in `docker ps` and can be used with other docker commands e.g. `docker restart budibase`
+
+When the container runs you should be able to access the container over http at your host address e.g. http://1.2.3.4/ or using your custom domain e.g. https://my.custom.domain/
+
+When the Budibase UI appears you will be prompted to create an account to get started.
+
+### Podman
+The single image container builds fine when using podman in place of docker. You may be prompted for the registry to use for the CouchDB image and the HEALTHCHECK parameter is not OCI compliant so is ignored.
+
+### Check
+There are many things that could go wrong so if your container is not building or running as expected please check the following before opening a support issue.
+Verify the healthcheck status of the container:
+```
+docker ps
+```
+Check the container logs:
+```
+docker logs budibase
+```
+
+### Support
+This single image build is still a work-in-progress so if you open an issue please provide the following information:
+- The OS and OS version you are building on
+- The versions you are using of docker, docker-compose, yarn, node, lerna
+- For build errors please provide zipped output
+- For container errors please provide zipped container logs
diff --git a/hosting/single/nginx-default-site.conf b/hosting/single/nginx-default-site.conf
new file mode 100644
index 0000000000..964313fa73
--- /dev/null
+++ b/hosting/single/nginx-default-site.conf
@@ -0,0 +1,94 @@
+server {
+ listen 80 default_server;
+ listen [::]:80 default_server;
+ server_name _;
+
+ client_max_body_size 1000m;
+ ignore_invalid_headers off;
+ proxy_buffering off;
+ # port_in_redirect off;
+
+ location ^~ /.well-known/acme-challenge/ {
+ default_type "text/plain";
+ root /var/www/html;
+ break;
+ }
+ location = /.well-known/acme-challenge/ {
+ return 404;
+ }
+
+ location /app {
+ proxy_pass http://127.0.0.1:4001;
+ }
+
+ location = / {
+ proxy_pass http://127.0.0.1:4001;
+ }
+
+ location ~ ^/(builder|app_) {
+ proxy_http_version 1.1;
+ proxy_set_header Connection $connection_upgrade;
+ proxy_set_header Upgrade $http_upgrade;
+ proxy_set_header X-Real-IP $remote_addr;
+ proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
+ proxy_pass http://127.0.0.1:4001;
+ }
+
+ location ~ ^/api/(system|admin|global)/ {
+ proxy_pass http://127.0.0.1:4002;
+ }
+
+ location /worker/ {
+ proxy_pass http://127.0.0.1:4002;
+ rewrite ^/worker/(.*)$ /$1 break;
+ }
+
+ location /api/ {
+ # calls to the API are rate limited with bursting
+ limit_req zone=ratelimit burst=20 nodelay;
+
+ # 120s timeout on API requests
+ proxy_read_timeout 120s;
+ proxy_connect_timeout 120s;
+ proxy_send_timeout 120s;
+
+ proxy_http_version 1.1;
+ proxy_set_header Connection $connection_upgrade;
+ proxy_set_header Upgrade $http_upgrade;
+ proxy_set_header X-Real-IP $remote_addr;
+ proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
+
+ proxy_pass http://127.0.0.1:4001;
+ }
+
+ location /db/ {
+ proxy_pass http://127.0.0.1:5984;
+ rewrite ^/db/(.*)$ /$1 break;
+ }
+
+ location / {
+ proxy_set_header X-Real-IP $remote_addr;
+ proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
+ proxy_set_header X-Forwarded-Proto $scheme;
+
+ proxy_connect_timeout 300;
+ proxy_http_version 1.1;
+ proxy_set_header Connection "";
+ chunked_transfer_encoding off;
+ proxy_pass http://127.0.0.1:9000;
+ }
+
+ client_header_timeout 60;
+ client_body_timeout 60;
+ keepalive_timeout 60;
+
+ # gzip
+ gzip on;
+ gzip_vary on;
+ gzip_proxied any;
+ gzip_comp_level 6;
+ gzip_types text/plain text/css text/xml application/json application/javascript application/rss+xml application/atom+xml image/svg+xml;
+
+
+
+}
diff --git a/hosting/single/nginx.conf b/hosting/single/nginx.conf
index 86938ced4e..1e5d1c20d2 100644
--- a/hosting/single/nginx.conf
+++ b/hosting/single/nginx.conf
@@ -1,6 +1,6 @@
-user www www;
-error_log /etc/nginx/logs/error.log;
-pid /etc/nginx/logs/nginx.pid;
+user www-data www-data;
+error_log /var/log/nginx/error.log;
+pid /var/run/nginx.pid;
worker_processes auto;
worker_rlimit_nofile 8192;
@@ -32,85 +32,6 @@ http {
default "upgrade";
}
- server {
- listen 10000 default_server;
- listen [::]:10000 default_server;
- server_name _;
- client_max_body_size 1000m;
- ignore_invalid_headers off;
- proxy_buffering off;
- # port_in_redirect off;
+ include /etc/nginx/sites-enabled/*;
- location /app {
- proxy_pass http://127.0.0.1:4001;
- }
-
- location = / {
- proxy_pass http://127.0.0.1:4001;
- }
-
- location ~ ^/(builder|app_) {
- proxy_http_version 1.1;
- proxy_set_header Connection $connection_upgrade;
- proxy_set_header Upgrade $http_upgrade;
- proxy_set_header X-Real-IP $remote_addr;
- proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
- proxy_pass http://127.0.0.1:4001;
- }
-
- location ~ ^/api/(system|admin|global)/ {
- proxy_pass http://127.0.0.1:4002;
- }
-
- location /worker/ {
- proxy_pass http://127.0.0.1:4002;
- rewrite ^/worker/(.*)$ /$1 break;
- }
-
- location /api/ {
- # calls to the API are rate limited with bursting
- limit_req zone=ratelimit burst=20 nodelay;
-
- # 120s timeout on API requests
- proxy_read_timeout 120s;
- proxy_connect_timeout 120s;
- proxy_send_timeout 120s;
-
- proxy_http_version 1.1;
- proxy_set_header Connection $connection_upgrade;
- proxy_set_header Upgrade $http_upgrade;
- proxy_set_header X-Real-IP $remote_addr;
- proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
-
- proxy_pass http://127.0.0.1:4001;
- }
-
- location /db/ {
- proxy_pass http://127.0.0.1:5984;
- rewrite ^/db/(.*)$ /$1 break;
- }
-
- location / {
- proxy_set_header X-Real-IP $remote_addr;
- proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
- proxy_set_header X-Forwarded-Proto $scheme;
-
- proxy_connect_timeout 300;
- proxy_http_version 1.1;
- proxy_set_header Connection "";
- chunked_transfer_encoding off;
- proxy_pass http://127.0.0.1:9000;
- }
-
- client_header_timeout 60;
- client_body_timeout 60;
- keepalive_timeout 60;
-
- # gzip
- gzip on;
- gzip_vary on;
- gzip_proxied any;
- gzip_comp_level 6;
- gzip_types text/plain text/css text/xml application/json application/javascript application/rss+xml application/atom+xml image/svg+xml;
- }
}
diff --git a/hosting/single/runner.sh b/hosting/single/runner.sh
index fab8431796..6f3d247842 100644
--- a/hosting/single/runner.sh
+++ b/hosting/single/runner.sh
@@ -2,6 +2,15 @@ redis-server --requirepass $REDIS_PASSWORD &
/opt/clouseau/bin/clouseau &
/minio/minio server /minio &
/docker-entrypoint.sh /opt/couchdb/bin/couchdb &
+/etc/init.d/nginx restart
+if [[ ! -z "${CUSTOM_DOMAIN}" ]]; then
+ # Add monthly cron job to renew certbot certificate
+ echo -n "* * 2 * * root exec /app/letsencrypt/certificate-renew.sh ${CUSTOM_DOMAIN}" >> /etc/cron.d/certificate-renew
+ chmod +x /etc/cron.d/certificate-renew
+ # Request the certbot certificate
+ /app/letsencrypt/certificate-request.sh ${CUSTOM_DOMAIN}
+fi
+
/etc/init.d/nginx restart
pushd app
pm2 start --name app "yarn run:docker"
@@ -10,7 +19,6 @@ pushd worker
pm2 start --name worker "yarn run:docker"
popd
sleep 10
-URL=http://${COUCHDB_USER}:${COUCHDB_PASSWORD}@localhost:5984
-curl -X PUT ${URL}/_users
-curl -X PUT ${URL}/_replicator
-sleep infinity
\ No newline at end of file
+curl -X PUT ${COUCH_DB_URL}/_users
+curl -X PUT ${COUCH_DB_URL}/_replicator
+sleep infinity
diff --git a/hosting/single/test.sh b/hosting/single/test.sh
new file mode 100755
index 0000000000..c7ef53f994
--- /dev/null
+++ b/hosting/single/test.sh
@@ -0,0 +1,4 @@
+#!/bin/bash
+id=$(docker run -t -d -p 80:80 budibase:latest)
+docker exec -it $id bash
+docker kill $id
diff --git a/lerna.json b/lerna.json
index 080372bc14..e65ce20401 100644
--- a/lerna.json
+++ b/lerna.json
@@ -1,5 +1,5 @@
{
- "version": "1.0.206",
+ "version": "1.0.212-alpha.6",
"npmClient": "yarn",
"packages": [
"packages/*"
diff --git a/package.json b/package.json
index 014cafcfc1..9c35af497f 100644
--- a/package.json
+++ b/package.json
@@ -62,6 +62,7 @@
"build:docker:develop": "node scripts/pinVersions && lerna run build:docker && npm run build:docker:proxy:compose && cd hosting/scripts/linux/ && ./release-to-docker-hub.sh develop && cd -",
"build:docker:airgap": "node hosting/scripts/airgapped/airgappedDockerBuild",
"build:digitalocean": "cd hosting/digitalocean && ./build.sh && cd -",
+ "build:docker:single:multiarch": "docker buildx build --platform linux/arm64,linux/amd64 -f hosting/single/Dockerfile -t budibase:latest .",
"build:docker:single:image": "docker build -f hosting/single/Dockerfile -t budibase:latest .",
"build:docker:single": "lerna run build && lerna run predocker && npm run build:docker:single:image",
"build:docs": "lerna run build:docs",
diff --git a/packages/backend-core/cache.js b/packages/backend-core/cache.js
index 932fd7b901..6b319357c4 100644
--- a/packages/backend-core/cache.js
+++ b/packages/backend-core/cache.js
@@ -3,5 +3,6 @@ const generic = require("./src/cache/generic")
module.exports = {
user: require("./src/cache/user"),
app: require("./src/cache/appMetadata"),
+ writethrough: require("./src/cache/writethrough"),
...generic,
}
diff --git a/packages/backend-core/package.json b/packages/backend-core/package.json
index 04d50eff1a..3f36482120 100644
--- a/packages/backend-core/package.json
+++ b/packages/backend-core/package.json
@@ -1,6 +1,6 @@
{
"name": "@budibase/backend-core",
- "version": "1.0.206",
+ "version": "1.0.212-alpha.6",
"description": "Budibase backend core libraries used in server and worker",
"main": "dist/src/index.js",
"types": "dist/src/index.d.ts",
@@ -20,6 +20,7 @@
"test:watch": "jest --watchAll"
},
"dependencies": {
+ "@budibase/types": "^1.0.212-alpha.6",
"@techpass/passport-openidconnect": "0.3.2",
"aws-sdk": "2.1030.0",
"bcrypt": "5.0.1",
@@ -57,12 +58,12 @@
]
},
"devDependencies": {
- "@budibase/types": "^1.0.206",
"@shopify/jest-koa-mocks": "3.1.5",
"@types/jest": "27.5.1",
"@types/koa": "2.0.52",
"@types/node": "14.18.20",
"@types/node-fetch": "2.6.1",
+ "@types/pouchdb": "6.4.0",
"@types/redlock": "4.0.3",
"@types/semver": "7.3.7",
"@types/tar-fs": "2.0.1",
diff --git a/packages/backend-core/redis.js b/packages/backend-core/redis.js
index badce3be20..1f7a48540a 100644
--- a/packages/backend-core/redis.js
+++ b/packages/backend-core/redis.js
@@ -1,5 +1,5 @@
module.exports = {
Client: require("./src/redis"),
utils: require("./src/redis/utils"),
- clients: require("./src/redis/authRedis"),
+ clients: require("./src/redis/init"),
}
diff --git a/packages/backend-core/src/cache/appMetadata.js b/packages/backend-core/src/cache/appMetadata.js
index effdc886d7..b0d9481cbd 100644
--- a/packages/backend-core/src/cache/appMetadata.js
+++ b/packages/backend-core/src/cache/appMetadata.js
@@ -1,4 +1,4 @@
-const redis = require("../redis/authRedis")
+const redis = require("../redis/init")
const { doWithDB } = require("../db")
const { DocumentTypes } = require("../db/constants")
diff --git a/packages/backend-core/src/cache/base/index.ts b/packages/backend-core/src/cache/base/index.ts
new file mode 100644
index 0000000000..f3216531f4
--- /dev/null
+++ b/packages/backend-core/src/cache/base/index.ts
@@ -0,0 +1,92 @@
+import { getTenantId } from "../../context"
+import redis from "../../redis/init"
+import RedisWrapper from "../../redis"
+
+function generateTenantKey(key: string) {
+ const tenantId = getTenantId()
+ return `${key}:${tenantId}`
+}
+
+export = class BaseCache {
+ client: RedisWrapper | undefined
+
+ constructor(client: RedisWrapper | undefined = undefined) {
+ this.client = client
+ }
+
+ async getClient() {
+ return !this.client ? await redis.getCacheClient() : this.client
+ }
+
+ async keys(pattern: string) {
+ const client = await this.getClient()
+ return client.keys(pattern)
+ }
+
+ /**
+ * Read only from the cache.
+ */
+ async get(key: string, opts = { useTenancy: true }) {
+ key = opts.useTenancy ? generateTenantKey(key) : key
+ const client = await this.getClient()
+ return client.get(key)
+ }
+
+ /**
+ * Write to the cache.
+ */
+ async store(
+ key: string,
+ value: any,
+ ttl: number | null = null,
+ opts = { useTenancy: true }
+ ) {
+ key = opts.useTenancy ? generateTenantKey(key) : key
+ const client = await this.getClient()
+ await client.store(key, value, ttl)
+ }
+
+ /**
+ * Remove from cache.
+ */
+ async delete(key: string, opts = { useTenancy: true }) {
+ key = opts.useTenancy ? generateTenantKey(key) : key
+ const client = await this.getClient()
+ return client.delete(key)
+ }
+
+ /**
+ * Read from the cache. Write to the cache if not exists.
+ */
+ async withCache(
+ key: string,
+ ttl: number,
+ fetchFn: any,
+ opts = { useTenancy: true }
+ ) {
+ const cachedValue = await this.get(key, opts)
+ if (cachedValue) {
+ return cachedValue
+ }
+
+ try {
+ const fetchedValue = await fetchFn()
+
+ await this.store(key, fetchedValue, ttl, opts)
+ return fetchedValue
+ } catch (err) {
+ console.error("Error fetching before cache - ", err)
+ throw err
+ }
+ }
+
+ async bustCache(key: string, opts = { client: null }) {
+ const client = await this.getClient()
+ try {
+ await client.delete(generateTenantKey(key))
+ } catch (err) {
+ console.error("Error busting cache - ", err)
+ throw err
+ }
+ }
+}
diff --git a/packages/backend-core/src/cache/generic.js b/packages/backend-core/src/cache/generic.js
index 8f0ef78954..e2f3915339 100644
--- a/packages/backend-core/src/cache/generic.js
+++ b/packages/backend-core/src/cache/generic.js
@@ -1,5 +1,6 @@
-const redis = require("../redis/authRedis")
-const { getTenantId } = require("../context")
+const BaseCache = require("./base")
+
+const GENERIC = new BaseCache()
exports.CacheKeys = {
CHECKLIST: "checklist",
@@ -16,67 +17,13 @@ exports.TTL = {
ONE_DAY: 86400,
}
-function generateTenantKey(key) {
- const tenantId = getTenantId()
- return `${key}:${tenantId}`
+function performExport(funcName) {
+ return (...args) => GENERIC[funcName](...args)
}
-exports.keys = async pattern => {
- const client = await redis.getCacheClient()
- return client.keys(pattern)
-}
-
-/**
- * Read only from the cache.
- */
-exports.get = async (key, opts = { useTenancy: true }) => {
- key = opts.useTenancy ? generateTenantKey(key) : key
- const client = await redis.getCacheClient()
- const value = await client.get(key)
- return value
-}
-
-/**
- * Write to the cache.
- */
-exports.store = async (key, value, ttl, opts = { useTenancy: true }) => {
- key = opts.useTenancy ? generateTenantKey(key) : key
- const client = await redis.getCacheClient()
- await client.store(key, value, ttl)
-}
-
-exports.delete = async (key, opts = { useTenancy: true }) => {
- key = opts.useTenancy ? generateTenantKey(key) : key
- const client = await redis.getCacheClient()
- return client.delete(key)
-}
-
-/**
- * Read from the cache. Write to the cache if not exists.
- */
-exports.withCache = async (key, ttl, fetchFn, opts = { useTenancy: true }) => {
- const cachedValue = await exports.get(key, opts)
- if (cachedValue) {
- return cachedValue
- }
-
- try {
- const fetchedValue = await fetchFn()
-
- await exports.store(key, fetchedValue, ttl, opts)
- return fetchedValue
- } catch (err) {
- console.error("Error fetching before cache - ", err)
- throw err
- }
-}
-
-exports.bustCache = async key => {
- const client = await redis.getCacheClient()
- try {
- await client.delete(generateTenantKey(key))
- } catch (err) {
- console.error("Error busting cache - ", err)
- throw err
- }
-}
+exports.keys = performExport("keys")
+exports.get = performExport("get")
+exports.store = performExport("store")
+exports.delete = performExport("delete")
+exports.withCache = performExport("withCache")
+exports.bustCache = performExport("bustCache")
diff --git a/packages/backend-core/src/cache/tests/writethrough.spec.js b/packages/backend-core/src/cache/tests/writethrough.spec.js
new file mode 100644
index 0000000000..68db24b325
--- /dev/null
+++ b/packages/backend-core/src/cache/tests/writethrough.spec.js
@@ -0,0 +1,59 @@
+require("../../../tests/utilities/TestConfiguration")
+const { Writethrough } = require("../writethrough")
+const { dangerousGetDB } = require("../../db")
+const tk = require("timekeeper")
+
+const START_DATE = Date.now()
+tk.freeze(START_DATE)
+
+const DELAY = 5000
+
+const db = dangerousGetDB("test")
+const db2 = dangerousGetDB("test2")
+const writethrough = new Writethrough(db, DELAY), writethrough2 = new Writethrough(db2, DELAY)
+
+describe("writethrough", () => {
+ describe("put", () => {
+ let first
+ it("should be able to store, will go to DB", async () => {
+ const response = await writethrough.put({ _id: "test", value: 1 })
+ const output = await db.get(response.id)
+ first = output
+ expect(output.value).toBe(1)
+ })
+
+ it("second put shouldn't update DB", async () => {
+ const response = await writethrough.put({ ...first, value: 2 })
+ const output = await db.get(response.id)
+ expect(first._rev).toBe(output._rev)
+ expect(output.value).toBe(1)
+ })
+
+ it("should put it again after delay period", async () => {
+ tk.freeze(START_DATE + DELAY + 1)
+ const response = await writethrough.put({ ...first, value: 3 })
+ const output = await db.get(response.id)
+ expect(response.rev).not.toBe(first._rev)
+ expect(output.value).toBe(3)
+ })
+ })
+
+ describe("get", () => {
+ it("should be able to retrieve", async () => {
+ const response = await writethrough.get("test")
+ expect(response.value).toBe(3)
+ })
+ })
+
+ describe("same doc, different databases (tenancy)", () => {
+ it("should be able to two different databases", async () => {
+ const resp1 = await writethrough.put({ _id: "db1", value: "first" })
+ const resp2 = await writethrough2.put({ _id: "db1", value: "second" })
+ expect(resp1.rev).toBeDefined()
+ expect(resp2.rev).toBeDefined()
+ expect((await db.get("db1")).value).toBe("first")
+ expect((await db2.get("db1")).value).toBe("second")
+ })
+ })
+})
+
diff --git a/packages/backend-core/src/cache/user.js b/packages/backend-core/src/cache/user.js
index faac6de725..130da1915e 100644
--- a/packages/backend-core/src/cache/user.js
+++ b/packages/backend-core/src/cache/user.js
@@ -1,4 +1,4 @@
-const redis = require("../redis/authRedis")
+const redis = require("../redis/init")
const { getTenantId, lookupTenantId, doWithGlobalDB } = require("../tenancy")
const env = require("../environment")
const accounts = require("../cloud/accounts")
diff --git a/packages/backend-core/src/cache/writethrough.ts b/packages/backend-core/src/cache/writethrough.ts
new file mode 100644
index 0000000000..e11ca0acaa
--- /dev/null
+++ b/packages/backend-core/src/cache/writethrough.ts
@@ -0,0 +1,120 @@
+import BaseCache from "./base"
+import { getWritethroughClient } from "../redis/init"
+
+const DEFAULT_WRITE_RATE_MS = 10000
+let CACHE: BaseCache | null = null
+
+interface CacheItem {
+ doc: any
+ lastWrite: number
+}
+
+async function getCache() {
+ if (!CACHE) {
+ const client = await getWritethroughClient()
+ CACHE = new BaseCache(client)
+ }
+ return CACHE
+}
+
+function makeCacheKey(db: PouchDB.Database, key: string) {
+ return db.name + key
+}
+
+function makeCacheItem(doc: any, lastWrite: number | null = null): CacheItem {
+ return { doc, lastWrite: lastWrite || Date.now() }
+}
+
+export async function put(
+ db: PouchDB.Database,
+ doc: any,
+ writeRateMs: number = DEFAULT_WRITE_RATE_MS
+) {
+ const cache = await getCache()
+ const key = doc._id
+ let cacheItem: CacheItem | undefined = await cache.get(makeCacheKey(db, key))
+ const updateDb = !cacheItem || cacheItem.lastWrite < Date.now() - writeRateMs
+ let output = doc
+ if (updateDb) {
+ const writeDb = async (toWrite: any) => {
+ // doc should contain the _id and _rev
+ const response = await db.put(toWrite)
+ output = {
+ ...doc,
+ _id: response.id,
+ _rev: response.rev,
+ }
+ }
+ try {
+ await writeDb(doc)
+ } catch (err: any) {
+ if (err.status !== 409) {
+ throw err
+ } else {
+ // get the rev, update over it - this is risky, may change in future
+ const readDoc = await db.get(doc._id)
+ doc._rev = readDoc._rev
+ await writeDb(doc)
+ }
+ }
+ }
+ // if we are updating the DB then need to set the lastWrite to now
+ cacheItem = makeCacheItem(output, updateDb ? null : cacheItem?.lastWrite)
+ await cache.store(makeCacheKey(db, key), cacheItem)
+ return { ok: true, id: output._id, rev: output._rev }
+}
+
+export async function get(db: PouchDB.Database, id: string): Promise {
+ const cache = await getCache()
+ const cacheKey = makeCacheKey(db, id)
+ let cacheItem: CacheItem = await cache.get(cacheKey)
+ if (!cacheItem) {
+ const doc = await db.get(id)
+ cacheItem = makeCacheItem(doc)
+ await cache.store(cacheKey, cacheItem)
+ }
+ return cacheItem.doc
+}
+
+export async function remove(
+ db: PouchDB.Database,
+ docOrId: any,
+ rev?: any
+): Promise {
+ const cache = await getCache()
+ if (!docOrId) {
+ throw new Error("No ID/Rev provided.")
+ }
+ const id = typeof docOrId === "string" ? docOrId : docOrId._id
+ rev = typeof docOrId === "string" ? rev : docOrId._rev
+ try {
+ await cache.delete(makeCacheKey(db, id))
+ } finally {
+ await db.remove(id, rev)
+ }
+}
+
+export class Writethrough {
+ db: PouchDB.Database
+ writeRateMs: number
+
+ constructor(
+ db: PouchDB.Database,
+ writeRateMs: number = DEFAULT_WRITE_RATE_MS
+ ) {
+ this.db = db
+ this.writeRateMs = writeRateMs
+ }
+
+ async put(doc: any) {
+ return put(this.db, doc, this.writeRateMs)
+ }
+
+ async get(id: string) {
+ return get(this.db, id)
+ }
+
+ async remove(docOrId: any, rev?: any) {
+ return remove(this.db, docOrId, rev)
+ }
+}
diff --git a/packages/backend-core/src/db/pouch.js b/packages/backend-core/src/db/pouch.js
index 76390ac644..59b7ff8ae7 100644
--- a/packages/backend-core/src/db/pouch.js
+++ b/packages/backend-core/src/db/pouch.js
@@ -1,21 +1,42 @@
const PouchDB = require("pouchdb")
const env = require("../environment")
-function getUrlInfo() {
- let url = env.COUCH_DB_URL
- let username, password, host
- const [protocol, rest] = url.split("://")
- if (url.includes("@")) {
- const hostParts = rest.split("@")
- host = hostParts[1]
- const authParts = hostParts[0].split(":")
- username = authParts[0]
- password = authParts[1]
- } else {
- host = rest
+exports.getUrlInfo = (url = env.COUCH_DB_URL) => {
+ let cleanUrl, username, password, host
+ if (url) {
+ // Ensure the URL starts with a protocol
+ const protoRegex = /^https?:\/\//i
+ if (!protoRegex.test(url)) {
+ url = `http://${url}`
+ }
+
+ // Split into protocol and remainder
+ const split = url.split("://")
+ const protocol = split[0]
+ const rest = split.slice(1).join("://")
+
+ // Extract auth if specified
+ if (url.includes("@")) {
+ // Split into host and remainder
+ let parts = rest.split("@")
+ host = parts[parts.length - 1]
+ let auth = parts.slice(0, -1).join("@")
+
+ // Split auth into username and password
+ if (auth.includes(":")) {
+ const authParts = auth.split(":")
+ username = authParts[0]
+ password = authParts.slice(1).join(":")
+ } else {
+ username = auth
+ }
+ } else {
+ host = rest
+ }
+ cleanUrl = `${protocol}://${host}`
}
return {
- url: `${protocol}://${host}`,
+ url: cleanUrl,
auth: {
username,
password,
@@ -24,7 +45,7 @@ function getUrlInfo() {
}
exports.getCouchInfo = () => {
- const urlInfo = getUrlInfo()
+ const urlInfo = exports.getUrlInfo()
let username
let password
if (env.COUCH_DB_USERNAME) {
diff --git a/packages/backend-core/src/db/tests/pouch.spec.js b/packages/backend-core/src/db/tests/pouch.spec.js
new file mode 100644
index 0000000000..30cdd0f5ec
--- /dev/null
+++ b/packages/backend-core/src/db/tests/pouch.spec.js
@@ -0,0 +1,62 @@
+require("../../../tests/utilities/TestConfiguration")
+const getUrlInfo = require("../pouch").getUrlInfo
+
+describe("pouch", () => {
+ describe("Couch DB URL parsing", () => {
+ it("should handle a null Couch DB URL", () => {
+ const info = getUrlInfo(null)
+ expect(info.url).toBeUndefined()
+ expect(info.auth.username).toBeUndefined()
+ })
+ it("should be able to parse a basic Couch DB URL", () => {
+ const info = getUrlInfo("http://host.com")
+ expect(info.url).toBe("http://host.com")
+ expect(info.auth.username).toBeUndefined()
+ })
+ it("should be able to parse a Couch DB basic URL with HTTPS", () => {
+ const info = getUrlInfo("https://host.com")
+ expect(info.url).toBe("https://host.com")
+ expect(info.auth.username).toBeUndefined()
+ })
+ it("should be able to parse a basic Couch DB URL with a custom port", () => {
+ const info = getUrlInfo("https://host.com:1234")
+ expect(info.url).toBe("https://host.com:1234")
+ expect(info.auth.username).toBeUndefined()
+ })
+ it("should be able to parse a Couch DB URL with auth", () => {
+ const info = getUrlInfo("https://user:pass@host.com:1234")
+ expect(info.url).toBe("https://host.com:1234")
+ expect(info.auth.username).toBe("user")
+ expect(info.auth.password).toBe("pass")
+ })
+ it("should be able to parse a Couch DB URL with auth and special chars", () => {
+ const info = getUrlInfo("https://user:s:p@s://@://:d@;][~s@host.com:1234")
+ expect(info.url).toBe("https://host.com:1234")
+ expect(info.auth.username).toBe("user")
+ expect(info.auth.password).toBe("s:p@s://@://:d@;][~s")
+ })
+ it("should be able to parse a Couch DB URL without a protocol", () => {
+ const info = getUrlInfo("host.com:1234")
+ expect(info.url).toBe("http://host.com:1234")
+ expect(info.auth.username).toBeUndefined()
+ })
+ it("should be able to parse a Couch DB URL with auth and without a protocol", () => {
+ const info = getUrlInfo("user:s:p@s://@://:d@;][~s@host.com:1234")
+ expect(info.url).toBe("http://host.com:1234")
+ expect(info.auth.username).toBe("user")
+ expect(info.auth.password).toBe("s:p@s://@://:d@;][~s")
+ })
+ it("should be able to parse a Couch DB URL with only username auth", () => {
+ const info = getUrlInfo("https://user@host.com:1234")
+ expect(info.url).toBe("https://host.com:1234")
+ expect(info.auth.username).toBe("user")
+ expect(info.auth.password).toBeUndefined()
+ })
+ it("should be able to parse a Couch DB URL with only username auth and without a protocol", () => {
+ const info = getUrlInfo("user@host.com:1234")
+ expect(info.url).toBe("http://host.com:1234")
+ expect(info.auth.username).toBe("user")
+ expect(info.auth.password).toBeUndefined()
+ })
+ })
+})
diff --git a/packages/backend-core/src/environment.ts b/packages/backend-core/src/environment.ts
index 1d9febc6f9..0a17c82873 100644
--- a/packages/backend-core/src/environment.ts
+++ b/packages/backend-core/src/environment.ts
@@ -16,7 +16,7 @@ if (!LOADED && isDev() && !isTest()) {
LOADED = true
}
-const env: any = {
+const env = {
isTest,
isDev,
JWT_SECRET: process.env.JWT_SECRET,
@@ -66,6 +66,7 @@ const env: any = {
for (let [key, value] of Object.entries(env)) {
// handle the edge case of "0" to disable an environment variable
if (value === "0") {
+ // @ts-ignore
env[key] = 0
}
}
diff --git a/packages/backend-core/src/objectStore/index.ts b/packages/backend-core/src/objectStore/index.ts
index 9bb0760f5b..a7e0b0c134 100644
--- a/packages/backend-core/src/objectStore/index.ts
+++ b/packages/backend-core/src/objectStore/index.ts
@@ -294,6 +294,16 @@ export const uploadDirectory = async (
await Promise.all(uploads)
}
+exports.downloadTarballDirect = async (url: string, path: string) => {
+ path = sanitizeKey(path)
+ const response = await fetch(url)
+ if (!response.ok) {
+ throw new Error(`unexpected response ${response.statusText}`)
+ }
+
+ await streamPipeline(response.body, zlib.Unzip(), tar.extract(path))
+}
+
export const downloadTarball = async (url: any, bucketName: any, path: any) => {
bucketName = sanitizeBucket(bucketName)
path = sanitizeKey(path)
diff --git a/packages/backend-core/src/pkg/redis.ts b/packages/backend-core/src/pkg/redis.ts
index caa5526e55..65ab186d9a 100644
--- a/packages/backend-core/src/pkg/redis.ts
+++ b/packages/backend-core/src/pkg/redis.ts
@@ -2,7 +2,7 @@
// The outer exports can't be used as they now reference dist directly
import Client from "../redis"
import utils from "../redis/utils"
-import clients from "../redis/authRedis"
+import clients from "../redis/init"
export = {
Client,
diff --git a/packages/backend-core/src/redis/index.js b/packages/backend-core/src/redis/index.ts
similarity index 55%
rename from packages/backend-core/src/redis/index.js
rename to packages/backend-core/src/redis/index.ts
index bd35f95916..206110366f 100644
--- a/packages/backend-core/src/redis/index.js
+++ b/packages/backend-core/src/redis/index.ts
@@ -1,3 +1,4 @@
+import RedisWrapper from "../redis"
const env = require("../environment")
// ioredis mock is all in memory
const Redis = env.isTest() ? require("ioredis-mock") : require("ioredis")
@@ -6,24 +7,34 @@ const {
removeDbPrefix,
getRedisOptions,
SEPARATOR,
+ SelectableDatabases,
} = require("./utils")
const RETRY_PERIOD_MS = 2000
const STARTUP_TIMEOUT_MS = 5000
const CLUSTERED = false
+const DEFAULT_SELECT_DB = SelectableDatabases.DEFAULT
// for testing just generate the client once
let CLOSED = false
-let CLIENT = env.isTest() ? new Redis(getRedisOptions()) : null
+let CLIENTS: { [key: number]: any } = {}
// if in test always connected
-let CONNECTED = !!env.isTest()
+let CONNECTED = env.isTest()
-function connectionError(timeout, err) {
+function pickClient(selectDb: number): any {
+ return CLIENTS[selectDb]
+}
+
+function connectionError(
+ selectDb: number,
+ timeout: NodeJS.Timeout,
+ err: Error | string
+) {
// manually shut down, ignore errors
if (CLOSED) {
return
}
- CLIENT.disconnect()
+ pickClient(selectDb).disconnect()
CLOSED = true
// always clear this on error
clearTimeout(timeout)
@@ -38,59 +49,69 @@ function connectionError(timeout, err) {
* Inits the system, will error if unable to connect to redis cluster (may take up to 10 seconds) otherwise
* will return the ioredis client which will be ready to use.
*/
-function init() {
- let timeout
+function init(selectDb = DEFAULT_SELECT_DB) {
+ let timeout: NodeJS.Timeout
CLOSED = false
- // testing uses a single in memory client
- if (env.isTest() || (CLIENT && CONNECTED)) {
+ let client = pickClient(selectDb)
+ // already connected, ignore
+ if (client && CONNECTED) {
return
}
+ // testing uses a single in memory client
+ if (env.isTest()) {
+ CLIENTS[selectDb] = new Redis(getRedisOptions())
+ }
// start the timer - only allowed 5 seconds to connect
timeout = setTimeout(() => {
if (!CONNECTED) {
- connectionError(timeout, "Did not successfully connect in timeout")
+ connectionError(
+ selectDb,
+ timeout,
+ "Did not successfully connect in timeout"
+ )
}
}, STARTUP_TIMEOUT_MS)
// disconnect any lingering client
- if (CLIENT) {
- CLIENT.disconnect()
+ if (client) {
+ client.disconnect()
}
const { redisProtocolUrl, opts, host, port } = getRedisOptions(CLUSTERED)
if (CLUSTERED) {
- CLIENT = new Redis.Cluster([{ host, port }], opts)
+ client = new Redis.Cluster([{ host, port }], opts)
} else if (redisProtocolUrl) {
- CLIENT = new Redis(redisProtocolUrl)
+ client = new Redis(redisProtocolUrl)
} else {
- CLIENT = new Redis(opts)
+ client = new Redis(opts)
}
// attach handlers
- CLIENT.on("end", err => {
- connectionError(timeout, err)
+ client.on("end", (err: Error) => {
+ connectionError(selectDb, timeout, err)
})
- CLIENT.on("error", err => {
- connectionError(timeout, err)
+ client.on("error", (err: Error) => {
+ connectionError(selectDb, timeout, err)
})
- CLIENT.on("connect", () => {
+ client.on("connect", () => {
clearTimeout(timeout)
CONNECTED = true
})
+ CLIENTS[selectDb] = client
}
-function waitForConnection() {
+function waitForConnection(selectDb: number = DEFAULT_SELECT_DB) {
return new Promise(resolve => {
- if (CLIENT == null) {
+ if (pickClient(selectDb) == null) {
init()
} else if (CONNECTED) {
- resolve()
+ resolve("")
return
}
// check if the connection is ready
const interval = setInterval(() => {
if (CONNECTED) {
clearInterval(interval)
- resolve()
+ resolve("")
}
}, 500)
})
@@ -100,25 +121,26 @@ function waitForConnection() {
* Utility function, takes a redis stream and converts it to a promisified response -
* this can only be done with redis streams because they will have an end.
* @param stream A redis stream, specifically as this type of stream will have an end.
+ * @param client The client to use for further lookups.
* @return {Promise