Merge branch 'master' into budi-5641-justified-text-is-not-working-properly-in-paragraph
This commit is contained in:
commit
3abbd7e363
|
@ -1,48 +0,0 @@
|
||||||
name: Budibase Deploy Production
|
|
||||||
|
|
||||||
on:
|
|
||||||
workflow_dispatch:
|
|
||||||
inputs:
|
|
||||||
version:
|
|
||||||
description: Budibase release version. For example - 1.0.0
|
|
||||||
required: false
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
release:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
|
|
||||||
steps:
|
|
||||||
- name: Fail if not a tag
|
|
||||||
run: |
|
|
||||||
if [[ $GITHUB_REF != refs/tags/* ]]; then
|
|
||||||
echo "Workflow Dispatch can only be run on tags"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
with:
|
|
||||||
fetch-depth: 0
|
|
||||||
|
|
||||||
- name: Fail if tag is not in master
|
|
||||||
run: |
|
|
||||||
if ! git merge-base --is-ancestor ${{ github.sha }} origin/master; then
|
|
||||||
echo "Tag is not in master. This pipeline can only execute tags that are present on the master branch"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
- name: Get the latest budibase release version
|
|
||||||
id: version
|
|
||||||
run: |
|
|
||||||
if [ -z "${{ github.event.inputs.version }}" ]; then
|
|
||||||
release_version=$(cat lerna.json | jq -r '.version')
|
|
||||||
else
|
|
||||||
release_version=${{ github.event.inputs.version }}
|
|
||||||
fi
|
|
||||||
echo "RELEASE_VERSION=$release_version" >> $GITHUB_ENV
|
|
||||||
|
|
||||||
- uses: passeidireto/trigger-external-workflow-action@main
|
|
||||||
env:
|
|
||||||
PAYLOAD_VERSION: ${{ env.RELEASE_VERSION }}
|
|
||||||
with:
|
|
||||||
repository: budibase/budibase-deploys
|
|
||||||
event: budicloud-prod-deploy
|
|
||||||
github_pat: ${{ secrets.GH_ACCESS_TOKEN }}
|
|
|
@ -1,178 +0,0 @@
|
||||||
name: Budibase Release
|
|
||||||
concurrency:
|
|
||||||
group: release
|
|
||||||
cancel-in-progress: false
|
|
||||||
|
|
||||||
on:
|
|
||||||
push:
|
|
||||||
tags:
|
|
||||||
- "[0-9]+.[0-9]+.[0-9]+"
|
|
||||||
# Exclude all pre-releases
|
|
||||||
- "!*[0-9]+.[0-9]+.[0-9]+-*"
|
|
||||||
|
|
||||||
env:
|
|
||||||
# Posthog token used by ui at build time
|
|
||||||
POSTHOG_TOKEN: phc_bIjZL7oh2GEUd2vqvTBH8WvrX0fWTFQMs6H5KQxiUxU
|
|
||||||
INTERCOM_TOKEN: ${{ secrets.INTERCOM_TOKEN }}
|
|
||||||
PERSONAL_ACCESS_TOKEN: ${{ secrets.PERSONAL_ACCESS_TOKEN }}
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
release-images:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
with:
|
|
||||||
submodules: true
|
|
||||||
token: ${{ secrets.PERSONAL_ACCESS_TOKEN }}
|
|
||||||
fetch-depth: 0
|
|
||||||
|
|
||||||
- name: Fail if tag is not in master
|
|
||||||
run: |
|
|
||||||
if ! git merge-base --is-ancestor ${{ github.sha }} origin/master; then
|
|
||||||
echo "Tag is not in master. This pipeline can only execute tags that are present on the master branch"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
- uses: actions/setup-node@v1
|
|
||||||
with:
|
|
||||||
node-version: 18.x
|
|
||||||
cache: yarn
|
|
||||||
|
|
||||||
- run: yarn install --frozen-lockfile
|
|
||||||
- name: Update versions
|
|
||||||
run: ./scripts/updateVersions.sh
|
|
||||||
- run: yarn lint
|
|
||||||
- run: yarn build
|
|
||||||
- run: yarn build:sdk
|
|
||||||
|
|
||||||
- name: Publish budibase packages to NPM
|
|
||||||
env:
|
|
||||||
NPM_TOKEN: ${{ secrets.NPM_TOKEN }}
|
|
||||||
run: |
|
|
||||||
# setup the username and email. I tend to use 'GitHub Actions Bot' with no email by default
|
|
||||||
git config --global user.name "Budibase Release Bot"
|
|
||||||
git config --global user.email "<>"
|
|
||||||
git submodule foreach git commit -a -m 'Release process'
|
|
||||||
git commit -a -m 'Release process'
|
|
||||||
echo //registry.npmjs.org/:_authToken=${NPM_TOKEN} >> .npmrc
|
|
||||||
yarn release
|
|
||||||
|
|
||||||
- name: "Get Current tag"
|
|
||||||
id: currenttag
|
|
||||||
run: |
|
|
||||||
version=$(./scripts/getCurrentVersion.sh)
|
|
||||||
echo "Using tag $version"
|
|
||||||
echo "version=$version" >> "$GITHUB_OUTPUT"
|
|
||||||
|
|
||||||
- name: Setup Docker Buildx
|
|
||||||
id: buildx
|
|
||||||
uses: docker/setup-buildx-action@v1
|
|
||||||
|
|
||||||
- name: Docker login
|
|
||||||
run: |
|
|
||||||
docker login -u $DOCKER_USER -p $DOCKER_PASSWORD
|
|
||||||
env:
|
|
||||||
DOCKER_USER: ${{ secrets.DOCKER_USERNAME }}
|
|
||||||
DOCKER_PASSWORD: ${{ secrets.DOCKER_API_KEY }}
|
|
||||||
|
|
||||||
- name: Build worker docker
|
|
||||||
uses: docker/build-push-action@v5
|
|
||||||
with:
|
|
||||||
context: .
|
|
||||||
push: true
|
|
||||||
platforms: linux/amd64,linux/arm64
|
|
||||||
build-args: |
|
|
||||||
BUDIBASE_VERSION=${{ env.BUDIBASE_VERSION }}
|
|
||||||
tags: ${{ env.IMAGE_NAME }}:${{ env.IMAGE_TAG }}
|
|
||||||
file: ./packages/worker/Dockerfile.v2
|
|
||||||
cache-from: type=registry,ref=${{ env.IMAGE_NAME }}:latest
|
|
||||||
cache-to: type=inline
|
|
||||||
env:
|
|
||||||
IMAGE_NAME: budibase/worker
|
|
||||||
IMAGE_TAG: ${{ steps.currenttag.outputs.version }}
|
|
||||||
BUDIBASE_VERSION: ${{ steps.currenttag.outputs.version }}
|
|
||||||
|
|
||||||
- name: Build server docker
|
|
||||||
uses: docker/build-push-action@v5
|
|
||||||
with:
|
|
||||||
context: .
|
|
||||||
push: true
|
|
||||||
platforms: linux/amd64,linux/arm64
|
|
||||||
build-args: |
|
|
||||||
BUDIBASE_VERSION=${{ env.BUDIBASE_VERSION }}
|
|
||||||
tags: ${{ env.IMAGE_NAME }}:${{ env.IMAGE_TAG }}
|
|
||||||
file: ./packages/server/Dockerfile.v2
|
|
||||||
cache-from: type=registry,ref=${{ env.IMAGE_NAME }}:latest
|
|
||||||
cache-to: type=inline
|
|
||||||
env:
|
|
||||||
IMAGE_NAME: budibase/apps
|
|
||||||
IMAGE_TAG: ${{ steps.currenttag.outputs.version }}
|
|
||||||
BUDIBASE_VERSION: ${{ steps.currenttag.outputs.version }}
|
|
||||||
|
|
||||||
- name: Build proxy docker
|
|
||||||
uses: docker/build-push-action@v5
|
|
||||||
with:
|
|
||||||
context: ./hosting/proxy
|
|
||||||
push: true
|
|
||||||
platforms: linux/amd64,linux/arm64
|
|
||||||
tags: ${{ env.IMAGE_NAME }}:${{ env.IMAGE_TAG }}
|
|
||||||
file: ./hosting/proxy/Dockerfile
|
|
||||||
cache-from: type=registry,ref=${{ env.IMAGE_NAME }}:latest
|
|
||||||
cache-to: type=inline
|
|
||||||
env:
|
|
||||||
IMAGE_NAME: budibase/proxy
|
|
||||||
IMAGE_TAG: ${{ steps.currenttag.outputs.version }}
|
|
||||||
|
|
||||||
release-helm-chart:
|
|
||||||
needs: [release-images]
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
- name: Setup Helm
|
|
||||||
uses: azure/setup-helm@v1
|
|
||||||
id: helm-install
|
|
||||||
|
|
||||||
- name: Get the latest budibase release version
|
|
||||||
id: version
|
|
||||||
run: |
|
|
||||||
release_version=$(cat lerna.json | jq -r '.version')
|
|
||||||
echo "RELEASE_VERSION=$release_version" >> $GITHUB_ENV
|
|
||||||
|
|
||||||
# due to helm repo index issue: https://github.com/helm/helm/issues/7363
|
|
||||||
# we need to create new package in a different dir, merge the index and move the package back
|
|
||||||
- name: Build and release helm chart
|
|
||||||
run: |
|
|
||||||
git config user.name "Budibase Helm Bot"
|
|
||||||
git config user.email "<>"
|
|
||||||
git reset --hard
|
|
||||||
git fetch
|
|
||||||
mkdir sync
|
|
||||||
echo "Packaging chart to sync dir"
|
|
||||||
helm package charts/budibase --version 0.0.0-master --app-version "$RELEASE_VERSION" --destination sync
|
|
||||||
echo "Packaging successful"
|
|
||||||
git checkout gh-pages
|
|
||||||
echo "Indexing helm repo"
|
|
||||||
helm repo index --merge docs/index.yaml sync
|
|
||||||
mv -f sync/* docs
|
|
||||||
rm -rf sync
|
|
||||||
echo "Pushing new helm release"
|
|
||||||
git add -A
|
|
||||||
git commit -m "Helm Release: ${{ env.RELEASE_VERSION }}"
|
|
||||||
git push
|
|
||||||
|
|
||||||
trigger-deploy-to-qa-env:
|
|
||||||
needs: [release-helm-chart]
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
|
|
||||||
- uses: peter-evans/repository-dispatch@v2
|
|
||||||
with:
|
|
||||||
repository: budibase/budibase-deploys
|
|
||||||
event-type: budicloud-qa-deploy
|
|
||||||
token: ${{ secrets.GH_ACCESS_TOKEN }}
|
|
||||||
client-payload: |-
|
|
||||||
{
|
|
||||||
"VERSION": "${{ github.ref_name }}",
|
|
||||||
"REF_NAME": "${{ github.ref_name}}"
|
|
||||||
}
|
|
|
@ -1,125 +0,0 @@
|
||||||
name: Budibase Release Selfhost
|
|
||||||
|
|
||||||
on:
|
|
||||||
workflow_dispatch:
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
release:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
|
|
||||||
steps:
|
|
||||||
- name: Fail if not a tag
|
|
||||||
run: |
|
|
||||||
if [[ $GITHUB_REF != refs/tags/* ]]; then
|
|
||||||
echo "Workflow Dispatch can only be run on tags"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
with:
|
|
||||||
submodules: true
|
|
||||||
token: ${{ secrets.PERSONAL_ACCESS_TOKEN }}
|
|
||||||
fetch-depth: 0
|
|
||||||
|
|
||||||
- name: Fail if tag is not in master
|
|
||||||
run: |
|
|
||||||
if ! git merge-base --is-ancestor ${{ github.sha }} origin/master; then
|
|
||||||
echo "Tag is not in master. This pipeline can only execute tags that are present on the master branch"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
- name: Use Node.js 18.x
|
|
||||||
uses: actions/setup-node@v1
|
|
||||||
with:
|
|
||||||
node-version: 18.x
|
|
||||||
|
|
||||||
- name: Get the latest budibase release version
|
|
||||||
id: version
|
|
||||||
run: |
|
|
||||||
release_version=$(cat lerna.json | jq -r '.version')
|
|
||||||
echo "RELEASE_VERSION=$release_version" >> $GITHUB_ENV
|
|
||||||
|
|
||||||
- name: Tag and release Docker images (Self Host)
|
|
||||||
run: |
|
|
||||||
docker login -u $DOCKER_USER -p $DOCKER_PASSWORD
|
|
||||||
|
|
||||||
release_tag=${{ env.RELEASE_VERSION }}
|
|
||||||
|
|
||||||
# Pull apps and worker images
|
|
||||||
docker pull budibase/apps:$release_tag
|
|
||||||
docker pull budibase/worker:$release_tag
|
|
||||||
docker pull budibase/proxy:$release_tag
|
|
||||||
|
|
||||||
# Tag apps and worker images
|
|
||||||
docker tag budibase/apps:$release_tag budibase/apps:$SELFHOST_TAG
|
|
||||||
docker tag budibase/worker:$release_tag budibase/worker:$SELFHOST_TAG
|
|
||||||
docker tag budibase/proxy:$release_tag budibase/proxy:$SELFHOST_TAG
|
|
||||||
|
|
||||||
# Push images
|
|
||||||
docker push budibase/apps:$SELFHOST_TAG
|
|
||||||
docker push budibase/worker:$SELFHOST_TAG
|
|
||||||
docker push budibase/proxy:$SELFHOST_TAG
|
|
||||||
env:
|
|
||||||
DOCKER_USER: ${{ secrets.DOCKER_USERNAME }}
|
|
||||||
DOCKER_PASSWORD: ${{ secrets.DOCKER_API_KEY }}
|
|
||||||
SELFHOST_TAG: latest
|
|
||||||
|
|
||||||
- name: Bootstrap and build (CLI)
|
|
||||||
run: |
|
|
||||||
yarn
|
|
||||||
yarn build
|
|
||||||
|
|
||||||
- name: Build OpenAPI spec
|
|
||||||
run: |
|
|
||||||
pushd packages/server
|
|
||||||
yarn
|
|
||||||
yarn specs
|
|
||||||
popd
|
|
||||||
|
|
||||||
- name: Setup Helm
|
|
||||||
uses: azure/setup-helm@v1
|
|
||||||
id: helm-install
|
|
||||||
|
|
||||||
# due to helm repo index issue: https://github.com/helm/helm/issues/7363
|
|
||||||
# we need to create new package in a different dir, merge the index and move the package back
|
|
||||||
- name: Build and release helm chart
|
|
||||||
run: |
|
|
||||||
git config user.name "Budibase Helm Bot"
|
|
||||||
git config user.email "<>"
|
|
||||||
git reset --hard
|
|
||||||
git fetch
|
|
||||||
mkdir sync
|
|
||||||
echo "Packaging chart to sync dir"
|
|
||||||
helm package charts/budibase --version "$RELEASE_VERSION" --app-version "$RELEASE_VERSION" --destination sync
|
|
||||||
echo "Packaging successful"
|
|
||||||
git checkout gh-pages
|
|
||||||
echo "Indexing helm repo"
|
|
||||||
helm repo index --merge docs/index.yaml sync
|
|
||||||
mv -f sync/* docs
|
|
||||||
rm -rf sync
|
|
||||||
echo "Pushing new helm release"
|
|
||||||
git add -A
|
|
||||||
git commit -m "Helm Release: ${{ env.RELEASE_VERSION }}"
|
|
||||||
git push
|
|
||||||
env:
|
|
||||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
|
||||||
|
|
||||||
- name: Perform Github Release
|
|
||||||
uses: softprops/action-gh-release@v1
|
|
||||||
with:
|
|
||||||
name: ${{ env.RELEASE_VERSION }}
|
|
||||||
tag_name: ${{ env.RELEASE_VERSION }}
|
|
||||||
generate_release_notes: true
|
|
||||||
files: |
|
|
||||||
packages/cli/build/cli-win.exe
|
|
||||||
packages/cli/build/cli-linux
|
|
||||||
packages/cli/build/cli-macos
|
|
||||||
packages/server/specs/openapi.yaml
|
|
||||||
packages/server/specs/openapi.json
|
|
||||||
|
|
||||||
- name: Discord Webhook Action
|
|
||||||
uses: tsickert/discord-webhook@v4.0.0
|
|
||||||
with:
|
|
||||||
webhook-url: ${{ secrets.PROD_DEPLOY_WEBHOOK_URL }}
|
|
||||||
content: "Self Host Deployment Complete: ${{ env.RELEASE_VERSION }} deployed to Self Host."
|
|
||||||
embed-title: ${{ env.RELEASE_VERSION }}
|
|
|
@ -1,86 +0,0 @@
|
||||||
name: Deploy Budibase Single Container Image to DockerHub
|
|
||||||
|
|
||||||
on:
|
|
||||||
workflow_dispatch:
|
|
||||||
|
|
||||||
env:
|
|
||||||
CI: true
|
|
||||||
PERSONAL_ACCESS_TOKEN: ${{ secrets.PERSONAL_ACCESS_TOKEN }}
|
|
||||||
REGISTRY_URL: registry.hub.docker.com
|
|
||||||
jobs:
|
|
||||||
build:
|
|
||||||
name: "build"
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
strategy:
|
|
||||||
matrix:
|
|
||||||
node-version: [18.x]
|
|
||||||
steps:
|
|
||||||
- name: Maximize build space
|
|
||||||
uses: easimon/maximize-build-space@master
|
|
||||||
with:
|
|
||||||
root-reserve-mb: 30000
|
|
||||||
swap-size-mb: 1024
|
|
||||||
remove-android: "true"
|
|
||||||
remove-dotnet: "true"
|
|
||||||
- name: Fail if not a tag
|
|
||||||
run: |
|
|
||||||
if [[ $GITHUB_REF != refs/tags/* ]]; then
|
|
||||||
echo "Workflow Dispatch can only be run on tags"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
- name: "Checkout"
|
|
||||||
uses: actions/checkout@v2
|
|
||||||
with:
|
|
||||||
submodules: true
|
|
||||||
token: ${{ secrets.PERSONAL_ACCESS_TOKEN }}
|
|
||||||
|
|
||||||
- name: Use Node.js ${{ matrix.node-version }}
|
|
||||||
uses: actions/setup-node@v1
|
|
||||||
with:
|
|
||||||
node-version: ${{ matrix.node-version }}
|
|
||||||
- name: Setup QEMU
|
|
||||||
uses: docker/setup-qemu-action@v1
|
|
||||||
- name: Setup Docker Buildx
|
|
||||||
id: buildx
|
|
||||||
uses: docker/setup-buildx-action@v1
|
|
||||||
- name: Run Yarn
|
|
||||||
run: yarn
|
|
||||||
- name: Update versions
|
|
||||||
run: ./scripts/updateVersions.sh
|
|
||||||
- name: Run Yarn Build
|
|
||||||
run: yarn build
|
|
||||||
- name: Login to Docker Hub
|
|
||||||
uses: docker/login-action@v2
|
|
||||||
with:
|
|
||||||
username: ${{ secrets.DOCKER_USERNAME }}
|
|
||||||
password: ${{ secrets.DOCKER_API_KEY }}
|
|
||||||
- name: Get the latest release version
|
|
||||||
id: version
|
|
||||||
run: |
|
|
||||||
release_version=$(cat lerna.json | jq -r '.version')
|
|
||||||
echo $release_version
|
|
||||||
echo "RELEASE_VERSION=$release_version" >> $GITHUB_ENV
|
|
||||||
- name: Tag and release Budibase service docker image
|
|
||||||
uses: docker/build-push-action@v2
|
|
||||||
with:
|
|
||||||
context: .
|
|
||||||
push: true
|
|
||||||
platforms: linux/amd64,linux/arm64
|
|
||||||
build-args: BUDIBASE_VERSION=${{ env.BUDIBASE_VERSION }}
|
|
||||||
tags: budibase/budibase,budibase/budibase:${{ env.RELEASE_VERSION }}
|
|
||||||
file: ./hosting/single/Dockerfile.v2
|
|
||||||
env:
|
|
||||||
BUDIBASE_VERSION: ${{ env.RELEASE_VERSION }}
|
|
||||||
- name: Tag and release Budibase Azure App Service docker image
|
|
||||||
uses: docker/build-push-action@v2
|
|
||||||
with:
|
|
||||||
context: .
|
|
||||||
push: true
|
|
||||||
platforms: linux/amd64
|
|
||||||
build-args: |
|
|
||||||
TARGETBUILD=aas
|
|
||||||
BUDIBASE_VERSION=${{ env.BUDIBASE_VERSION }}
|
|
||||||
tags: budibase/budibase-aas,budibase/budibase-aas:${{ env.RELEASE_VERSION }}
|
|
||||||
file: ./hosting/single/Dockerfile.v2
|
|
||||||
env:
|
|
||||||
BUDIBASE_VERSION: ${{ env.RELEASE_VERSION }}
|
|
|
@ -109,7 +109,7 @@ export class DatabaseImpl implements Database {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
async get<T>(id?: string): Promise<T | any> {
|
async get<T extends Document>(id?: string): Promise<T> {
|
||||||
const db = await this.checkSetup()
|
const db = await this.checkSetup()
|
||||||
if (!id) {
|
if (!id) {
|
||||||
throw new Error("Unable to get doc without a valid _id.")
|
throw new Error("Unable to get doc without a valid _id.")
|
||||||
|
@ -117,6 +117,28 @@ export class DatabaseImpl implements Database {
|
||||||
return this.updateOutput(() => db.get(id))
|
return this.updateOutput(() => db.get(id))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async getMultiple<T extends Document>(
|
||||||
|
ids: string[],
|
||||||
|
opts?: { allowMissing?: boolean }
|
||||||
|
): Promise<T[]> {
|
||||||
|
// get unique
|
||||||
|
ids = [...new Set(ids)]
|
||||||
|
const response = await this.allDocs<T>({
|
||||||
|
keys: ids,
|
||||||
|
include_docs: true,
|
||||||
|
})
|
||||||
|
const NOT_FOUND = "not_found"
|
||||||
|
const rows = response.rows.filter(row => row.error !== NOT_FOUND)
|
||||||
|
const someMissing = rows.length !== response.rows.length
|
||||||
|
// some were filtered out - means some missing
|
||||||
|
if (!opts?.allowMissing && someMissing) {
|
||||||
|
const missing = response.rows.filter(row => row.error === NOT_FOUND)
|
||||||
|
const missingIds = missing.map(row => row.key).join(", ")
|
||||||
|
throw new Error(`Unable to get documents: ${missingIds}`)
|
||||||
|
}
|
||||||
|
return rows.map(row => row.doc!)
|
||||||
|
}
|
||||||
|
|
||||||
async remove(idOrDoc: string | Document, rev?: string) {
|
async remove(idOrDoc: string | Document, rev?: string) {
|
||||||
const db = await this.checkSetup()
|
const db = await this.checkSetup()
|
||||||
let _id: string
|
let _id: string
|
||||||
|
|
|
@ -404,7 +404,7 @@
|
||||||
datasource = $datasources.list.find(ds => ds._id === query?.datasourceId)
|
datasource = $datasources.list.find(ds => ds._id === query?.datasourceId)
|
||||||
const datasourceUrl = datasource?.config.url
|
const datasourceUrl = datasource?.config.url
|
||||||
const qs = query?.fields.queryString
|
const qs = query?.fields.queryString
|
||||||
breakQs = restUtils.breakQueryString(qs)
|
breakQs = restUtils.breakQueryString(encodeURI(qs))
|
||||||
breakQs = runtimeToReadableMap(mergedBindings, breakQs)
|
breakQs = runtimeToReadableMap(mergedBindings, breakQs)
|
||||||
|
|
||||||
const path = query.fields.path
|
const path = query.fields.path
|
||||||
|
@ -652,7 +652,7 @@
|
||||||
<div class="bottom">
|
<div class="bottom">
|
||||||
<Layout paddingY="S" gap="S">
|
<Layout paddingY="S" gap="S">
|
||||||
<Divider />
|
<Divider />
|
||||||
{#if !response && Object.keys(schema).length === 0}
|
{#if !response && Object.keys(schema || {}).length === 0}
|
||||||
<Heading size="M">Response</Heading>
|
<Heading size="M">Response</Heading>
|
||||||
<div class="placeholder">
|
<div class="placeholder">
|
||||||
<div class="placeholder-internal">
|
<div class="placeholder-internal">
|
||||||
|
|
|
@ -1 +1 @@
|
||||||
Subproject commit ad9a0085bee0c4f3184acd86cadd872ea9917e88
|
Subproject commit e202f415d9fa540d08cc2ba6e27394fbc22f357b
|
|
@ -1,9 +1,5 @@
|
||||||
import * as linkRows from "../../../db/linkedRows"
|
import * as linkRows from "../../../db/linkedRows"
|
||||||
import {
|
import { generateRowID, InternalTables } from "../../../db/utils"
|
||||||
generateRowID,
|
|
||||||
getMultiIDParams,
|
|
||||||
InternalTables,
|
|
||||||
} from "../../../db/utils"
|
|
||||||
import * as userController from "../user"
|
import * as userController from "../user"
|
||||||
import {
|
import {
|
||||||
cleanupAttachments,
|
cleanupAttachments,
|
||||||
|
@ -240,8 +236,10 @@ export async function fetchEnrichedRow(ctx: UserCtx) {
|
||||||
const linkVals = links as LinkDocumentValue[]
|
const linkVals = links as LinkDocumentValue[]
|
||||||
|
|
||||||
// look up the actual rows based on the ids
|
// look up the actual rows based on the ids
|
||||||
const params = getMultiIDParams(linkVals.map(linkVal => linkVal.id))
|
let linkedRows = await db.getMultiple<Row>(
|
||||||
let linkedRows = (await db.allDocs<Row>(params)).rows.map(row => row.doc!)
|
linkVals.map(linkVal => linkVal.id),
|
||||||
|
{ allowMissing: true }
|
||||||
|
)
|
||||||
|
|
||||||
// get the linked tables
|
// get the linked tables
|
||||||
const linkTableIds = getLinkedTableIDs(table as Table)
|
const linkTableIds = getLinkedTableIDs(table as Table)
|
||||||
|
|
|
@ -1,21 +1,9 @@
|
||||||
import { InternalTables } from "../../../db/utils"
|
import { InternalTables } from "../../../db/utils"
|
||||||
import * as userController from "../user"
|
import * as userController from "../user"
|
||||||
import { context } from "@budibase/backend-core"
|
import { context } from "@budibase/backend-core"
|
||||||
import {
|
import { Ctx, Row, UserCtx } from "@budibase/types"
|
||||||
Ctx,
|
|
||||||
FieldType,
|
|
||||||
ManyToOneRelationshipFieldMetadata,
|
|
||||||
OneToManyRelationshipFieldMetadata,
|
|
||||||
Row,
|
|
||||||
SearchFilters,
|
|
||||||
Table,
|
|
||||||
UserCtx,
|
|
||||||
} from "@budibase/types"
|
|
||||||
import { FieldTypes, NoEmptyFilterStrings } from "../../../constants"
|
|
||||||
import sdk from "../../../sdk"
|
|
||||||
|
|
||||||
import validateJs from "validate.js"
|
import validateJs from "validate.js"
|
||||||
import { cloneDeep } from "lodash/fp"
|
|
||||||
|
|
||||||
validateJs.extend(validateJs.validators.datetime, {
|
validateJs.extend(validateJs.validators.datetime, {
|
||||||
parse: function (value: string) {
|
parse: function (value: string) {
|
||||||
|
|
|
@ -8,7 +8,7 @@ import {
|
||||||
getLinkedTable,
|
getLinkedTable,
|
||||||
} from "./linkUtils"
|
} from "./linkUtils"
|
||||||
import flatten from "lodash/flatten"
|
import flatten from "lodash/flatten"
|
||||||
import { getMultiIDParams, USER_METDATA_PREFIX } from "../utils"
|
import { USER_METDATA_PREFIX } from "../utils"
|
||||||
import partition from "lodash/partition"
|
import partition from "lodash/partition"
|
||||||
import { getGlobalUsersFromMetadata } from "../../utilities/global"
|
import { getGlobalUsersFromMetadata } from "../../utilities/global"
|
||||||
import { processFormulas } from "../../utilities/rowProcessor"
|
import { processFormulas } from "../../utilities/rowProcessor"
|
||||||
|
@ -79,9 +79,7 @@ async function getFullLinkedDocs(links: LinkDocumentValue[]) {
|
||||||
const db = context.getAppDB()
|
const db = context.getAppDB()
|
||||||
const linkedRowIds = links.map(link => link.id)
|
const linkedRowIds = links.map(link => link.id)
|
||||||
const uniqueRowIds = [...new Set(linkedRowIds)]
|
const uniqueRowIds = [...new Set(linkedRowIds)]
|
||||||
let dbRows = (await db.allDocs<Row>(getMultiIDParams(uniqueRowIds))).rows.map(
|
let dbRows = await db.getMultiple<Row>(uniqueRowIds, { allowMissing: true })
|
||||||
row => row.doc!
|
|
||||||
)
|
|
||||||
// convert the unique db rows back to a full list of linked rows
|
// convert the unique db rows back to a full list of linked rows
|
||||||
const linked = linkedRowIds
|
const linked = linkedRowIds
|
||||||
.map(id => dbRows.find(row => row && row._id === id))
|
.map(id => dbRows.find(row => row && row._id === id))
|
||||||
|
|
|
@ -283,16 +283,6 @@ export function generatePluginID(name: string) {
|
||||||
return `${DocumentType.PLUGIN}${SEPARATOR}${name}`
|
return `${DocumentType.PLUGIN}${SEPARATOR}${name}`
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* This can be used with the db.allDocs to get a list of IDs
|
|
||||||
*/
|
|
||||||
export function getMultiIDParams(ids: string[]) {
|
|
||||||
return {
|
|
||||||
keys: ids,
|
|
||||||
include_docs: true,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Generates a new view ID.
|
* Generates a new view ID.
|
||||||
* @returns The new view ID which the view doc can be stored under.
|
* @returns The new view ID which the view doc can be stored under.
|
||||||
|
|
|
@ -165,10 +165,22 @@ class RedisIntegration {
|
||||||
// commands split line by line
|
// commands split line by line
|
||||||
const commands = query.json.trim().split("\n")
|
const commands = query.json.trim().split("\n")
|
||||||
let pipelineCommands = []
|
let pipelineCommands = []
|
||||||
|
let tokenised
|
||||||
|
|
||||||
// process each command separately
|
// process each command separately
|
||||||
for (let command of commands) {
|
for (let command of commands) {
|
||||||
const tokenised = command.trim().split(" ")
|
const valueToken = command.trim().match(/".*"/)
|
||||||
|
if (valueToken?.[0]) {
|
||||||
|
tokenised = [
|
||||||
|
...command
|
||||||
|
.substring(0, command.indexOf(valueToken[0]) - 1)
|
||||||
|
.trim()
|
||||||
|
.split(" "),
|
||||||
|
valueToken?.[0],
|
||||||
|
]
|
||||||
|
} else {
|
||||||
|
tokenised = command.trim().split(" ")
|
||||||
|
}
|
||||||
// Pipeline only accepts lower case commands
|
// Pipeline only accepts lower case commands
|
||||||
tokenised[0] = tokenised[0].toLowerCase()
|
tokenised[0] = tokenised[0].toLowerCase()
|
||||||
pipelineCommands.push(tokenised)
|
pipelineCommands.push(tokenised)
|
||||||
|
|
|
@ -85,4 +85,21 @@ describe("Redis Integration", () => {
|
||||||
["get", "foo"],
|
["get", "foo"],
|
||||||
])
|
])
|
||||||
})
|
})
|
||||||
|
|
||||||
|
it("calls the pipeline method with double quoted phrase values", async () => {
|
||||||
|
const body = {
|
||||||
|
json: 'SET foo "What a wonderful world!"\nGET foo',
|
||||||
|
}
|
||||||
|
|
||||||
|
// ioredis-mock doesn't support pipelines
|
||||||
|
config.integration.client.pipeline = jest.fn(() => ({
|
||||||
|
exec: jest.fn(() => [[]]),
|
||||||
|
}))
|
||||||
|
|
||||||
|
await config.integration.command(body)
|
||||||
|
expect(config.integration.client.pipeline).toHaveBeenCalledWith([
|
||||||
|
["set", "foo", '"What a wonderful world!"'],
|
||||||
|
["get", "foo"],
|
||||||
|
])
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
import { context } from "@budibase/backend-core"
|
import { context } from "@budibase/backend-core"
|
||||||
import { getMultiIDParams, getTableParams } from "../../../db/utils"
|
import { getTableParams } from "../../../db/utils"
|
||||||
import {
|
import {
|
||||||
breakExternalTableId,
|
breakExternalTableId,
|
||||||
isExternalTableID,
|
isExternalTableID,
|
||||||
|
@ -17,6 +17,9 @@ import datasources from "../datasources"
|
||||||
import sdk from "../../../sdk"
|
import sdk from "../../../sdk"
|
||||||
|
|
||||||
export function processTable(table: Table): Table {
|
export function processTable(table: Table): Table {
|
||||||
|
if (!table) {
|
||||||
|
return table
|
||||||
|
}
|
||||||
if (table._id && isExternalTableID(table._id)) {
|
if (table._id && isExternalTableID(table._id)) {
|
||||||
return {
|
return {
|
||||||
...table,
|
...table,
|
||||||
|
@ -73,6 +76,9 @@ export async function getExternalTable(
|
||||||
tableName: string
|
tableName: string
|
||||||
): Promise<Table> {
|
): Promise<Table> {
|
||||||
const entities = await getExternalTablesInDatasource(datasourceId)
|
const entities = await getExternalTablesInDatasource(datasourceId)
|
||||||
|
if (!entities[tableName]) {
|
||||||
|
throw new Error(`Unable to find table named "${tableName}"`)
|
||||||
|
}
|
||||||
return processTable(entities[tableName])
|
return processTable(entities[tableName])
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -124,10 +130,10 @@ export async function getTables(tableIds: string[]): Promise<Table[]> {
|
||||||
}
|
}
|
||||||
if (internalTableIds.length) {
|
if (internalTableIds.length) {
|
||||||
const db = context.getAppDB()
|
const db = context.getAppDB()
|
||||||
const internalTableDocs = await db.allDocs<Table>(
|
const internalTables = await db.getMultiple<Table>(internalTableIds, {
|
||||||
getMultiIDParams(internalTableIds)
|
allowMissing: true,
|
||||||
)
|
})
|
||||||
tables = tables.concat(internalTableDocs.rows.map(row => row.doc!))
|
tables = tables.concat(internalTables)
|
||||||
}
|
}
|
||||||
return processTables(tables)
|
return processTables(tables)
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,39 @@
|
||||||
|
import TestConfig from "../../tests/utilities/TestConfiguration"
|
||||||
|
import { basicTable } from "../../tests/utilities/structures"
|
||||||
|
import { Table } from "@budibase/types"
|
||||||
|
import sdk from "../"
|
||||||
|
|
||||||
|
describe("tables", () => {
|
||||||
|
const config = new TestConfig()
|
||||||
|
let table: Table
|
||||||
|
|
||||||
|
beforeAll(async () => {
|
||||||
|
await config.init()
|
||||||
|
table = await config.api.table.create(basicTable())
|
||||||
|
})
|
||||||
|
|
||||||
|
describe("getTables", () => {
|
||||||
|
it("should be able to retrieve tables", async () => {
|
||||||
|
await config.doInContext(config.appId, async () => {
|
||||||
|
const tables = await sdk.tables.getTables([table._id!])
|
||||||
|
expect(tables.length).toBe(1)
|
||||||
|
expect(tables[0]._id).toBe(table._id)
|
||||||
|
expect(tables[0].name).toBe(table.name)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
it("shouldn't fail when retrieving tables that don't exist", async () => {
|
||||||
|
await config.doInContext(config.appId, async () => {
|
||||||
|
const tables = await sdk.tables.getTables(["unknown"])
|
||||||
|
expect(tables.length).toBe(0)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should de-duplicate the IDs", async () => {
|
||||||
|
await config.doInContext(config.appId, async () => {
|
||||||
|
const tables = await sdk.tables.getTables([table._id!, table._id!])
|
||||||
|
expect(tables.length).toBe(1)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
})
|
||||||
|
})
|
|
@ -1,4 +1,4 @@
|
||||||
import { getMultiIDParams, getGlobalIDFromUserMetadataID } from "../db/utils"
|
import { getGlobalIDFromUserMetadataID } from "../db/utils"
|
||||||
import {
|
import {
|
||||||
roles,
|
roles,
|
||||||
db as dbCore,
|
db as dbCore,
|
||||||
|
@ -96,9 +96,7 @@ export async function getRawGlobalUsers(userIds?: string[]): Promise<User[]> {
|
||||||
const db = tenancy.getGlobalDB()
|
const db = tenancy.getGlobalDB()
|
||||||
let globalUsers: User[]
|
let globalUsers: User[]
|
||||||
if (userIds) {
|
if (userIds) {
|
||||||
globalUsers = (await db.allDocs<User>(getMultiIDParams(userIds))).rows.map(
|
globalUsers = await db.getMultiple<User>(userIds, { allowMissing: true })
|
||||||
row => row.doc!
|
|
||||||
)
|
|
||||||
} else {
|
} else {
|
||||||
globalUsers = (
|
globalUsers = (
|
||||||
await db.allDocs<User>(
|
await db.allDocs<User>(
|
||||||
|
|
|
@ -122,7 +122,11 @@ export interface Database {
|
||||||
|
|
||||||
exists(): Promise<boolean>
|
exists(): Promise<boolean>
|
||||||
checkSetup(): Promise<Nano.DocumentScope<any>>
|
checkSetup(): Promise<Nano.DocumentScope<any>>
|
||||||
get<T>(id?: string): Promise<T>
|
get<T extends Document>(id?: string): Promise<T>
|
||||||
|
getMultiple<T extends Document>(
|
||||||
|
ids: string[],
|
||||||
|
opts?: { allowMissing?: boolean }
|
||||||
|
): Promise<T[]>
|
||||||
remove(
|
remove(
|
||||||
id: string | Document,
|
id: string | Document,
|
||||||
rev?: string
|
rev?: string
|
||||||
|
|
|
@ -7,7 +7,9 @@ export enum PlanType {
|
||||||
/** @deprecated */
|
/** @deprecated */
|
||||||
PREMIUM = "premium",
|
PREMIUM = "premium",
|
||||||
PREMIUM_PLUS = "premium_plus",
|
PREMIUM_PLUS = "premium_plus",
|
||||||
|
/** @deprecated */
|
||||||
BUSINESS = "business",
|
BUSINESS = "business",
|
||||||
|
ENTERPRISE_BASIC = "enterprise_basic",
|
||||||
ENTERPRISE = "enterprise",
|
ENTERPRISE = "enterprise",
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue