Merge pull request #13359 from Budibase/reuse-containers

Change the way we manage `testcontainers`
This commit is contained in:
Sam Rose 2024-04-03 14:38:26 +01:00 committed by GitHub
commit 86da22cbd7
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
26 changed files with 630 additions and 466 deletions

View File

@ -91,6 +91,9 @@ jobs:
test-libraries: test-libraries:
runs-on: ubuntu-latest runs-on: ubuntu-latest
env:
DEBUG: testcontainers,testcontainers:exec,testcontainers:build,testcontainers:pull
REUSE_CONTAINERS: true
steps: steps:
- name: Checkout repo - name: Checkout repo
uses: actions/checkout@v4 uses: actions/checkout@v4
@ -104,6 +107,14 @@ jobs:
with: with:
node-version: 20.x node-version: 20.x
cache: yarn cache: yarn
- name: Pull testcontainers images
run: |
docker pull testcontainers/ryuk:0.5.1 &
docker pull budibase/couchdb &
docker pull redis &
wait $(jobs -p)
- run: yarn --frozen-lockfile - run: yarn --frozen-lockfile
- name: Test - name: Test
run: | run: |
@ -138,9 +149,10 @@ jobs:
fi fi
test-server: test-server:
runs-on: ubuntu-latest runs-on: budi-tubby-tornado-quad-core-150gb
env: env:
DEBUG: testcontainers,testcontainers:exec,testcontainers:build,testcontainers:pull DEBUG: testcontainers,testcontainers:exec,testcontainers:build,testcontainers:pull
REUSE_CONTAINERS: true
steps: steps:
- name: Checkout repo - name: Checkout repo
uses: actions/checkout@v4 uses: actions/checkout@v4
@ -157,13 +169,16 @@ jobs:
- name: Pull testcontainers images - name: Pull testcontainers images
run: | run: |
docker pull mcr.microsoft.com/mssql/server:2022-latest docker pull mcr.microsoft.com/mssql/server:2022-latest &
docker pull mysql:8.3 docker pull mysql:8.3 &
docker pull postgres:16.1-bullseye docker pull postgres:16.1-bullseye &
docker pull mongo:7.0-jammy docker pull mongo:7.0-jammy &
docker pull mariadb:lts docker pull mariadb:lts &
docker pull testcontainers/ryuk:0.5.1 docker pull testcontainers/ryuk:0.5.1 &
docker pull budibase/couchdb docker pull budibase/couchdb &
docker pull redis &
wait $(jobs -p)
- run: yarn --frozen-lockfile - run: yarn --frozen-lockfile

View File

@ -1,25 +1,47 @@
import { GenericContainer, Wait } from "testcontainers" import { GenericContainer, Wait } from "testcontainers"
import path from "path"
import lockfile from "proper-lockfile"
export default async function setup() { export default async function setup() {
await new GenericContainer("budibase/couchdb") const lockPath = path.resolve(__dirname, "globalSetup.ts")
.withExposedPorts(5984) if (process.env.REUSE_CONTAINERS) {
.withEnvironment({ // If you run multiple tests at the same time, it's possible for the CouchDB
COUCHDB_PASSWORD: "budibase", // shared container to get started multiple times despite having an
COUCHDB_USER: "budibase", // identical reuse hash. To avoid that, we do a filesystem-based lock so
}) // that only one globalSetup.ts is running at a time.
.withCopyContentToContainer([ lockfile.lockSync(lockPath)
{ }
content: `
try {
let couchdb = new GenericContainer("budibase/couchdb")
.withExposedPorts(5984)
.withEnvironment({
COUCHDB_PASSWORD: "budibase",
COUCHDB_USER: "budibase",
})
.withCopyContentToContainer([
{
content: `
[log] [log]
level = warn level = warn
`, `,
target: "/opt/couchdb/etc/local.d/test-couchdb.ini", target: "/opt/couchdb/etc/local.d/test-couchdb.ini",
}, },
]) ])
.withWaitStrategy( .withWaitStrategy(
Wait.forSuccessfulCommand( Wait.forSuccessfulCommand(
"curl http://budibase:budibase@localhost:5984/_up" "curl http://budibase:budibase@localhost:5984/_up"
).withStartupTimeout(20000) ).withStartupTimeout(20000)
) )
.start()
if (process.env.REUSE_CONTAINERS) {
couchdb = couchdb.withReuse()
}
await couchdb.start()
} finally {
if (process.env.REUSE_CONTAINERS) {
lockfile.unlockSync(lockPath)
}
}
} }

View File

@ -7,6 +7,7 @@
"@babel/preset-env": "^7.22.5", "@babel/preset-env": "^7.22.5",
"@esbuild-plugins/tsconfig-paths": "^0.1.2", "@esbuild-plugins/tsconfig-paths": "^0.1.2",
"@types/node": "20.10.0", "@types/node": "20.10.0",
"@types/proper-lockfile": "^4.1.4",
"@typescript-eslint/parser": "6.9.0", "@typescript-eslint/parser": "6.9.0",
"esbuild": "^0.18.17", "esbuild": "^0.18.17",
"esbuild-node-externals": "^1.8.0", "esbuild-node-externals": "^1.8.0",
@ -23,6 +24,7 @@
"nx-cloud": "16.0.5", "nx-cloud": "16.0.5",
"prettier": "2.8.8", "prettier": "2.8.8",
"prettier-plugin-svelte": "^2.3.0", "prettier-plugin-svelte": "^2.3.0",
"proper-lockfile": "^4.1.2",
"svelte": "^4.2.10", "svelte": "^4.2.10",
"svelte-eslint-parser": "^0.33.1", "svelte-eslint-parser": "^0.33.1",
"typescript": "5.2.2", "typescript": "5.2.2",

View File

@ -1,6 +1,7 @@
import { DatabaseImpl } from "../../../src/db"
import { execSync } from "child_process" import { execSync } from "child_process"
const IPV4_PORT_REGEX = new RegExp(`0\\.0\\.0\\.0:(\\d+)->(\\d+)/tcp`, "g")
interface ContainerInfo { interface ContainerInfo {
Command: string Command: string
CreatedAt: string CreatedAt: string
@ -19,7 +20,10 @@ interface ContainerInfo {
} }
function getTestcontainers(): ContainerInfo[] { function getTestcontainers(): ContainerInfo[] {
return execSync("docker ps --format json") // We use --format json to make sure the output is nice and machine-readable,
// and we use --no-trunc so that the command returns full container IDs so we
// can filter on them correctly.
return execSync("docker ps --format json --no-trunc")
.toString() .toString()
.split("\n") .split("\n")
.filter(x => x.length > 0) .filter(x => x.length > 0)
@ -27,32 +31,55 @@ function getTestcontainers(): ContainerInfo[] {
.filter(x => x.Labels.includes("org.testcontainers=true")) .filter(x => x.Labels.includes("org.testcontainers=true"))
} }
function getContainerByImage(image: string) { export function getContainerByImage(image: string) {
return getTestcontainers().find(x => x.Image.startsWith(image)) const containers = getTestcontainers().filter(x => x.Image.startsWith(image))
if (containers.length > 1) {
let errorMessage = `Multiple containers found starting with image: "${image}"\n\n`
for (const container of containers) {
errorMessage += JSON.stringify(container, null, 2)
}
throw new Error(errorMessage)
}
return containers[0]
} }
function getExposedPort(container: ContainerInfo, port: number) { export function getContainerById(id: string) {
const match = container.Ports.match(new RegExp(`0.0.0.0:(\\d+)->${port}/tcp`)) return getTestcontainers().find(x => x.ID === id)
if (!match) { }
return undefined
export interface Port {
host: number
container: number
}
export function getExposedV4Ports(container: ContainerInfo): Port[] {
let ports: Port[] = []
for (const match of container.Ports.matchAll(IPV4_PORT_REGEX)) {
ports.push({ host: parseInt(match[1]), container: parseInt(match[2]) })
} }
return parseInt(match[1]) return ports
}
export function getExposedV4Port(container: ContainerInfo, port: number) {
return getExposedV4Ports(container).find(x => x.container === port)?.host
} }
export function setupEnv(...envs: any[]) { export function setupEnv(...envs: any[]) {
// We start couchdb in globalSetup.ts, in the root of the monorepo, so it
// should be relatively safe to look for it by its image name.
const couch = getContainerByImage("budibase/couchdb") const couch = getContainerByImage("budibase/couchdb")
if (!couch) { if (!couch) {
throw new Error("CouchDB container not found") throw new Error("CouchDB container not found")
} }
const couchPort = getExposedPort(couch, 5984) const couchPort = getExposedV4Port(couch, 5984)
if (!couchPort) { if (!couchPort) {
throw new Error("CouchDB port not found") throw new Error("CouchDB port not found")
} }
const configs = [ const configs = [
{ key: "COUCH_DB_PORT", value: `${couchPort}` }, { key: "COUCH_DB_PORT", value: `${couchPort}` },
{ key: "COUCH_DB_URL", value: `http://localhost:${couchPort}` }, { key: "COUCH_DB_URL", value: `http://127.0.0.1:${couchPort}` },
] ]
for (const config of configs.filter(x => !!x.value)) { for (const config of configs.filter(x => !!x.value)) {
@ -60,7 +87,4 @@ export function setupEnv(...envs: any[]) {
env._set(config.key, config.value) env._set(config.key, config.value)
} }
} }
// @ts-expect-error
DatabaseImpl.nano = undefined
} }

View File

@ -4,8 +4,8 @@ set -e
if [[ -n $CI ]] if [[ -n $CI ]]
then then
export NODE_OPTIONS="--max-old-space-size=4096 --no-node-snapshot $NODE_OPTIONS" export NODE_OPTIONS="--max-old-space-size=4096 --no-node-snapshot $NODE_OPTIONS"
echo "jest --coverage --maxWorkers=2 --forceExit --workerIdleMemoryLimit=2000MB --bail $@" echo "jest --coverage --maxWorkers=4 --forceExit --workerIdleMemoryLimit=2000MB --bail $@"
jest --coverage --maxWorkers=2 --forceExit --workerIdleMemoryLimit=2000MB --bail $@ jest --coverage --maxWorkers=4 --forceExit --workerIdleMemoryLimit=2000MB --bail $@
else else
# --maxWorkers performs better in development # --maxWorkers performs better in development
export NODE_OPTIONS="--no-node-snapshot $NODE_OPTIONS" export NODE_OPTIONS="--no-node-snapshot $NODE_OPTIONS"

View File

@ -1,7 +1,5 @@
const setup = require("../../tests/utilities") const setup = require("../../tests/utilities")
jest.setTimeout(30000)
describe("/metrics", () => { describe("/metrics", () => {
let request = setup.getRequest() let request = setup.getRequest()
let config = setup.getConfig() let config = setup.getConfig()

View File

@ -1,7 +1,6 @@
import * as setup from "./utilities" import * as setup from "./utilities"
import path from "path" import path from "path"
jest.setTimeout(15000)
const PASSWORD = "testtest" const PASSWORD = "testtest"
describe("/applications/:appId/import", () => { describe("/applications/:appId/import", () => {

View File

@ -23,8 +23,6 @@ let {
collectAutomation, collectAutomation,
} = setup.structures } = setup.structures
jest.setTimeout(30000)
describe("/automations", () => { describe("/automations", () => {
let request = setup.getRequest() let request = setup.getRequest()
let config = setup.getConfig() let config = setup.getConfig()

View File

@ -1,9 +1,10 @@
import { Datasource, Query, SourceName } from "@budibase/types" import { Datasource, Query, SourceName } from "@budibase/types"
import * as setup from "../utilities" import * as setup from "../utilities"
import { databaseTestProviders } from "../../../../integrations/tests/utils" import {
import pg from "pg" DatabaseName,
import mysql from "mysql2/promise" getDatasource,
import mssql from "mssql" rawQuery,
} from "../../../../integrations/tests/utils"
jest.unmock("pg") jest.unmock("pg")
@ -34,13 +35,16 @@ const createTableSQL: Record<string, string> = {
const insertSQL = `INSERT INTO test_table (name) VALUES ('one'), ('two'), ('three'), ('four'), ('five')` const insertSQL = `INSERT INTO test_table (name) VALUES ('one'), ('two'), ('three'), ('four'), ('five')`
const dropTableSQL = `DROP TABLE test_table;` const dropTableSQL = `DROP TABLE test_table;`
describe.each([ describe.each(
["postgres", databaseTestProviders.postgres], [
["mysql", databaseTestProviders.mysql], DatabaseName.POSTGRES,
["mssql", databaseTestProviders.mssql], DatabaseName.MYSQL,
["mariadb", databaseTestProviders.mariadb], DatabaseName.SQL_SERVER,
])("queries (%s)", (dbName, dsProvider) => { DatabaseName.MARIADB,
].map(name => [name, getDatasource(name)])
)("queries (%s)", (dbName, dsProvider) => {
const config = setup.getConfig() const config = setup.getConfig()
let rawDatasource: Datasource
let datasource: Datasource let datasource: Datasource
async function createQuery(query: Partial<Query>): Promise<Query> { async function createQuery(query: Partial<Query>): Promise<Query> {
@ -57,62 +61,22 @@ describe.each([
return await config.api.query.save({ ...defaultQuery, ...query }) return await config.api.query.save({ ...defaultQuery, ...query })
} }
async function rawQuery(sql: string): Promise<any> {
// We re-fetch the datasource here because the one returned by
// config.api.datasource.create has the password field blanked out, and we
// need the password to connect to the database.
const ds = await dsProvider.datasource()
switch (ds.source) {
case SourceName.POSTGRES: {
const client = new pg.Client(ds.config!)
await client.connect()
try {
const { rows } = await client.query(sql)
return rows
} finally {
await client.end()
}
}
case SourceName.MYSQL: {
const con = await mysql.createConnection(ds.config!)
try {
const [rows] = await con.query(sql)
return rows
} finally {
con.end()
}
}
case SourceName.SQL_SERVER: {
const pool = new mssql.ConnectionPool(ds.config! as mssql.config)
const client = await pool.connect()
try {
const { recordset } = await client.query(sql)
return recordset
} finally {
await pool.close()
}
}
}
}
beforeAll(async () => { beforeAll(async () => {
await config.init() await config.init()
datasource = await config.api.datasource.create( rawDatasource = await dsProvider
await dsProvider.datasource() datasource = await config.api.datasource.create(rawDatasource)
)
}) })
beforeEach(async () => { beforeEach(async () => {
await rawQuery(createTableSQL[datasource.source]) await rawQuery(rawDatasource, createTableSQL[datasource.source])
await rawQuery(insertSQL) await rawQuery(rawDatasource, insertSQL)
}) })
afterEach(async () => { afterEach(async () => {
await rawQuery(dropTableSQL) await rawQuery(rawDatasource, dropTableSQL)
}) })
afterAll(async () => { afterAll(async () => {
await dsProvider.stop()
setup.afterAll() setup.afterAll()
}) })
@ -143,7 +107,10 @@ describe.each([
}, },
]) ])
const rows = await rawQuery("SELECT * FROM test_table WHERE name = 'baz'") const rows = await rawQuery(
rawDatasource,
"SELECT * FROM test_table WHERE name = 'baz'"
)
expect(rows).toHaveLength(1) expect(rows).toHaveLength(1)
}) })
@ -171,6 +138,7 @@ describe.each([
expect(result.data).toEqual([{ created: true }]) expect(result.data).toEqual([{ created: true }])
const rows = await rawQuery( const rows = await rawQuery(
rawDatasource,
`SELECT * FROM test_table WHERE birthday = '${date.toISOString()}'` `SELECT * FROM test_table WHERE birthday = '${date.toISOString()}'`
) )
expect(rows).toHaveLength(1) expect(rows).toHaveLength(1)
@ -202,6 +170,7 @@ describe.each([
expect(result.data).toEqual([{ created: true }]) expect(result.data).toEqual([{ created: true }])
const rows = await rawQuery( const rows = await rawQuery(
rawDatasource,
`SELECT * FROM test_table WHERE name = '${notDateStr}'` `SELECT * FROM test_table WHERE name = '${notDateStr}'`
) )
expect(rows).toHaveLength(1) expect(rows).toHaveLength(1)
@ -338,7 +307,10 @@ describe.each([
}, },
]) ])
const rows = await rawQuery("SELECT * FROM test_table WHERE id = 1") const rows = await rawQuery(
rawDatasource,
"SELECT * FROM test_table WHERE id = 1"
)
expect(rows).toEqual([ expect(rows).toEqual([
{ id: 1, name: "foo", birthday: null, number: null }, { id: 1, name: "foo", birthday: null, number: null },
]) ])
@ -406,7 +378,10 @@ describe.each([
}, },
]) ])
const rows = await rawQuery("SELECT * FROM test_table WHERE id = 1") const rows = await rawQuery(
rawDatasource,
"SELECT * FROM test_table WHERE id = 1"
)
expect(rows).toHaveLength(0) expect(rows).toHaveLength(0)
}) })
}) })
@ -443,7 +418,7 @@ describe.each([
} catch (err: any) { } catch (err: any) {
error = err.message error = err.message
} }
if (dbName === "mssql") { if (dbName === DatabaseName.SQL_SERVER) {
expect(error).toBeUndefined() expect(error).toBeUndefined()
} else { } else {
expect(error).toBeDefined() expect(error).toBeDefined()

View File

@ -1,14 +1,17 @@
import { Datasource, Query } from "@budibase/types" import { Datasource, Query } from "@budibase/types"
import * as setup from "../utilities" import * as setup from "../utilities"
import { databaseTestProviders } from "../../../../integrations/tests/utils" import {
import { MongoClient, type Collection, BSON } from "mongodb" DatabaseName,
getDatasource,
const collection = "test_collection" } from "../../../../integrations/tests/utils"
import { MongoClient, type Collection, BSON, Db } from "mongodb"
import { generator } from "@budibase/backend-core/tests"
const expectValidId = expect.stringMatching(/^\w{24}$/) const expectValidId = expect.stringMatching(/^\w{24}$/)
const expectValidBsonObjectId = expect.any(BSON.ObjectId) const expectValidBsonObjectId = expect.any(BSON.ObjectId)
describe("/queries", () => { describe("/queries", () => {
let collection: string
let config = setup.getConfig() let config = setup.getConfig()
let datasource: Datasource let datasource: Datasource
@ -37,8 +40,7 @@ describe("/queries", () => {
async function withClient<T>( async function withClient<T>(
callback: (client: MongoClient) => Promise<T> callback: (client: MongoClient) => Promise<T>
): Promise<T> { ): Promise<T> {
const ds = await databaseTestProviders.mongodb.datasource() const client = new MongoClient(datasource.config!.connectionString)
const client = new MongoClient(ds.config!.connectionString)
await client.connect() await client.connect()
try { try {
return await callback(client) return await callback(client)
@ -47,30 +49,33 @@ describe("/queries", () => {
} }
} }
async function withDb<T>(callback: (db: Db) => Promise<T>): Promise<T> {
return await withClient(async client => {
return await callback(client.db(datasource.config!.db))
})
}
async function withCollection<T>( async function withCollection<T>(
callback: (collection: Collection) => Promise<T> callback: (collection: Collection) => Promise<T>
): Promise<T> { ): Promise<T> {
return await withClient(async client => { return await withDb(async db => {
const db = client.db(
(await databaseTestProviders.mongodb.datasource()).config!.db
)
return await callback(db.collection(collection)) return await callback(db.collection(collection))
}) })
} }
afterAll(async () => { afterAll(async () => {
await databaseTestProviders.mongodb.stop()
setup.afterAll() setup.afterAll()
}) })
beforeAll(async () => { beforeAll(async () => {
await config.init() await config.init()
datasource = await config.api.datasource.create( datasource = await config.api.datasource.create(
await databaseTestProviders.mongodb.datasource() await getDatasource(DatabaseName.MONGODB)
) )
}) })
beforeEach(async () => { beforeEach(async () => {
collection = generator.guid()
await withCollection(async collection => { await withCollection(async collection => {
await collection.insertMany([ await collection.insertMany([
{ name: "one" }, { name: "one" },

View File

@ -1,4 +1,4 @@
import { databaseTestProviders } from "../../../integrations/tests/utils" import { DatabaseName, getDatasource } from "../../../integrations/tests/utils"
import tk from "timekeeper" import tk from "timekeeper"
import { outputProcessing } from "../../../utilities/rowProcessor" import { outputProcessing } from "../../../utilities/rowProcessor"
@ -34,10 +34,10 @@ jest.unmock("pg")
describe.each([ describe.each([
["internal", undefined], ["internal", undefined],
["postgres", databaseTestProviders.postgres], [DatabaseName.POSTGRES, getDatasource(DatabaseName.POSTGRES)],
["mysql", databaseTestProviders.mysql], [DatabaseName.MYSQL, getDatasource(DatabaseName.MYSQL)],
["mssql", databaseTestProviders.mssql], [DatabaseName.SQL_SERVER, getDatasource(DatabaseName.SQL_SERVER)],
["mariadb", databaseTestProviders.mariadb], [DatabaseName.MARIADB, getDatasource(DatabaseName.MARIADB)],
])("/rows (%s)", (__, dsProvider) => { ])("/rows (%s)", (__, dsProvider) => {
const isInternal = dsProvider === undefined const isInternal = dsProvider === undefined
const config = setup.getConfig() const config = setup.getConfig()
@ -49,23 +49,23 @@ describe.each([
await config.init() await config.init()
if (dsProvider) { if (dsProvider) {
datasource = await config.createDatasource({ datasource = await config.createDatasource({
datasource: await dsProvider.datasource(), datasource: await dsProvider,
}) })
} }
}) })
afterAll(async () => { afterAll(async () => {
if (dsProvider) {
await dsProvider.stop()
}
setup.afterAll() setup.afterAll()
}) })
function saveTableRequest( function saveTableRequest(
...overrides: Partial<SaveTableRequest>[] // We omit the name field here because it's generated in the function with a
// high likelihood to be unique. Tests should not have any reason to control
// the table name they're writing to.
...overrides: Partial<Omit<SaveTableRequest, "name">>[]
): SaveTableRequest { ): SaveTableRequest {
const req: SaveTableRequest = { const req: SaveTableRequest = {
name: uuid.v4().substring(0, 16), name: uuid.v4().substring(0, 10),
type: "table", type: "table",
sourceType: datasource sourceType: datasource
? TableSourceType.EXTERNAL ? TableSourceType.EXTERNAL
@ -87,7 +87,10 @@ describe.each([
} }
function defaultTable( function defaultTable(
...overrides: Partial<SaveTableRequest>[] // We omit the name field here because it's generated in the function with a
// high likelihood to be unique. Tests should not have any reason to control
// the table name they're writing to.
...overrides: Partial<Omit<SaveTableRequest, "name">>[]
): SaveTableRequest { ): SaveTableRequest {
return saveTableRequest( return saveTableRequest(
{ {
@ -194,7 +197,6 @@ describe.each([
const newTable = await config.api.table.save( const newTable = await config.api.table.save(
saveTableRequest({ saveTableRequest({
name: "TestTableAuto",
schema: { schema: {
"Row ID": { "Row ID": {
name: "Row ID", name: "Row ID",
@ -383,11 +385,9 @@ describe.each([
isInternal && isInternal &&
it("doesn't allow creating in user table", async () => { it("doesn't allow creating in user table", async () => {
const userTableId = InternalTable.USER_METADATA
const response = await config.api.row.save( const response = await config.api.row.save(
userTableId, InternalTable.USER_METADATA,
{ {
tableId: userTableId,
firstName: "Joe", firstName: "Joe",
lastName: "Joe", lastName: "Joe",
email: "joe@joe.com", email: "joe@joe.com",
@ -462,7 +462,6 @@ describe.each([
table = await config.api.table.save(defaultTable()) table = await config.api.table.save(defaultTable())
otherTable = await config.api.table.save( otherTable = await config.api.table.save(
defaultTable({ defaultTable({
name: "a",
schema: { schema: {
relationship: { relationship: {
name: "relationship", name: "relationship",
@ -898,8 +897,8 @@ describe.each([
let o2mTable: Table let o2mTable: Table
let m2mTable: Table let m2mTable: Table
beforeAll(async () => { beforeAll(async () => {
o2mTable = await config.api.table.save(defaultTable({ name: "o2m" })) o2mTable = await config.api.table.save(defaultTable())
m2mTable = await config.api.table.save(defaultTable({ name: "m2m" })) m2mTable = await config.api.table.save(defaultTable())
}) })
describe.each([ describe.each([
@ -1256,7 +1255,6 @@ describe.each([
otherTable = await config.api.table.save(defaultTable()) otherTable = await config.api.table.save(defaultTable())
table = await config.api.table.save( table = await config.api.table.save(
saveTableRequest({ saveTableRequest({
name: "b",
schema: { schema: {
links: { links: {
name: "links", name: "links",
@ -1354,7 +1352,6 @@ describe.each([
const table = await config.api.table.save( const table = await config.api.table.save(
saveTableRequest({ saveTableRequest({
name: "table",
schema: { schema: {
text: { text: {
name: "text", name: "text",

View File

@ -3,8 +3,6 @@ import { checkPermissionsEndpoint } from "./utilities/TestFunctions"
import * as setup from "./utilities" import * as setup from "./utilities"
import { UserMetadata } from "@budibase/types" import { UserMetadata } from "@budibase/types"
jest.setTimeout(30000)
jest.mock("../../../utilities/workerRequests", () => ({ jest.mock("../../../utilities/workerRequests", () => ({
getGlobalUsers: jest.fn(() => { getGlobalUsers: jest.fn(() => {
return {} return {}

View File

@ -19,8 +19,7 @@ import {
ViewV2, ViewV2,
} from "@budibase/types" } from "@budibase/types"
import { generator, mocks } from "@budibase/backend-core/tests" import { generator, mocks } from "@budibase/backend-core/tests"
import * as uuid from "uuid" import { DatabaseName, getDatasource } from "../../../integrations/tests/utils"
import { databaseTestProviders } from "../../../integrations/tests/utils"
import merge from "lodash/merge" import merge from "lodash/merge"
import { quotas } from "@budibase/pro" import { quotas } from "@budibase/pro"
import { roles } from "@budibase/backend-core" import { roles } from "@budibase/backend-core"
@ -30,10 +29,10 @@ jest.unmock("pg")
describe.each([ describe.each([
["internal", undefined], ["internal", undefined],
["postgres", databaseTestProviders.postgres], [DatabaseName.POSTGRES, getDatasource(DatabaseName.POSTGRES)],
["mysql", databaseTestProviders.mysql], [DatabaseName.MYSQL, getDatasource(DatabaseName.MYSQL)],
["mssql", databaseTestProviders.mssql], [DatabaseName.SQL_SERVER, getDatasource(DatabaseName.SQL_SERVER)],
["mariadb", databaseTestProviders.mariadb], [DatabaseName.MARIADB, getDatasource(DatabaseName.MARIADB)],
])("/v2/views (%s)", (_, dsProvider) => { ])("/v2/views (%s)", (_, dsProvider) => {
const config = setup.getConfig() const config = setup.getConfig()
const isInternal = !dsProvider const isInternal = !dsProvider
@ -42,10 +41,10 @@ describe.each([
let datasource: Datasource let datasource: Datasource
function saveTableRequest( function saveTableRequest(
...overrides: Partial<SaveTableRequest>[] ...overrides: Partial<Omit<SaveTableRequest, "name">>[]
): SaveTableRequest { ): SaveTableRequest {
const req: SaveTableRequest = { const req: SaveTableRequest = {
name: uuid.v4().substring(0, 16), name: generator.guid().replaceAll("-", "").substring(0, 16),
type: "table", type: "table",
sourceType: datasource sourceType: datasource
? TableSourceType.EXTERNAL ? TableSourceType.EXTERNAL
@ -90,16 +89,13 @@ describe.each([
if (dsProvider) { if (dsProvider) {
datasource = await config.createDatasource({ datasource = await config.createDatasource({
datasource: await dsProvider.datasource(), datasource: await dsProvider,
}) })
} }
table = await config.api.table.save(priceTable()) table = await config.api.table.save(priceTable())
}) })
afterAll(async () => { afterAll(async () => {
if (dsProvider) {
await dsProvider.stop()
}
setup.afterAll() setup.afterAll()
}) })
@ -231,7 +227,7 @@ describe.each([
view = await config.api.viewV2.create({ view = await config.api.viewV2.create({
tableId: table._id!, tableId: table._id!,
name: "View A", name: generator.guid(),
}) })
}) })
@ -307,12 +303,13 @@ describe.each([
it("can update an existing view name", async () => { it("can update an existing view name", async () => {
const tableId = table._id! const tableId = table._id!
await config.api.viewV2.update({ ...view, name: "View B" }) const newName = generator.guid()
await config.api.viewV2.update({ ...view, name: newName })
expect(await config.api.table.get(tableId)).toEqual( expect(await config.api.table.get(tableId)).toEqual(
expect.objectContaining({ expect.objectContaining({
views: { views: {
"View B": { ...view, name: "View B", schema: expect.anything() }, [newName]: { ...view, name: newName, schema: expect.anything() },
}, },
}) })
) )
@ -507,7 +504,6 @@ describe.each([
it("views have extra data trimmed", async () => { it("views have extra data trimmed", async () => {
const table = await config.api.table.save( const table = await config.api.table.save(
saveTableRequest({ saveTableRequest({
name: "orders",
schema: { schema: {
Country: { Country: {
type: FieldType.STRING, type: FieldType.STRING,
@ -523,7 +519,7 @@ describe.each([
const view = await config.api.viewV2.create({ const view = await config.api.viewV2.create({
tableId: table._id!, tableId: table._id!,
name: uuid.v4(), name: generator.guid(),
schema: { schema: {
Country: { Country: {
visible: true, visible: true,
@ -853,7 +849,6 @@ describe.each([
beforeAll(async () => { beforeAll(async () => {
table = await config.api.table.save( table = await config.api.table.save(
saveTableRequest({ saveTableRequest({
name: `users_${uuid.v4()}`,
type: "table", type: "table",
schema: { schema: {
name: { name: {

View File

@ -3,7 +3,6 @@ import {
generateMakeRequest, generateMakeRequest,
MakeRequestResponse, MakeRequestResponse,
} from "../api/routes/public/tests/utils" } from "../api/routes/public/tests/utils"
import { v4 as uuidv4 } from "uuid"
import * as setup from "../api/routes/tests/utilities" import * as setup from "../api/routes/tests/utilities"
import { import {
Datasource, Datasource,
@ -12,12 +11,23 @@ import {
TableRequest, TableRequest,
TableSourceType, TableSourceType,
} from "@budibase/types" } from "@budibase/types"
import { databaseTestProviders } from "../integrations/tests/utils" import {
import mysql from "mysql2/promise" DatabaseName,
getDatasource,
rawQuery,
} from "../integrations/tests/utils"
import { builderSocket } from "../websockets" import { builderSocket } from "../websockets"
import { generator } from "@budibase/backend-core/tests"
// @ts-ignore // @ts-ignore
fetch.mockSearch() fetch.mockSearch()
function uniqueTableName(length?: number): string {
return generator
.guid()
.replaceAll("-", "_")
.substring(0, length || 10)
}
const config = setup.getConfig()! const config = setup.getConfig()!
jest.mock("../websockets", () => ({ jest.mock("../websockets", () => ({
@ -37,7 +47,8 @@ jest.mock("../websockets", () => ({
describe("mysql integrations", () => { describe("mysql integrations", () => {
let makeRequest: MakeRequestResponse, let makeRequest: MakeRequestResponse,
mysqlDatasource: Datasource, rawDatasource: Datasource,
datasource: Datasource,
primaryMySqlTable: Table primaryMySqlTable: Table
beforeAll(async () => { beforeAll(async () => {
@ -46,18 +57,13 @@ describe("mysql integrations", () => {
makeRequest = generateMakeRequest(apiKey, true) makeRequest = generateMakeRequest(apiKey, true)
mysqlDatasource = await config.api.datasource.create( rawDatasource = await getDatasource(DatabaseName.MYSQL)
await databaseTestProviders.mysql.datasource() datasource = await config.api.datasource.create(rawDatasource)
)
})
afterAll(async () => {
await databaseTestProviders.mysql.stop()
}) })
beforeEach(async () => { beforeEach(async () => {
primaryMySqlTable = await config.createTable({ primaryMySqlTable = await config.createTable({
name: uuidv4(), name: uniqueTableName(),
type: "table", type: "table",
primary: ["id"], primary: ["id"],
schema: { schema: {
@ -79,7 +85,7 @@ describe("mysql integrations", () => {
type: FieldType.NUMBER, type: FieldType.NUMBER,
}, },
}, },
sourceId: mysqlDatasource._id, sourceId: datasource._id,
sourceType: TableSourceType.EXTERNAL, sourceType: TableSourceType.EXTERNAL,
}) })
}) })
@ -87,18 +93,15 @@ describe("mysql integrations", () => {
afterAll(config.end) afterAll(config.end)
it("validate table schema", async () => { it("validate table schema", async () => {
const res = await makeRequest( const res = await makeRequest("get", `/api/datasources/${datasource._id}`)
"get",
`/api/datasources/${mysqlDatasource._id}`
)
expect(res.status).toBe(200) expect(res.status).toBe(200)
expect(res.body).toEqual({ expect(res.body).toEqual({
config: { config: {
database: "mysql", database: expect.any(String),
host: mysqlDatasource.config!.host, host: datasource.config!.host,
password: "--secret-value--", password: "--secret-value--",
port: mysqlDatasource.config!.port, port: datasource.config!.port,
user: "root", user: "root",
}, },
plus: true, plus: true,
@ -117,7 +120,7 @@ describe("mysql integrations", () => {
it("should be able to verify the connection", async () => { it("should be able to verify the connection", async () => {
await config.api.datasource.verify( await config.api.datasource.verify(
{ {
datasource: await databaseTestProviders.mysql.datasource(), datasource: rawDatasource,
}, },
{ {
body: { body: {
@ -128,13 +131,12 @@ describe("mysql integrations", () => {
}) })
it("should state an invalid datasource cannot connect", async () => { it("should state an invalid datasource cannot connect", async () => {
const dbConfig = await databaseTestProviders.mysql.datasource()
await config.api.datasource.verify( await config.api.datasource.verify(
{ {
datasource: { datasource: {
...dbConfig, ...rawDatasource,
config: { config: {
...dbConfig.config, ...rawDatasource.config,
password: "wrongpassword", password: "wrongpassword",
}, },
}, },
@ -154,7 +156,7 @@ describe("mysql integrations", () => {
it("should fetch information about mysql datasource", async () => { it("should fetch information about mysql datasource", async () => {
const primaryName = primaryMySqlTable.name const primaryName = primaryMySqlTable.name
const response = await makeRequest("post", "/api/datasources/info", { const response = await makeRequest("post", "/api/datasources/info", {
datasource: mysqlDatasource, datasource: datasource,
}) })
expect(response.status).toBe(200) expect(response.status).toBe(200)
expect(response.body.tableNames).toBeDefined() expect(response.body.tableNames).toBeDefined()
@ -163,40 +165,38 @@ describe("mysql integrations", () => {
}) })
describe("Integration compatibility with mysql search_path", () => { describe("Integration compatibility with mysql search_path", () => {
let client: mysql.Connection, pathDatasource: Datasource let datasource: Datasource, rawDatasource: Datasource
const database = "test1" const database = generator.guid()
const database2 = "test-2" const database2 = generator.guid()
beforeAll(async () => { beforeAll(async () => {
const dsConfig = await databaseTestProviders.mysql.datasource() rawDatasource = await getDatasource(DatabaseName.MYSQL)
const dbConfig = dsConfig.config!
client = await mysql.createConnection(dbConfig) await rawQuery(rawDatasource, `CREATE DATABASE \`${database}\`;`)
await client.query(`CREATE DATABASE \`${database}\`;`) await rawQuery(rawDatasource, `CREATE DATABASE \`${database2}\`;`)
await client.query(`CREATE DATABASE \`${database2}\`;`)
const pathConfig: any = { const pathConfig: any = {
...dsConfig, ...rawDatasource,
config: { config: {
...dbConfig, ...rawDatasource.config!,
database, database,
}, },
} }
pathDatasource = await config.api.datasource.create(pathConfig) datasource = await config.api.datasource.create(pathConfig)
}) })
afterAll(async () => { afterAll(async () => {
await client.query(`DROP DATABASE \`${database}\`;`) await rawQuery(rawDatasource, `DROP DATABASE \`${database}\`;`)
await client.query(`DROP DATABASE \`${database2}\`;`) await rawQuery(rawDatasource, `DROP DATABASE \`${database2}\`;`)
await client.end()
}) })
it("discovers tables from any schema in search path", async () => { it("discovers tables from any schema in search path", async () => {
await client.query( await rawQuery(
rawDatasource,
`CREATE TABLE \`${database}\`.table1 (id1 SERIAL PRIMARY KEY);` `CREATE TABLE \`${database}\`.table1 (id1 SERIAL PRIMARY KEY);`
) )
const response = await makeRequest("post", "/api/datasources/info", { const response = await makeRequest("post", "/api/datasources/info", {
datasource: pathDatasource, datasource: datasource,
}) })
expect(response.status).toBe(200) expect(response.status).toBe(200)
expect(response.body.tableNames).toBeDefined() expect(response.body.tableNames).toBeDefined()
@ -207,15 +207,17 @@ describe("mysql integrations", () => {
it("does not mix columns from different tables", async () => { it("does not mix columns from different tables", async () => {
const repeated_table_name = "table_same_name" const repeated_table_name = "table_same_name"
await client.query( await rawQuery(
rawDatasource,
`CREATE TABLE \`${database}\`.${repeated_table_name} (id SERIAL PRIMARY KEY, val1 TEXT);` `CREATE TABLE \`${database}\`.${repeated_table_name} (id SERIAL PRIMARY KEY, val1 TEXT);`
) )
await client.query( await rawQuery(
rawDatasource,
`CREATE TABLE \`${database2}\`.${repeated_table_name} (id2 SERIAL PRIMARY KEY, val2 TEXT);` `CREATE TABLE \`${database2}\`.${repeated_table_name} (id2 SERIAL PRIMARY KEY, val2 TEXT);`
) )
const response = await makeRequest( const response = await makeRequest(
"post", "post",
`/api/datasources/${pathDatasource._id}/schema`, `/api/datasources/${datasource._id}/schema`,
{ {
tablesFilter: [repeated_table_name], tablesFilter: [repeated_table_name],
} }
@ -231,30 +233,14 @@ describe("mysql integrations", () => {
}) })
describe("POST /api/tables/", () => { describe("POST /api/tables/", () => {
let client: mysql.Connection
const emitDatasourceUpdateMock = jest.fn() const emitDatasourceUpdateMock = jest.fn()
beforeEach(async () => {
client = await mysql.createConnection(
(
await databaseTestProviders.mysql.datasource()
).config!
)
mysqlDatasource = await config.api.datasource.create(
await databaseTestProviders.mysql.datasource()
)
})
afterEach(async () => {
await client.end()
})
it("will emit the datasource entity schema with externalType to the front-end when adding a new column", async () => { it("will emit the datasource entity schema with externalType to the front-end when adding a new column", async () => {
const addColumnToTable: TableRequest = { const addColumnToTable: TableRequest = {
type: "table", type: "table",
sourceType: TableSourceType.EXTERNAL, sourceType: TableSourceType.EXTERNAL,
name: "table", name: uniqueTableName(),
sourceId: mysqlDatasource._id!, sourceId: datasource._id!,
primary: ["id"], primary: ["id"],
schema: { schema: {
id: { id: {
@ -301,14 +287,16 @@ describe("mysql integrations", () => {
}, },
}, },
created: true, created: true,
_id: `${mysqlDatasource._id}__table`, _id: `${datasource._id}__${addColumnToTable.name}`,
} }
delete expectedTable._add delete expectedTable._add
expect(emitDatasourceUpdateMock).toHaveBeenCalledTimes(1) expect(emitDatasourceUpdateMock).toHaveBeenCalledTimes(1)
const emittedDatasource: Datasource = const emittedDatasource: Datasource =
emitDatasourceUpdateMock.mock.calls[0][1] emitDatasourceUpdateMock.mock.calls[0][1]
expect(emittedDatasource.entities!["table"]).toEqual(expectedTable) expect(emittedDatasource.entities![expectedTable.name]).toEqual(
expectedTable
)
}) })
it("will rename a column", async () => { it("will rename a column", async () => {
@ -346,17 +334,18 @@ describe("mysql integrations", () => {
"/api/tables/", "/api/tables/",
renameColumnOnTable renameColumnOnTable
) )
mysqlDatasource = (
await makeRequest( const ds = (
"post", await makeRequest("post", `/api/datasources/${datasource._id}/schema`)
`/api/datasources/${mysqlDatasource._id}/schema`
)
).body.datasource ).body.datasource
expect(response.status).toEqual(200) expect(response.status).toEqual(200)
expect( expect(Object.keys(ds.entities![primaryMySqlTable.name].schema)).toEqual([
Object.keys(mysqlDatasource.entities![primaryMySqlTable.name].schema) "id",
).toEqual(["id", "name", "description", "age"]) "name",
"description",
"age",
])
}) })
}) })
}) })

View File

@ -16,8 +16,12 @@ import {
import _ from "lodash" import _ from "lodash"
import { generator } from "@budibase/backend-core/tests" import { generator } from "@budibase/backend-core/tests"
import { utils } from "@budibase/backend-core" import { utils } from "@budibase/backend-core"
import { databaseTestProviders } from "../integrations/tests/utils" import {
import { Client } from "pg" DatabaseName,
getDatasource,
rawQuery,
} from "../integrations/tests/utils"
// @ts-ignore // @ts-ignore
fetch.mockSearch() fetch.mockSearch()
@ -28,7 +32,8 @@ jest.mock("../websockets")
describe("postgres integrations", () => { describe("postgres integrations", () => {
let makeRequest: MakeRequestResponse, let makeRequest: MakeRequestResponse,
postgresDatasource: Datasource, rawDatasource: Datasource,
datasource: Datasource,
primaryPostgresTable: Table, primaryPostgresTable: Table,
oneToManyRelationshipInfo: ForeignTableInfo, oneToManyRelationshipInfo: ForeignTableInfo,
manyToOneRelationshipInfo: ForeignTableInfo, manyToOneRelationshipInfo: ForeignTableInfo,
@ -40,19 +45,17 @@ describe("postgres integrations", () => {
makeRequest = generateMakeRequest(apiKey, true) makeRequest = generateMakeRequest(apiKey, true)
postgresDatasource = await config.api.datasource.create( rawDatasource = await getDatasource(DatabaseName.POSTGRES)
await databaseTestProviders.postgres.datasource() datasource = await config.api.datasource.create(rawDatasource)
)
})
afterAll(async () => {
await databaseTestProviders.postgres.stop()
}) })
beforeEach(async () => { beforeEach(async () => {
async function createAuxTable(prefix: string) { async function createAuxTable(prefix: string) {
return await config.createTable({ return await config.createTable({
name: `${prefix}_${generator.word({ length: 6 })}`, name: `${prefix}_${generator
.guid()
.replaceAll("-", "")
.substring(0, 6)}`,
type: "table", type: "table",
primary: ["id"], primary: ["id"],
primaryDisplay: "title", primaryDisplay: "title",
@ -67,7 +70,7 @@ describe("postgres integrations", () => {
type: FieldType.STRING, type: FieldType.STRING,
}, },
}, },
sourceId: postgresDatasource._id, sourceId: datasource._id,
sourceType: TableSourceType.EXTERNAL, sourceType: TableSourceType.EXTERNAL,
}) })
} }
@ -89,7 +92,7 @@ describe("postgres integrations", () => {
} }
primaryPostgresTable = await config.createTable({ primaryPostgresTable = await config.createTable({
name: `p_${generator.word({ length: 6 })}`, name: `p_${generator.guid().replaceAll("-", "").substring(0, 6)}`,
type: "table", type: "table",
primary: ["id"], primary: ["id"],
schema: { schema: {
@ -144,7 +147,7 @@ describe("postgres integrations", () => {
main: true, main: true,
}, },
}, },
sourceId: postgresDatasource._id, sourceId: datasource._id,
sourceType: TableSourceType.EXTERNAL, sourceType: TableSourceType.EXTERNAL,
}) })
}) })
@ -251,7 +254,7 @@ describe("postgres integrations", () => {
async function createDefaultPgTable() { async function createDefaultPgTable() {
return await config.createTable({ return await config.createTable({
name: generator.word({ length: 10 }), name: generator.guid().replaceAll("-", "").substring(0, 10),
type: "table", type: "table",
primary: ["id"], primary: ["id"],
schema: { schema: {
@ -261,7 +264,7 @@ describe("postgres integrations", () => {
autocolumn: true, autocolumn: true,
}, },
}, },
sourceId: postgresDatasource._id, sourceId: datasource._id,
sourceType: TableSourceType.EXTERNAL, sourceType: TableSourceType.EXTERNAL,
}) })
} }
@ -299,19 +302,16 @@ describe("postgres integrations", () => {
} }
it("validate table schema", async () => { it("validate table schema", async () => {
const res = await makeRequest( const res = await makeRequest("get", `/api/datasources/${datasource._id}`)
"get",
`/api/datasources/${postgresDatasource._id}`
)
expect(res.status).toBe(200) expect(res.status).toBe(200)
expect(res.body).toEqual({ expect(res.body).toEqual({
config: { config: {
ca: false, ca: false,
database: "postgres", database: expect.any(String),
host: postgresDatasource.config!.host, host: datasource.config!.host,
password: "--secret-value--", password: "--secret-value--",
port: postgresDatasource.config!.port, port: datasource.config!.port,
rejectUnauthorized: false, rejectUnauthorized: false,
schema: "public", schema: "public",
ssl: false, ssl: false,
@ -1043,7 +1043,7 @@ describe("postgres integrations", () => {
it("should be able to verify the connection", async () => { it("should be able to verify the connection", async () => {
await config.api.datasource.verify( await config.api.datasource.verify(
{ {
datasource: await databaseTestProviders.postgres.datasource(), datasource: await getDatasource(DatabaseName.POSTGRES),
}, },
{ {
body: { body: {
@ -1054,7 +1054,7 @@ describe("postgres integrations", () => {
}) })
it("should state an invalid datasource cannot connect", async () => { it("should state an invalid datasource cannot connect", async () => {
const dbConfig = await databaseTestProviders.postgres.datasource() const dbConfig = await getDatasource(DatabaseName.POSTGRES)
await config.api.datasource.verify( await config.api.datasource.verify(
{ {
datasource: { datasource: {
@ -1079,7 +1079,7 @@ describe("postgres integrations", () => {
it("should fetch information about postgres datasource", async () => { it("should fetch information about postgres datasource", async () => {
const primaryName = primaryPostgresTable.name const primaryName = primaryPostgresTable.name
const response = await makeRequest("post", "/api/datasources/info", { const response = await makeRequest("post", "/api/datasources/info", {
datasource: postgresDatasource, datasource: datasource,
}) })
expect(response.status).toBe(200) expect(response.status).toBe(200)
expect(response.body.tableNames).toBeDefined() expect(response.body.tableNames).toBeDefined()
@ -1088,86 +1088,88 @@ describe("postgres integrations", () => {
}) })
describe("POST /api/datasources/:datasourceId/schema", () => { describe("POST /api/datasources/:datasourceId/schema", () => {
let client: Client let tableName: string
beforeEach(async () => { beforeEach(async () => {
client = new Client( tableName = generator.guid().replaceAll("-", "").substring(0, 10)
(await databaseTestProviders.postgres.datasource()).config!
)
await client.connect()
}) })
afterEach(async () => { afterEach(async () => {
await client.query(`DROP TABLE IF EXISTS "table"`) await rawQuery(rawDatasource, `DROP TABLE IF EXISTS "${tableName}"`)
await client.end()
}) })
it("recognises when a table has no primary key", async () => { it("recognises when a table has no primary key", async () => {
await client.query(`CREATE TABLE "table" (id SERIAL)`) await rawQuery(rawDatasource, `CREATE TABLE "${tableName}" (id SERIAL)`)
const response = await makeRequest( const response = await makeRequest(
"post", "post",
`/api/datasources/${postgresDatasource._id}/schema` `/api/datasources/${datasource._id}/schema`
) )
expect(response.body.errors).toEqual({ expect(response.body.errors).toEqual({
table: "Table must have a primary key.", [tableName]: "Table must have a primary key.",
}) })
}) })
it("recognises when a table is using a reserved column name", async () => { it("recognises when a table is using a reserved column name", async () => {
await client.query(`CREATE TABLE "table" (_id SERIAL PRIMARY KEY) `) await rawQuery(
rawDatasource,
`CREATE TABLE "${tableName}" (_id SERIAL PRIMARY KEY) `
)
const response = await makeRequest( const response = await makeRequest(
"post", "post",
`/api/datasources/${postgresDatasource._id}/schema` `/api/datasources/${datasource._id}/schema`
) )
expect(response.body.errors).toEqual({ expect(response.body.errors).toEqual({
table: "Table contains invalid columns.", [tableName]: "Table contains invalid columns.",
}) })
}) })
}) })
describe("Integration compatibility with postgres search_path", () => { describe("Integration compatibility with postgres search_path", () => {
let client: Client, pathDatasource: Datasource let rawDatasource: Datasource,
const schema1 = "test1", datasource: Datasource,
schema2 = "test-2" schema1: string,
schema2: string
beforeAll(async () => { beforeEach(async () => {
const dsConfig = await databaseTestProviders.postgres.datasource() schema1 = generator.guid().replaceAll("-", "")
const dbConfig = dsConfig.config! schema2 = generator.guid().replaceAll("-", "")
client = new Client(dbConfig) rawDatasource = await getDatasource(DatabaseName.POSTGRES)
await client.connect() const dbConfig = rawDatasource.config!
await client.query(`CREATE SCHEMA "${schema1}";`)
await client.query(`CREATE SCHEMA "${schema2}";`) await rawQuery(rawDatasource, `CREATE SCHEMA "${schema1}";`)
await rawQuery(rawDatasource, `CREATE SCHEMA "${schema2}";`)
const pathConfig: any = { const pathConfig: any = {
...dsConfig, ...rawDatasource,
config: { config: {
...dbConfig, ...dbConfig,
schema: `${schema1}, ${schema2}`, schema: `${schema1}, ${schema2}`,
}, },
} }
pathDatasource = await config.api.datasource.create(pathConfig) datasource = await config.api.datasource.create(pathConfig)
}) })
afterAll(async () => { afterEach(async () => {
await client.query(`DROP SCHEMA "${schema1}" CASCADE;`) await rawQuery(rawDatasource, `DROP SCHEMA "${schema1}" CASCADE;`)
await client.query(`DROP SCHEMA "${schema2}" CASCADE;`) await rawQuery(rawDatasource, `DROP SCHEMA "${schema2}" CASCADE;`)
await client.end()
}) })
it("discovers tables from any schema in search path", async () => { it("discovers tables from any schema in search path", async () => {
await client.query( await rawQuery(
rawDatasource,
`CREATE TABLE "${schema1}".table1 (id1 SERIAL PRIMARY KEY);` `CREATE TABLE "${schema1}".table1 (id1 SERIAL PRIMARY KEY);`
) )
await client.query( await rawQuery(
rawDatasource,
`CREATE TABLE "${schema2}".table2 (id2 SERIAL PRIMARY KEY);` `CREATE TABLE "${schema2}".table2 (id2 SERIAL PRIMARY KEY);`
) )
const response = await makeRequest("post", "/api/datasources/info", { const response = await makeRequest("post", "/api/datasources/info", {
datasource: pathDatasource, datasource: datasource,
}) })
expect(response.status).toBe(200) expect(response.status).toBe(200)
expect(response.body.tableNames).toBeDefined() expect(response.body.tableNames).toBeDefined()
@ -1178,15 +1180,17 @@ describe("postgres integrations", () => {
it("does not mix columns from different tables", async () => { it("does not mix columns from different tables", async () => {
const repeated_table_name = "table_same_name" const repeated_table_name = "table_same_name"
await client.query( await rawQuery(
rawDatasource,
`CREATE TABLE "${schema1}".${repeated_table_name} (id SERIAL PRIMARY KEY, val1 TEXT);` `CREATE TABLE "${schema1}".${repeated_table_name} (id SERIAL PRIMARY KEY, val1 TEXT);`
) )
await client.query( await rawQuery(
rawDatasource,
`CREATE TABLE "${schema2}".${repeated_table_name} (id2 SERIAL PRIMARY KEY, val2 TEXT);` `CREATE TABLE "${schema2}".${repeated_table_name} (id2 SERIAL PRIMARY KEY, val2 TEXT);`
) )
const response = await makeRequest( const response = await makeRequest(
"post", "post",
`/api/datasources/${pathDatasource._id}/schema`, `/api/datasources/${datasource._id}/schema`,
{ {
tablesFilter: [repeated_table_name], tablesFilter: [repeated_table_name],
} }

View File

@ -1,25 +1,90 @@
jest.unmock("pg") jest.unmock("pg")
import { Datasource } from "@budibase/types" import { Datasource, SourceName } from "@budibase/types"
import * as postgres from "./postgres" import * as postgres from "./postgres"
import * as mongodb from "./mongodb" import * as mongodb from "./mongodb"
import * as mysql from "./mysql" import * as mysql from "./mysql"
import * as mssql from "./mssql" import * as mssql from "./mssql"
import * as mariadb from "./mariadb" import * as mariadb from "./mariadb"
import { StartedTestContainer } from "testcontainers" import { GenericContainer } from "testcontainers"
import { testContainerUtils } from "@budibase/backend-core/tests"
jest.setTimeout(30000) export type DatasourceProvider = () => Promise<Datasource>
export interface DatabaseProvider { export enum DatabaseName {
start(): Promise<StartedTestContainer> POSTGRES = "postgres",
stop(): Promise<void> MONGODB = "mongodb",
datasource(): Promise<Datasource> MYSQL = "mysql",
SQL_SERVER = "mssql",
MARIADB = "mariadb",
} }
export const databaseTestProviders = { const providers: Record<DatabaseName, DatasourceProvider> = {
postgres, [DatabaseName.POSTGRES]: postgres.getDatasource,
mongodb, [DatabaseName.MONGODB]: mongodb.getDatasource,
mysql, [DatabaseName.MYSQL]: mysql.getDatasource,
mssql, [DatabaseName.SQL_SERVER]: mssql.getDatasource,
mariadb, [DatabaseName.MARIADB]: mariadb.getDatasource,
}
export function getDatasourceProviders(
...sourceNames: DatabaseName[]
): Promise<Datasource>[] {
return sourceNames.map(sourceName => providers[sourceName]())
}
export function getDatasourceProvider(
sourceName: DatabaseName
): DatasourceProvider {
return providers[sourceName]
}
export function getDatasource(sourceName: DatabaseName): Promise<Datasource> {
return providers[sourceName]()
}
export async function getDatasources(
...sourceNames: DatabaseName[]
): Promise<Datasource[]> {
return Promise.all(sourceNames.map(sourceName => providers[sourceName]()))
}
export async function rawQuery(ds: Datasource, sql: string): Promise<any> {
switch (ds.source) {
case SourceName.POSTGRES: {
return postgres.rawQuery(ds, sql)
}
case SourceName.MYSQL: {
return mysql.rawQuery(ds, sql)
}
case SourceName.SQL_SERVER: {
return mssql.rawQuery(ds, sql)
}
default: {
throw new Error(`Unsupported source: ${ds.source}`)
}
}
}
export async function startContainer(container: GenericContainer) {
if (process.env.REUSE_CONTAINERS) {
container = container.withReuse()
}
const startedContainer = await container.start()
const info = testContainerUtils.getContainerById(startedContainer.getId())
if (!info) {
throw new Error("Container not found")
}
// Some Docker runtimes, when you expose a port, will bind it to both
// 127.0.0.1 and ::1, so ipv4 and ipv6. The port spaces of ipv4 and ipv6
// addresses are not shared, and testcontainers will sometimes give you back
// the ipv6 port. There's no way to know that this has happened, and if you
// try to then connect to `localhost:port` you may attempt to bind to the v4
// address which could be unbound or even an entirely different container. For
// that reason, we don't use testcontainers' `getExposedPort` function,
// preferring instead our own method that guaranteed v4 ports.
return testContainerUtils.getExposedV4Ports(info)
} }

View File

@ -1,8 +1,11 @@
import { Datasource, SourceName } from "@budibase/types" import { Datasource, SourceName } from "@budibase/types"
import { GenericContainer, Wait, StartedTestContainer } from "testcontainers" import { GenericContainer, Wait } from "testcontainers"
import { AbstractWaitStrategy } from "testcontainers/build/wait-strategies/wait-strategy" import { AbstractWaitStrategy } from "testcontainers/build/wait-strategies/wait-strategy"
import { rawQuery } from "./mysql"
import { generator, testContainerUtils } from "@budibase/backend-core/tests"
import { startContainer } from "."
let container: StartedTestContainer | undefined let ports: Promise<testContainerUtils.Port[]>
class MariaDBWaitStrategy extends AbstractWaitStrategy { class MariaDBWaitStrategy extends AbstractWaitStrategy {
async waitUntilReady(container: any, boundPorts: any, startTime?: Date) { async waitUntilReady(container: any, boundPorts: any, startTime?: Date) {
@ -21,38 +24,38 @@ class MariaDBWaitStrategy extends AbstractWaitStrategy {
} }
} }
export async function start(): Promise<StartedTestContainer> { export async function getDatasource(): Promise<Datasource> {
return await new GenericContainer("mariadb:lts") if (!ports) {
.withExposedPorts(3306) ports = startContainer(
.withEnvironment({ MARIADB_ROOT_PASSWORD: "password" }) new GenericContainer("mariadb:lts")
.withWaitStrategy(new MariaDBWaitStrategy()) .withExposedPorts(3306)
.start() .withEnvironment({ MARIADB_ROOT_PASSWORD: "password" })
} .withWaitStrategy(new MariaDBWaitStrategy())
)
export async function datasource(): Promise<Datasource> {
if (!container) {
container = await start()
} }
const host = container.getHost()
const port = container.getMappedPort(3306)
return { const port = (await ports).find(x => x.container === 3306)?.host
if (!port) {
throw new Error("MariaDB port not found")
}
const config = {
host: "127.0.0.1",
port,
user: "root",
password: "password",
database: "mysql",
}
const datasource = {
type: "datasource_plus", type: "datasource_plus",
source: SourceName.MYSQL, source: SourceName.MYSQL,
plus: true, plus: true,
config: { config,
host,
port,
user: "root",
password: "password",
database: "mysql",
},
} }
}
export async function stop() { const database = generator.guid().replaceAll("-", "")
if (container) { await rawQuery(datasource, `CREATE DATABASE \`${database}\``)
await container.stop() datasource.config.database = database
container = undefined return datasource
}
} }

View File

@ -1,43 +1,39 @@
import { generator, testContainerUtils } from "@budibase/backend-core/tests"
import { Datasource, SourceName } from "@budibase/types" import { Datasource, SourceName } from "@budibase/types"
import { GenericContainer, Wait, StartedTestContainer } from "testcontainers" import { GenericContainer, Wait } from "testcontainers"
import { startContainer } from "."
let container: StartedTestContainer | undefined let ports: Promise<testContainerUtils.Port[]>
export async function start(): Promise<StartedTestContainer> { export async function getDatasource(): Promise<Datasource> {
return await new GenericContainer("mongo:7.0-jammy") if (!ports) {
.withExposedPorts(27017) ports = startContainer(
.withEnvironment({ new GenericContainer("mongo:7.0-jammy")
MONGO_INITDB_ROOT_USERNAME: "mongo", .withExposedPorts(27017)
MONGO_INITDB_ROOT_PASSWORD: "password", .withEnvironment({
}) MONGO_INITDB_ROOT_USERNAME: "mongo",
.withWaitStrategy( MONGO_INITDB_ROOT_PASSWORD: "password",
Wait.forSuccessfulCommand( })
`mongosh --eval "db.version()"` .withWaitStrategy(
).withStartupTimeout(10000) Wait.forSuccessfulCommand(
`mongosh --eval "db.version()"`
).withStartupTimeout(10000)
)
) )
.start()
}
export async function datasource(): Promise<Datasource> {
if (!container) {
container = await start()
} }
const host = container.getHost()
const port = container.getMappedPort(27017) const port = (await ports).find(x => x.container === 27017)
if (!port) {
throw new Error("MongoDB port not found")
}
return { return {
type: "datasource", type: "datasource",
source: SourceName.MONGODB, source: SourceName.MONGODB,
plus: false, plus: false,
config: { config: {
connectionString: `mongodb://mongo:password@${host}:${port}`, connectionString: `mongodb://mongo:password@127.0.0.1:${port.host}`,
db: "mongo", db: generator.guid(),
}, },
} }
} }
export async function stop() {
if (container) {
await container.stop()
container = undefined
}
}

View File

@ -1,43 +1,41 @@
import { Datasource, SourceName } from "@budibase/types" import { Datasource, SourceName } from "@budibase/types"
import { GenericContainer, Wait, StartedTestContainer } from "testcontainers" import { GenericContainer, Wait } from "testcontainers"
import mssql from "mssql"
import { generator, testContainerUtils } from "@budibase/backend-core/tests"
import { startContainer } from "."
let container: StartedTestContainer | undefined let ports: Promise<testContainerUtils.Port[]>
export async function start(): Promise<StartedTestContainer> { export async function getDatasource(): Promise<Datasource> {
return await new GenericContainer( if (!ports) {
"mcr.microsoft.com/mssql/server:2022-latest" ports = startContainer(
) new GenericContainer("mcr.microsoft.com/mssql/server:2022-latest")
.withExposedPorts(1433) .withExposedPorts(1433)
.withEnvironment({ .withEnvironment({
ACCEPT_EULA: "Y", ACCEPT_EULA: "Y",
MSSQL_SA_PASSWORD: "Password_123", MSSQL_SA_PASSWORD: "Password_123",
// This is important, as Microsoft allow us to use the "Developer" edition // This is important, as Microsoft allow us to use the "Developer" edition
// of SQL Server for development and testing purposes. We can't use other // of SQL Server for development and testing purposes. We can't use other
// versions without a valid license, and we cannot use the Developer // versions without a valid license, and we cannot use the Developer
// version in production. // version in production.
MSSQL_PID: "Developer", MSSQL_PID: "Developer",
}) })
.withWaitStrategy( .withWaitStrategy(
Wait.forSuccessfulCommand( Wait.forSuccessfulCommand(
"/opt/mssql-tools/bin/sqlcmd -S localhost -U sa -P Password_123 -q 'SELECT 1'" "/opt/mssql-tools/bin/sqlcmd -S localhost -U sa -P Password_123 -q 'SELECT 1'"
) )
)
) )
.start()
}
export async function datasource(): Promise<Datasource> {
if (!container) {
container = await start()
} }
const host = container.getHost()
const port = container.getMappedPort(1433)
return { const port = (await ports).find(x => x.container === 1433)?.host
const datasource: Datasource = {
type: "datasource_plus", type: "datasource_plus",
source: SourceName.SQL_SERVER, source: SourceName.SQL_SERVER,
plus: true, plus: true,
config: { config: {
server: host, server: "127.0.0.1",
port, port,
user: "sa", user: "sa",
password: "Password_123", password: "Password_123",
@ -46,11 +44,28 @@ export async function datasource(): Promise<Datasource> {
}, },
}, },
} }
const database = generator.guid().replaceAll("-", "")
await rawQuery(datasource, `CREATE DATABASE "${database}"`)
datasource.config!.database = database
return datasource
} }
export async function stop() { export async function rawQuery(ds: Datasource, sql: string) {
if (container) { if (!ds.config) {
await container.stop() throw new Error("Datasource config is missing")
container = undefined }
if (ds.source !== SourceName.SQL_SERVER) {
throw new Error("Datasource source is not SQL Server")
}
const pool = new mssql.ConnectionPool(ds.config! as mssql.config)
const client = await pool.connect()
try {
const { recordset } = await client.query(sql)
return recordset
} finally {
await pool.close()
} }
} }

View File

@ -1,8 +1,11 @@
import { Datasource, SourceName } from "@budibase/types" import { Datasource, SourceName } from "@budibase/types"
import { GenericContainer, Wait, StartedTestContainer } from "testcontainers" import { GenericContainer, Wait } from "testcontainers"
import { AbstractWaitStrategy } from "testcontainers/build/wait-strategies/wait-strategy" import { AbstractWaitStrategy } from "testcontainers/build/wait-strategies/wait-strategy"
import mysql from "mysql2/promise"
import { generator, testContainerUtils } from "@budibase/backend-core/tests"
import { startContainer } from "."
let container: StartedTestContainer | undefined let ports: Promise<testContainerUtils.Port[]>
class MySQLWaitStrategy extends AbstractWaitStrategy { class MySQLWaitStrategy extends AbstractWaitStrategy {
async waitUntilReady(container: any, boundPorts: any, startTime?: Date) { async waitUntilReady(container: any, boundPorts: any, startTime?: Date) {
@ -24,38 +27,50 @@ class MySQLWaitStrategy extends AbstractWaitStrategy {
} }
} }
export async function start(): Promise<StartedTestContainer> { export async function getDatasource(): Promise<Datasource> {
return await new GenericContainer("mysql:8.3") if (!ports) {
.withExposedPorts(3306) ports = startContainer(
.withEnvironment({ MYSQL_ROOT_PASSWORD: "password" }) new GenericContainer("mysql:8.3")
.withWaitStrategy(new MySQLWaitStrategy().withStartupTimeout(10000)) .withExposedPorts(3306)
.start() .withEnvironment({ MYSQL_ROOT_PASSWORD: "password" })
} .withWaitStrategy(new MySQLWaitStrategy().withStartupTimeout(10000))
)
export async function datasource(): Promise<Datasource> {
if (!container) {
container = await start()
} }
const host = container.getHost()
const port = container.getMappedPort(3306)
return { const port = (await ports).find(x => x.container === 3306)?.host
const datasource: Datasource = {
type: "datasource_plus", type: "datasource_plus",
source: SourceName.MYSQL, source: SourceName.MYSQL,
plus: true, plus: true,
config: { config: {
host, host: "127.0.0.1",
port, port,
user: "root", user: "root",
password: "password", password: "password",
database: "mysql", database: "mysql",
}, },
} }
const database = generator.guid().replaceAll("-", "")
await rawQuery(datasource, `CREATE DATABASE \`${database}\``)
datasource.config!.database = database
return datasource
} }
export async function stop() { export async function rawQuery(ds: Datasource, sql: string) {
if (container) { if (!ds.config) {
await container.stop() throw new Error("Datasource config is missing")
container = undefined }
if (ds.source !== SourceName.MYSQL) {
throw new Error("Datasource source is not MySQL")
}
const connection = await mysql.createConnection(ds.config)
try {
const [rows] = await connection.query(sql)
return rows
} finally {
connection.end()
} }
} }

View File

@ -1,33 +1,33 @@
import { Datasource, SourceName } from "@budibase/types" import { Datasource, SourceName } from "@budibase/types"
import { GenericContainer, Wait, StartedTestContainer } from "testcontainers" import { GenericContainer, Wait } from "testcontainers"
import pg from "pg"
import { generator, testContainerUtils } from "@budibase/backend-core/tests"
import { startContainer } from "."
let container: StartedTestContainer | undefined let ports: Promise<testContainerUtils.Port[]>
export async function start(): Promise<StartedTestContainer> { export async function getDatasource(): Promise<Datasource> {
return await new GenericContainer("postgres:16.1-bullseye") if (!ports) {
.withExposedPorts(5432) ports = startContainer(
.withEnvironment({ POSTGRES_PASSWORD: "password" }) new GenericContainer("postgres:16.1-bullseye")
.withWaitStrategy( .withExposedPorts(5432)
Wait.forSuccessfulCommand( .withEnvironment({ POSTGRES_PASSWORD: "password" })
"pg_isready -h localhost -p 5432" .withWaitStrategy(
).withStartupTimeout(10000) Wait.forSuccessfulCommand(
"pg_isready -h localhost -p 5432"
).withStartupTimeout(10000)
)
) )
.start()
}
export async function datasource(): Promise<Datasource> {
if (!container) {
container = await start()
} }
const host = container.getHost()
const port = container.getMappedPort(5432)
return { const port = (await ports).find(x => x.container === 5432)?.host
const datasource: Datasource = {
type: "datasource_plus", type: "datasource_plus",
source: SourceName.POSTGRES, source: SourceName.POSTGRES,
plus: true, plus: true,
config: { config: {
host, host: "127.0.0.1",
port, port,
database: "postgres", database: "postgres",
user: "postgres", user: "postgres",
@ -38,11 +38,28 @@ export async function datasource(): Promise<Datasource> {
ca: false, ca: false,
}, },
} }
const database = generator.guid().replaceAll("-", "")
await rawQuery(datasource, `CREATE DATABASE "${database}"`)
datasource.config!.database = database
return datasource
} }
export async function stop() { export async function rawQuery(ds: Datasource, sql: string) {
if (container) { if (!ds.config) {
await container.stop() throw new Error("Datasource config is missing")
container = undefined }
if (ds.source !== SourceName.POSTGRES) {
throw new Error("Datasource source is not Postgres")
}
const client = new pg.Client(ds.config)
await client.connect()
try {
const { rows } = await client.query(sql)
return rows
} finally {
await client.end()
} }
} }

View File

@ -25,8 +25,6 @@ const clearMigrations = async () => {
} }
} }
jest.setTimeout(10000)
describe("migrations", () => { describe("migrations", () => {
const config = new TestConfig() const config = new TestConfig()

View File

@ -17,8 +17,6 @@ import {
generator, generator,
} from "@budibase/backend-core/tests" } from "@budibase/backend-core/tests"
jest.setTimeout(30000)
describe("external search", () => { describe("external search", () => {
const config = new TestConfiguration() const config = new TestConfiguration()

View File

@ -2,17 +2,11 @@ import env from "../environment"
import { env as coreEnv, timers } from "@budibase/backend-core" import { env as coreEnv, timers } from "@budibase/backend-core"
import { testContainerUtils } from "@budibase/backend-core/tests" import { testContainerUtils } from "@budibase/backend-core/tests"
if (!process.env.DEBUG) {
global.console.log = jest.fn() // console.log are ignored in tests
global.console.warn = jest.fn() // console.warn are ignored in tests
}
if (!process.env.CI) { if (!process.env.CI) {
// set a longer timeout in dev for debugging // set a longer timeout in dev for debugging 100 seconds
// 100 seconds
jest.setTimeout(100 * 1000) jest.setTimeout(100 * 1000)
} else { } else {
jest.setTimeout(10 * 1000) jest.setTimeout(30 * 1000)
} }
testContainerUtils.setupEnv(env, coreEnv) testContainerUtils.setupEnv(env, coreEnv)

View File

@ -1,6 +1,7 @@
import TestConfiguration from "../TestConfiguration" import TestConfiguration from "../TestConfiguration"
import { SuperTest, Test, Response } from "supertest" import request, { SuperTest, Test, Response } from "supertest"
import { ReadStream } from "fs" import { ReadStream } from "fs"
import { getServer } from "../../../app"
type Headers = Record<string, string | string[] | undefined> type Headers = Record<string, string | string[] | undefined>
type Method = "get" | "post" | "put" | "patch" | "delete" type Method = "get" | "post" | "put" | "patch" | "delete"
@ -76,7 +77,8 @@ export abstract class TestAPI {
protected _requestRaw = async ( protected _requestRaw = async (
method: "get" | "post" | "put" | "patch" | "delete", method: "get" | "post" | "put" | "patch" | "delete",
url: string, url: string,
opts?: RequestOpts opts?: RequestOpts,
attempt = 0
): Promise<Response> => { ): Promise<Response> => {
const { const {
headers = {}, headers = {},
@ -107,26 +109,29 @@ export abstract class TestAPI {
const headersFn = publicUser const headersFn = publicUser
? this.config.publicHeaders.bind(this.config) ? this.config.publicHeaders.bind(this.config)
: this.config.defaultHeaders.bind(this.config) : this.config.defaultHeaders.bind(this.config)
let request = this.request[method](url).set(
const app = getServer()
let req = request(app)[method](url)
req = req.set(
headersFn({ headersFn({
"x-budibase-include-stacktrace": "true", "x-budibase-include-stacktrace": "true",
}) })
) )
if (headers) { if (headers) {
request = request.set(headers) req = req.set(headers)
} }
if (body) { if (body) {
request = request.send(body) req = req.send(body)
} }
for (const [key, value] of Object.entries(fields)) { for (const [key, value] of Object.entries(fields)) {
request = request.field(key, value) req = req.field(key, value)
} }
for (const [key, value] of Object.entries(files)) { for (const [key, value] of Object.entries(files)) {
if (isAttachedFile(value)) { if (isAttachedFile(value)) {
request = request.attach(key, value.file, value.name) req = req.attach(key, value.file, value.name)
} else { } else {
request = request.attach(key, value as any) req = req.attach(key, value as any)
} }
} }
if (expectations?.headers) { if (expectations?.headers) {
@ -136,11 +141,25 @@ export abstract class TestAPI {
`Got an undefined expected value for header "${key}", if you want to check for the absence of a header, use headersNotPresent` `Got an undefined expected value for header "${key}", if you want to check for the absence of a header, use headersNotPresent`
) )
} }
request = request.expect(key, value as any) req = req.expect(key, value as any)
} }
} }
return await request try {
return await req
} catch (e: any) {
// We've found that occasionally the connection between supertest and the
// server supertest starts gets reset. Not sure why, but retrying it
// appears to work. I don't particularly like this, but it's better than
// flakiness.
if (e.code === "ECONNRESET") {
if (attempt > 2) {
throw e
}
return await this._requestRaw(method, url, opts, attempt + 1)
}
throw e
}
} }
protected _checkResponse = ( protected _checkResponse = (
@ -170,7 +189,18 @@ export abstract class TestAPI {
} }
} }
throw new Error(message) if (response.error) {
// Sometimes the error can be between supertest and the app, and when
// that happens response.error is sometimes populated with `text` that
// gives more detail about the error. The `message` is almost always
// useless from what I've seen.
if (response.error.text) {
response.error.message = response.error.text
}
throw new Error(message, { cause: response.error })
} else {
throw new Error(message)
}
} }
if (expectations?.headersNotPresent) { if (expectations?.headersNotPresent) {

View File

@ -5875,6 +5875,13 @@
"@types/pouchdb-node" "*" "@types/pouchdb-node" "*"
"@types/pouchdb-replication" "*" "@types/pouchdb-replication" "*"
"@types/proper-lockfile@^4.1.4":
version "4.1.4"
resolved "https://registry.yarnpkg.com/@types/proper-lockfile/-/proper-lockfile-4.1.4.tgz#cd9fab92bdb04730c1ada542c356f03620f84008"
integrity sha512-uo2ABllncSqg9F1D4nugVl9v93RmjxF6LJzQLMLDdPaXCUIDPeOJ21Gbqi43xNKzBi/WQ0Q0dICqufzQbMjipQ==
dependencies:
"@types/retry" "*"
"@types/qs@*": "@types/qs@*":
version "6.9.7" version "6.9.7"
resolved "https://registry.yarnpkg.com/@types/qs/-/qs-6.9.7.tgz#63bb7d067db107cc1e457c303bc25d511febf6cb" resolved "https://registry.yarnpkg.com/@types/qs/-/qs-6.9.7.tgz#63bb7d067db107cc1e457c303bc25d511febf6cb"
@ -5937,6 +5944,11 @@
dependencies: dependencies:
"@types/node" "*" "@types/node" "*"
"@types/retry@*":
version "0.12.5"
resolved "https://registry.yarnpkg.com/@types/retry/-/retry-0.12.5.tgz#f090ff4bd8d2e5b940ff270ab39fd5ca1834a07e"
integrity sha512-3xSjTp3v03X/lSQLkczaN9UIEwJMoMCA1+Nb5HfbJEQWogdeQIyVtTvxPXDQjZ5zws8rFQfVfRdz03ARihPJgw==
"@types/rimraf@^3.0.2": "@types/rimraf@^3.0.2":
version "3.0.2" version "3.0.2"
resolved "https://registry.yarnpkg.com/@types/rimraf/-/rimraf-3.0.2.tgz#a63d175b331748e5220ad48c901d7bbf1f44eef8" resolved "https://registry.yarnpkg.com/@types/rimraf/-/rimraf-3.0.2.tgz#a63d175b331748e5220ad48c901d7bbf1f44eef8"