Merge remote-tracking branch 'origin/master' into feature/screen-deselect
This commit is contained in:
commit
e922f08bb5
|
@ -1,5 +1,5 @@
|
||||||
{
|
{
|
||||||
"version": "2.22.16",
|
"version": "2.22.18",
|
||||||
"npmClient": "yarn",
|
"npmClient": "yarn",
|
||||||
"packages": [
|
"packages": [
|
||||||
"packages/*",
|
"packages/*",
|
||||||
|
|
|
@ -1 +1 @@
|
||||||
Subproject commit 532c4db35cecd346b5c24f0b89ab7b397a122a36
|
Subproject commit a0ee9cad8cefb8f9f40228705711be174f018fa9
|
|
@ -45,6 +45,7 @@ type GroupFns = {
|
||||||
getGroupBuilderAppIds: GroupBuildersFn
|
getGroupBuilderAppIds: GroupBuildersFn
|
||||||
}
|
}
|
||||||
type CreateAdminUserOpts = {
|
type CreateAdminUserOpts = {
|
||||||
|
password?: string
|
||||||
ssoId?: string
|
ssoId?: string
|
||||||
hashPassword?: boolean
|
hashPassword?: boolean
|
||||||
requirePassword?: boolean
|
requirePassword?: boolean
|
||||||
|
@ -501,9 +502,9 @@ export class UserDB {
|
||||||
static async createAdminUser(
|
static async createAdminUser(
|
||||||
email: string,
|
email: string,
|
||||||
tenantId: string,
|
tenantId: string,
|
||||||
password?: string,
|
|
||||||
opts?: CreateAdminUserOpts
|
opts?: CreateAdminUserOpts
|
||||||
) {
|
) {
|
||||||
|
const password = opts?.password
|
||||||
const user: User = {
|
const user: User = {
|
||||||
email: email,
|
email: email,
|
||||||
password,
|
password,
|
||||||
|
|
|
@ -72,7 +72,7 @@
|
||||||
"fast-json-patch": "^3.1.1",
|
"fast-json-patch": "^3.1.1",
|
||||||
"json-format-highlight": "^1.0.4",
|
"json-format-highlight": "^1.0.4",
|
||||||
"lodash": "4.17.21",
|
"lodash": "4.17.21",
|
||||||
"posthog-js": "^1.116.6",
|
"posthog-js": "^1.118.0",
|
||||||
"remixicon": "2.5.0",
|
"remixicon": "2.5.0",
|
||||||
"sanitize-html": "^2.7.0",
|
"sanitize-html": "^2.7.0",
|
||||||
"shortid": "2.2.15",
|
"shortid": "2.2.15",
|
||||||
|
|
|
@ -38,6 +38,10 @@ class AnalyticsHub {
|
||||||
intercom.show(user)
|
intercom.show(user)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
initPosthog() {
|
||||||
|
posthog.init()
|
||||||
|
}
|
||||||
|
|
||||||
async logout() {
|
async logout() {
|
||||||
posthog.logout()
|
posthog.logout()
|
||||||
intercom.logout()
|
intercom.logout()
|
||||||
|
|
|
@ -49,7 +49,10 @@
|
||||||
label: "Long Form Text",
|
label: "Long Form Text",
|
||||||
value: FIELDS.LONGFORM.type,
|
value: FIELDS.LONGFORM.type,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
label: "Attachment",
|
||||||
|
value: FIELDS.ATTACHMENT.type,
|
||||||
|
},
|
||||||
{
|
{
|
||||||
label: "User",
|
label: "User",
|
||||||
value: `${FIELDS.USER.type}${FIELDS.USER.subtype}`,
|
value: `${FIELDS.USER.type}${FIELDS.USER.subtype}`,
|
||||||
|
|
|
@ -33,13 +33,10 @@
|
||||||
import { TOUR_STEP_KEYS } from "components/portal/onboarding/tours.js"
|
import { TOUR_STEP_KEYS } from "components/portal/onboarding/tours.js"
|
||||||
import { goto } from "@roxi/routify"
|
import { goto } from "@roxi/routify"
|
||||||
import { onMount } from "svelte"
|
import { onMount } from "svelte"
|
||||||
import PosthogClient from "../../analytics/PosthogClient"
|
|
||||||
|
|
||||||
export let application
|
export let application
|
||||||
export let loaded
|
export let loaded
|
||||||
|
|
||||||
const posthog = new PosthogClient(process.env.POSTHOG_TOKEN)
|
|
||||||
|
|
||||||
let unpublishModal
|
let unpublishModal
|
||||||
let updateAppModal
|
let updateAppModal
|
||||||
let revertModal
|
let revertModal
|
||||||
|
@ -156,7 +153,7 @@
|
||||||
}
|
}
|
||||||
|
|
||||||
onMount(() => {
|
onMount(() => {
|
||||||
posthog.init()
|
analytics.initPosthog()
|
||||||
})
|
})
|
||||||
</script>
|
</script>
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,6 @@
|
||||||
import { Checkbox, Select, RadioGroup, Stepper, Input } from "@budibase/bbui"
|
import { Checkbox, Select, RadioGroup, Stepper, Input } from "@budibase/bbui"
|
||||||
|
import { licensing } from "stores/portal"
|
||||||
|
import { get } from "svelte/store"
|
||||||
import DataSourceSelect from "./controls/DataSourceSelect/DataSourceSelect.svelte"
|
import DataSourceSelect from "./controls/DataSourceSelect/DataSourceSelect.svelte"
|
||||||
import S3DataSourceSelect from "./controls/S3DataSourceSelect.svelte"
|
import S3DataSourceSelect from "./controls/S3DataSourceSelect.svelte"
|
||||||
import DataProviderSelect from "./controls/DataProviderSelect.svelte"
|
import DataProviderSelect from "./controls/DataProviderSelect.svelte"
|
||||||
|
@ -26,7 +28,8 @@ import FieldConfiguration from "./controls/FieldConfiguration/FieldConfiguration
|
||||||
import ButtonConfiguration from "./controls/ButtonConfiguration/ButtonConfiguration.svelte"
|
import ButtonConfiguration from "./controls/ButtonConfiguration/ButtonConfiguration.svelte"
|
||||||
import RelationshipFilterEditor from "./controls/RelationshipFilterEditor.svelte"
|
import RelationshipFilterEditor from "./controls/RelationshipFilterEditor.svelte"
|
||||||
import FormStepConfiguration from "./controls/FormStepConfiguration.svelte"
|
import FormStepConfiguration from "./controls/FormStepConfiguration.svelte"
|
||||||
import FormStepControls from "components/design/settings/controls/FormStepControls.svelte"
|
import FormStepControls from "./controls/FormStepControls.svelte"
|
||||||
|
import PaywalledSetting from "./controls/PaywalledSetting.svelte"
|
||||||
|
|
||||||
const componentMap = {
|
const componentMap = {
|
||||||
text: DrawerBindableInput,
|
text: DrawerBindableInput,
|
||||||
|
@ -86,11 +89,16 @@ const componentMap = {
|
||||||
}
|
}
|
||||||
|
|
||||||
export const getComponentForSetting = setting => {
|
export const getComponentForSetting = setting => {
|
||||||
const { type, showInBar, barStyle } = setting || {}
|
const { type, showInBar, barStyle, license } = setting || {}
|
||||||
if (!type) {
|
if (!type) {
|
||||||
return null
|
return null
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Check for paywalled settings
|
||||||
|
if (license && get(licensing).isFreePlan) {
|
||||||
|
return PaywalledSetting
|
||||||
|
}
|
||||||
|
|
||||||
// We can show a clone of the bar settings for certain select settings
|
// We can show a clone of the bar settings for certain select settings
|
||||||
if (showInBar && type === "select" && barStyle === "buttons") {
|
if (showInBar && type === "select" && barStyle === "buttons") {
|
||||||
return BarButtonList
|
return BarButtonList
|
||||||
|
|
|
@ -0,0 +1,23 @@
|
||||||
|
<script>
|
||||||
|
import { Tag, Tags } from "@budibase/bbui"
|
||||||
|
import { getFormattedPlanName } from "helpers/planTitle"
|
||||||
|
|
||||||
|
export let license
|
||||||
|
|
||||||
|
$: title = getFormattedPlanName(license)
|
||||||
|
</script>
|
||||||
|
|
||||||
|
<div>
|
||||||
|
<Tags>
|
||||||
|
<Tag icon="LockClosed">{title}</Tag>
|
||||||
|
</Tags>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<style>
|
||||||
|
div {
|
||||||
|
height: 32px;
|
||||||
|
display: flex;
|
||||||
|
flex-direction: column;
|
||||||
|
justify-content: center;
|
||||||
|
}
|
||||||
|
</style>
|
|
@ -183,6 +183,7 @@
|
||||||
props={{
|
props={{
|
||||||
// Generic settings
|
// Generic settings
|
||||||
placeholder: setting.placeholder || null,
|
placeholder: setting.placeholder || null,
|
||||||
|
license: setting.license,
|
||||||
|
|
||||||
// Select settings
|
// Select settings
|
||||||
options: setting.options || [],
|
options: setting.options || [],
|
||||||
|
|
|
@ -4610,6 +4610,35 @@
|
||||||
"key": "dataSource",
|
"key": "dataSource",
|
||||||
"required": true
|
"required": true
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"type": "select",
|
||||||
|
"label": "Auto-refresh",
|
||||||
|
"key": "autoRefresh",
|
||||||
|
"license": "premium",
|
||||||
|
"placeholder": "Never",
|
||||||
|
"options": [
|
||||||
|
{
|
||||||
|
"label": "10 seconds",
|
||||||
|
"value": 10
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "30 seconds",
|
||||||
|
"value": 30
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "1 minute",
|
||||||
|
"value": 60
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "5 minutes",
|
||||||
|
"value": 300
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "10 minutes",
|
||||||
|
"value": 600
|
||||||
|
}
|
||||||
|
]
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"type": "filter",
|
"type": "filter",
|
||||||
"label": "Filtering",
|
"label": "Filtering",
|
||||||
|
@ -4977,6 +5006,35 @@
|
||||||
"key": "dataSource",
|
"key": "dataSource",
|
||||||
"required": true
|
"required": true
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"type": "select",
|
||||||
|
"label": "Auto-refresh",
|
||||||
|
"key": "autoRefresh",
|
||||||
|
"license": "premium",
|
||||||
|
"placeholder": "Never",
|
||||||
|
"options": [
|
||||||
|
{
|
||||||
|
"label": "10 seconds",
|
||||||
|
"value": 10
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "30 seconds",
|
||||||
|
"value": 30
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "1 minute",
|
||||||
|
"value": 60
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "5 minutes",
|
||||||
|
"value": 300
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "10 minutes",
|
||||||
|
"value": 600
|
||||||
|
}
|
||||||
|
]
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"type": "text",
|
"type": "text",
|
||||||
"label": "Title",
|
"label": "Title",
|
||||||
|
@ -5445,6 +5503,35 @@
|
||||||
"key": "dataSource",
|
"key": "dataSource",
|
||||||
"required": true
|
"required": true
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"type": "select",
|
||||||
|
"label": "Auto-refresh",
|
||||||
|
"key": "autoRefresh",
|
||||||
|
"license": "premium",
|
||||||
|
"placeholder": "Never",
|
||||||
|
"options": [
|
||||||
|
{
|
||||||
|
"label": "10 seconds",
|
||||||
|
"value": 10
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "30 seconds",
|
||||||
|
"value": 30
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "1 minute",
|
||||||
|
"value": 60
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "5 minutes",
|
||||||
|
"value": 300
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "10 minutes",
|
||||||
|
"value": 600
|
||||||
|
}
|
||||||
|
]
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"type": "columns",
|
"type": "columns",
|
||||||
"label": "Columns",
|
"label": "Columns",
|
||||||
|
@ -5731,6 +5818,35 @@
|
||||||
"key": "dataSource",
|
"key": "dataSource",
|
||||||
"required": true
|
"required": true
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"type": "select",
|
||||||
|
"label": "Auto-refresh",
|
||||||
|
"key": "autoRefresh",
|
||||||
|
"license": "premium",
|
||||||
|
"placeholder": "Never",
|
||||||
|
"options": [
|
||||||
|
{
|
||||||
|
"label": "10 seconds",
|
||||||
|
"value": 10
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "30 seconds",
|
||||||
|
"value": 30
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "1 minute",
|
||||||
|
"value": 60
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "5 minutes",
|
||||||
|
"value": 300
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "10 minutes",
|
||||||
|
"value": 600
|
||||||
|
}
|
||||||
|
]
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"type": "searchfield",
|
"type": "searchfield",
|
||||||
"label": "Search columns",
|
"label": "Search columns",
|
||||||
|
@ -5908,6 +6024,35 @@
|
||||||
"key": "dataSource",
|
"key": "dataSource",
|
||||||
"required": true
|
"required": true
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"type": "select",
|
||||||
|
"label": "Auto-refresh",
|
||||||
|
"key": "autoRefresh",
|
||||||
|
"license": "premium",
|
||||||
|
"placeholder": "Never",
|
||||||
|
"options": [
|
||||||
|
{
|
||||||
|
"label": "10 seconds",
|
||||||
|
"value": 10
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "30 seconds",
|
||||||
|
"value": 30
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "1 minute",
|
||||||
|
"value": 60
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "5 minutes",
|
||||||
|
"value": 300
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "10 minutes",
|
||||||
|
"value": 600
|
||||||
|
}
|
||||||
|
]
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"type": "filter",
|
"type": "filter",
|
||||||
"label": "Filtering",
|
"label": "Filtering",
|
||||||
|
@ -6504,6 +6649,35 @@
|
||||||
"key": "dataSource",
|
"key": "dataSource",
|
||||||
"required": true
|
"required": true
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"type": "select",
|
||||||
|
"label": "Auto-refresh",
|
||||||
|
"key": "autoRefresh",
|
||||||
|
"license": "premium",
|
||||||
|
"placeholder": "Never",
|
||||||
|
"options": [
|
||||||
|
{
|
||||||
|
"label": "10 seconds",
|
||||||
|
"value": 10
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "30 seconds",
|
||||||
|
"value": 30
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "1 minute",
|
||||||
|
"value": 60
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "5 minutes",
|
||||||
|
"value": 300
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"label": "10 minutes",
|
||||||
|
"value": 600
|
||||||
|
}
|
||||||
|
]
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"type": "text",
|
"type": "text",
|
||||||
"label": "Height",
|
"label": "Height",
|
||||||
|
|
|
@ -5,29 +5,29 @@
|
||||||
import Provider from "./context/Provider.svelte"
|
import Provider from "./context/Provider.svelte"
|
||||||
import { onMount, getContext } from "svelte"
|
import { onMount, getContext } from "svelte"
|
||||||
import { enrichButtonActions } from "../utils/buttonActions.js"
|
import { enrichButtonActions } from "../utils/buttonActions.js"
|
||||||
|
import { memo } from "@budibase/frontend-core"
|
||||||
|
|
||||||
export let params = {}
|
export let params = {}
|
||||||
|
|
||||||
const context = getContext("context")
|
const context = getContext("context")
|
||||||
|
const onLoadActions = memo()
|
||||||
|
|
||||||
// Get the screen definition for the current route
|
// Get the screen definition for the current route
|
||||||
$: screenDefinition = $screenStore.activeScreen?.props
|
$: screenDefinition = $screenStore.activeScreen?.props
|
||||||
|
$: onLoadActions.set($screenStore.activeScreen?.onLoad)
|
||||||
$: runOnLoadActions(params)
|
$: runOnLoadActions($onLoadActions, params)
|
||||||
|
|
||||||
// Enrich and execute any on load actions.
|
// Enrich and execute any on load actions.
|
||||||
// We manually construct the full context here as this component is the
|
// We manually construct the full context here as this component is the
|
||||||
// one that provides the url context, so it is not available in $context yet
|
// one that provides the url context, so it is not available in $context yet
|
||||||
const runOnLoadActions = params => {
|
const runOnLoadActions = (actions, params) => {
|
||||||
const screenState = get(screenStore)
|
if (actions?.length && !get(builderStore).inBuilder) {
|
||||||
|
const enrichedActions = enrichButtonActions(actions, {
|
||||||
if (screenState.activeScreen?.onLoad && !get(builderStore).inBuilder) {
|
|
||||||
const actions = enrichButtonActions(screenState.activeScreen.onLoad, {
|
|
||||||
...get(context),
|
...get(context),
|
||||||
url: params,
|
url: params,
|
||||||
})
|
})
|
||||||
if (actions != null) {
|
if (enrichedActions != null) {
|
||||||
actions()
|
enrichedActions()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -9,17 +9,18 @@
|
||||||
export let sortOrder
|
export let sortOrder
|
||||||
export let limit
|
export let limit
|
||||||
export let paginate
|
export let paginate
|
||||||
|
export let autoRefresh
|
||||||
|
|
||||||
const { styleable, Provider, ActionTypes, API } = getContext("sdk")
|
const { styleable, Provider, ActionTypes, API } = getContext("sdk")
|
||||||
const component = getContext("component")
|
const component = getContext("component")
|
||||||
|
|
||||||
|
let interval
|
||||||
|
let queryExtensions = {}
|
||||||
|
|
||||||
// We need to manage our lucene query manually as we want to allow components
|
// We need to manage our lucene query manually as we want to allow components
|
||||||
// to extend it
|
// to extend it
|
||||||
let queryExtensions = {}
|
|
||||||
$: defaultQuery = LuceneUtils.buildLuceneQuery(filter)
|
$: defaultQuery = LuceneUtils.buildLuceneQuery(filter)
|
||||||
$: query = extendQuery(defaultQuery, queryExtensions)
|
$: query = extendQuery(defaultQuery, queryExtensions)
|
||||||
|
|
||||||
// Fetch data and refresh when needed
|
|
||||||
$: fetch = createFetch(dataSource)
|
$: fetch = createFetch(dataSource)
|
||||||
$: fetch.update({
|
$: fetch.update({
|
||||||
query,
|
query,
|
||||||
|
@ -28,11 +29,8 @@
|
||||||
limit,
|
limit,
|
||||||
paginate,
|
paginate,
|
||||||
})
|
})
|
||||||
|
|
||||||
// Sanitize schema to remove hidden fields
|
|
||||||
$: schema = sanitizeSchema($fetch.schema)
|
$: schema = sanitizeSchema($fetch.schema)
|
||||||
|
$: setUpAutoRefresh(autoRefresh)
|
||||||
// Build our action context
|
|
||||||
$: actions = [
|
$: actions = [
|
||||||
{
|
{
|
||||||
type: ActionTypes.RefreshDatasource,
|
type: ActionTypes.RefreshDatasource,
|
||||||
|
@ -63,8 +61,6 @@
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
]
|
]
|
||||||
|
|
||||||
// Build our data context
|
|
||||||
$: dataContext = {
|
$: dataContext = {
|
||||||
rows: $fetch.rows,
|
rows: $fetch.rows,
|
||||||
info: $fetch.info,
|
info: $fetch.info,
|
||||||
|
@ -140,6 +136,13 @@
|
||||||
})
|
})
|
||||||
return extendedQuery
|
return extendedQuery
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const setUpAutoRefresh = autoRefresh => {
|
||||||
|
clearInterval(interval)
|
||||||
|
if (autoRefresh) {
|
||||||
|
interval = setInterval(fetch.refresh, Math.max(10000, autoRefresh * 1000))
|
||||||
|
}
|
||||||
|
}
|
||||||
</script>
|
</script>
|
||||||
|
|
||||||
<div use:styleable={$component.styles} class="container">
|
<div use:styleable={$component.styles} class="container">
|
||||||
|
|
|
@ -18,6 +18,7 @@
|
||||||
export let columns = null
|
export let columns = null
|
||||||
export let onRowClick = null
|
export let onRowClick = null
|
||||||
export let buttons = null
|
export let buttons = null
|
||||||
|
export let repeat = null
|
||||||
|
|
||||||
const context = getContext("context")
|
const context = getContext("context")
|
||||||
const component = getContext("component")
|
const component = getContext("component")
|
||||||
|
@ -122,6 +123,7 @@
|
||||||
{fixedRowHeight}
|
{fixedRowHeight}
|
||||||
{columnWhitelist}
|
{columnWhitelist}
|
||||||
{schemaOverrides}
|
{schemaOverrides}
|
||||||
|
{repeat}
|
||||||
canAddRows={allowAddRows}
|
canAddRows={allowAddRows}
|
||||||
canEditRows={allowEditRows}
|
canEditRows={allowEditRows}
|
||||||
canDeleteRows={allowDeleteRows}
|
canDeleteRows={allowDeleteRows}
|
||||||
|
|
|
@ -31,6 +31,7 @@
|
||||||
export let cardButtonOnClick
|
export let cardButtonOnClick
|
||||||
export let linkColumn
|
export let linkColumn
|
||||||
export let noRowsMessage
|
export let noRowsMessage
|
||||||
|
export let autoRefresh
|
||||||
|
|
||||||
const context = getContext("context")
|
const context = getContext("context")
|
||||||
const { fetchDatasourceSchema, generateGoldenSample } = getContext("sdk")
|
const { fetchDatasourceSchema, generateGoldenSample } = getContext("sdk")
|
||||||
|
@ -184,6 +185,7 @@
|
||||||
sortOrder,
|
sortOrder,
|
||||||
paginate,
|
paginate,
|
||||||
limit,
|
limit,
|
||||||
|
autoRefresh,
|
||||||
}}
|
}}
|
||||||
order={1}
|
order={1}
|
||||||
>
|
>
|
||||||
|
|
|
@ -8,6 +8,7 @@
|
||||||
export let sortColumn
|
export let sortColumn
|
||||||
export let sortOrder
|
export let sortOrder
|
||||||
export let limit
|
export let limit
|
||||||
|
export let autoRefresh
|
||||||
|
|
||||||
// Block
|
// Block
|
||||||
export let chartTitle
|
export let chartTitle
|
||||||
|
@ -65,6 +66,7 @@
|
||||||
sortColumn,
|
sortColumn,
|
||||||
sortOrder,
|
sortOrder,
|
||||||
limit,
|
limit,
|
||||||
|
autoRefresh,
|
||||||
}}
|
}}
|
||||||
>
|
>
|
||||||
{#if dataProviderId && chartType}
|
{#if dataProviderId && chartType}
|
||||||
|
|
|
@ -17,6 +17,7 @@
|
||||||
export let hAlign
|
export let hAlign
|
||||||
export let vAlign
|
export let vAlign
|
||||||
export let gap
|
export let gap
|
||||||
|
export let autoRefresh
|
||||||
|
|
||||||
const component = getContext("component")
|
const component = getContext("component")
|
||||||
const context = getContext("context")
|
const context = getContext("context")
|
||||||
|
@ -47,6 +48,7 @@
|
||||||
sortOrder,
|
sortOrder,
|
||||||
limit,
|
limit,
|
||||||
paginate,
|
paginate,
|
||||||
|
autoRefresh,
|
||||||
}}
|
}}
|
||||||
>
|
>
|
||||||
{#if $component.empty}
|
{#if $component.empty}
|
||||||
|
|
|
@ -16,6 +16,7 @@
|
||||||
export let detailFields
|
export let detailFields
|
||||||
export let detailTitle
|
export let detailTitle
|
||||||
export let noRowsMessage
|
export let noRowsMessage
|
||||||
|
export let autoRefresh
|
||||||
|
|
||||||
const stateKey = generate()
|
const stateKey = generate()
|
||||||
const context = getContext("context")
|
const context = getContext("context")
|
||||||
|
@ -66,6 +67,7 @@
|
||||||
noValue: false,
|
noValue: false,
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
|
autoRefresh,
|
||||||
}}
|
}}
|
||||||
styles={{
|
styles={{
|
||||||
custom: `
|
custom: `
|
||||||
|
|
|
@ -33,6 +33,7 @@
|
||||||
export let sidePanelSaveLabel
|
export let sidePanelSaveLabel
|
||||||
export let sidePanelDeleteLabel
|
export let sidePanelDeleteLabel
|
||||||
export let notificationOverride
|
export let notificationOverride
|
||||||
|
export let autoRefresh
|
||||||
|
|
||||||
const { fetchDatasourceSchema, API, generateGoldenSample } = getContext("sdk")
|
const { fetchDatasourceSchema, API, generateGoldenSample } = getContext("sdk")
|
||||||
const component = getContext("component")
|
const component = getContext("component")
|
||||||
|
@ -243,6 +244,7 @@
|
||||||
sortOrder,
|
sortOrder,
|
||||||
paginate,
|
paginate,
|
||||||
limit: rowCount,
|
limit: rowCount,
|
||||||
|
autoRefresh,
|
||||||
}}
|
}}
|
||||||
context="provider"
|
context="provider"
|
||||||
order={1}
|
order={1}
|
||||||
|
|
|
@ -1 +1 @@
|
||||||
Subproject commit 6b62505be0c0b50a57b4f4980d86541ebdc86428
|
Subproject commit f8e8f87bd52081e1303a5ae92c432ea5b38f3bb4
|
|
@ -1,25 +0,0 @@
|
||||||
const query = jest.fn(() => ({
|
|
||||||
rows: [
|
|
||||||
{
|
|
||||||
a: "string",
|
|
||||||
b: 1,
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}))
|
|
||||||
|
|
||||||
class Client {
|
|
||||||
query = query
|
|
||||||
end = jest.fn(cb => {
|
|
||||||
if (cb) cb()
|
|
||||||
})
|
|
||||||
connect = jest.fn()
|
|
||||||
release = jest.fn()
|
|
||||||
}
|
|
||||||
|
|
||||||
const on = jest.fn()
|
|
||||||
|
|
||||||
module.exports = {
|
|
||||||
Client,
|
|
||||||
queryMock: query,
|
|
||||||
on,
|
|
||||||
}
|
|
|
@ -42,12 +42,6 @@ if (fs.existsSync("../pro/src")) {
|
||||||
|
|
||||||
const config: Config.InitialOptions = {
|
const config: Config.InitialOptions = {
|
||||||
projects: [
|
projects: [
|
||||||
{
|
|
||||||
...baseConfig,
|
|
||||||
displayName: "sequential test",
|
|
||||||
testMatch: ["<rootDir>/**/*.seq.spec.[jt]s"],
|
|
||||||
runner: "jest-serial-runner",
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
...baseConfig,
|
...baseConfig,
|
||||||
testMatch: ["<rootDir>/**/!(*.seq).spec.[jt]s"],
|
testMatch: ["<rootDir>/**/!(*.seq).spec.[jt]s"],
|
||||||
|
@ -60,6 +54,9 @@ const config: Config.InitialOptions = {
|
||||||
"!src/db/views/staticViews.*",
|
"!src/db/views/staticViews.*",
|
||||||
"!src/**/*.spec.{js,ts}",
|
"!src/**/*.spec.{js,ts}",
|
||||||
"!src/tests/**/*.{js,ts}",
|
"!src/tests/**/*.{js,ts}",
|
||||||
|
// The use of coverage in the JS runner breaks tests by inserting
|
||||||
|
// coverage functions into code that will run inside of the isolate.
|
||||||
|
"!src/jsRunner/**/*.{js,ts}",
|
||||||
],
|
],
|
||||||
coverageReporters: ["lcov", "json", "clover"],
|
coverageReporters: ["lcov", "json", "clover"],
|
||||||
}
|
}
|
||||||
|
|
|
@ -143,7 +143,7 @@
|
||||||
"jest": "29.7.0",
|
"jest": "29.7.0",
|
||||||
"jest-openapi": "0.14.2",
|
"jest-openapi": "0.14.2",
|
||||||
"jest-runner": "29.7.0",
|
"jest-runner": "29.7.0",
|
||||||
"jest-serial-runner": "1.2.1",
|
"nock": "13.5.4",
|
||||||
"nodemon": "2.0.15",
|
"nodemon": "2.0.15",
|
||||||
"openapi-typescript": "5.2.0",
|
"openapi-typescript": "5.2.0",
|
||||||
"path-to-regexp": "6.2.0",
|
"path-to-regexp": "6.2.0",
|
||||||
|
|
|
@ -4,11 +4,9 @@ set -e
|
||||||
if [[ -n $CI ]]
|
if [[ -n $CI ]]
|
||||||
then
|
then
|
||||||
export NODE_OPTIONS="--max-old-space-size=4096 --no-node-snapshot $NODE_OPTIONS"
|
export NODE_OPTIONS="--max-old-space-size=4096 --no-node-snapshot $NODE_OPTIONS"
|
||||||
echo "jest --coverage --maxWorkers=4 --forceExit --workerIdleMemoryLimit=2000MB --bail $@"
|
|
||||||
jest --coverage --maxWorkers=4 --forceExit --workerIdleMemoryLimit=2000MB --bail $@
|
jest --coverage --maxWorkers=4 --forceExit --workerIdleMemoryLimit=2000MB --bail $@
|
||||||
else
|
else
|
||||||
# --maxWorkers performs better in development
|
# --maxWorkers performs better in development
|
||||||
export NODE_OPTIONS="--no-node-snapshot $NODE_OPTIONS"
|
export NODE_OPTIONS="--no-node-snapshot $NODE_OPTIONS"
|
||||||
echo "jest --coverage --maxWorkers=2 --forceExit $@"
|
|
||||||
jest --coverage --maxWorkers=2 --forceExit $@
|
jest --coverage --maxWorkers=2 --forceExit $@
|
||||||
fi
|
fi
|
|
@ -1,6 +1,6 @@
|
||||||
import { getQueryParams, getTableParams } from "../../db/utils"
|
import { getQueryParams, getTableParams } from "../../db/utils"
|
||||||
import { getIntegration } from "../../integrations"
|
import { getIntegration } from "../../integrations"
|
||||||
import { invalidateDynamicVariables } from "../../threads/utils"
|
import { invalidateCachedVariable } from "../../threads/utils"
|
||||||
import { context, db as dbCore, events } from "@budibase/backend-core"
|
import { context, db as dbCore, events } from "@budibase/backend-core"
|
||||||
import {
|
import {
|
||||||
BuildSchemaFromSourceRequest,
|
BuildSchemaFromSourceRequest,
|
||||||
|
@ -121,7 +121,7 @@ async function invalidateVariables(
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
await invalidateDynamicVariables(toInvalidate)
|
await invalidateCachedVariable(toInvalidate)
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function update(
|
export async function update(
|
||||||
|
|
|
@ -2,7 +2,7 @@ import { generateQueryID } from "../../../db/utils"
|
||||||
import { Thread, ThreadType } from "../../../threads"
|
import { Thread, ThreadType } from "../../../threads"
|
||||||
import { save as saveDatasource } from "../datasource"
|
import { save as saveDatasource } from "../datasource"
|
||||||
import { RestImporter } from "./import"
|
import { RestImporter } from "./import"
|
||||||
import { invalidateDynamicVariables } from "../../../threads/utils"
|
import { invalidateCachedVariable } from "../../../threads/utils"
|
||||||
import env from "../../../environment"
|
import env from "../../../environment"
|
||||||
import { events, context, utils, constants } from "@budibase/backend-core"
|
import { events, context, utils, constants } from "@budibase/backend-core"
|
||||||
import sdk from "../../../sdk"
|
import sdk from "../../../sdk"
|
||||||
|
@ -281,49 +281,52 @@ export async function preview(
|
||||||
return { previewSchema, nestedSchemaFields }
|
return { previewSchema, nestedSchemaFields }
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const inputs: QueryEvent = {
|
||||||
|
appId: ctx.appId,
|
||||||
|
queryVerb: query.queryVerb,
|
||||||
|
fields: query.fields,
|
||||||
|
parameters: enrichParameters(query),
|
||||||
|
transformer: query.transformer,
|
||||||
|
schema: query.schema,
|
||||||
|
nullDefaultSupport: query.nullDefaultSupport,
|
||||||
|
queryId,
|
||||||
|
datasource,
|
||||||
|
// have to pass down to the thread runner - can't put into context now
|
||||||
|
environmentVariables: envVars,
|
||||||
|
ctx: {
|
||||||
|
user: ctx.user,
|
||||||
|
auth: { ...authConfigCtx },
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
let queryResponse: QueryResponse
|
||||||
try {
|
try {
|
||||||
const inputs: QueryEvent = {
|
queryResponse = await Runner.run<QueryResponse>(inputs)
|
||||||
appId: ctx.appId,
|
|
||||||
queryVerb: query.queryVerb,
|
|
||||||
fields: query.fields,
|
|
||||||
parameters: enrichParameters(query),
|
|
||||||
transformer: query.transformer,
|
|
||||||
schema: query.schema,
|
|
||||||
nullDefaultSupport: query.nullDefaultSupport,
|
|
||||||
queryId,
|
|
||||||
datasource,
|
|
||||||
// have to pass down to the thread runner - can't put into context now
|
|
||||||
environmentVariables: envVars,
|
|
||||||
ctx: {
|
|
||||||
user: ctx.user,
|
|
||||||
auth: { ...authConfigCtx },
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
const { rows, keys, info, extra } = await Runner.run<QueryResponse>(inputs)
|
|
||||||
const { previewSchema, nestedSchemaFields } = getSchemaFields(rows, keys)
|
|
||||||
|
|
||||||
// if existing schema, update to include any previous schema keys
|
|
||||||
if (existingSchema) {
|
|
||||||
for (let key of Object.keys(previewSchema)) {
|
|
||||||
if (existingSchema[key]) {
|
|
||||||
previewSchema[key] = existingSchema[key]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// remove configuration before sending event
|
|
||||||
delete datasource.config
|
|
||||||
await events.query.previewed(datasource, ctx.request.body)
|
|
||||||
ctx.body = {
|
|
||||||
rows,
|
|
||||||
nestedSchemaFields,
|
|
||||||
schema: previewSchema,
|
|
||||||
info,
|
|
||||||
extra,
|
|
||||||
}
|
|
||||||
} catch (err: any) {
|
} catch (err: any) {
|
||||||
ctx.throw(400, err)
|
ctx.throw(400, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const { rows, keys, info, extra } = queryResponse
|
||||||
|
const { previewSchema, nestedSchemaFields } = getSchemaFields(rows, keys)
|
||||||
|
|
||||||
|
// if existing schema, update to include any previous schema keys
|
||||||
|
if (existingSchema) {
|
||||||
|
for (let key of Object.keys(previewSchema)) {
|
||||||
|
if (existingSchema[key]) {
|
||||||
|
previewSchema[key] = existingSchema[key]
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// remove configuration before sending event
|
||||||
|
delete datasource.config
|
||||||
|
await events.query.previewed(datasource, ctx.request.body)
|
||||||
|
ctx.body = {
|
||||||
|
rows,
|
||||||
|
nestedSchemaFields,
|
||||||
|
schema: previewSchema,
|
||||||
|
info,
|
||||||
|
extra,
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
async function execute(
|
async function execute(
|
||||||
|
@ -416,7 +419,7 @@ const removeDynamicVariables = async (queryId: string) => {
|
||||||
const variablesToDelete = dynamicVariables!.filter(
|
const variablesToDelete = dynamicVariables!.filter(
|
||||||
(dv: any) => dv.queryId === queryId
|
(dv: any) => dv.queryId === queryId
|
||||||
)
|
)
|
||||||
await invalidateDynamicVariables(variablesToDelete)
|
await invalidateCachedVariable(variablesToDelete)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -84,8 +84,8 @@ export async function save(ctx: UserCtx<SaveTableRequest, SaveTableResponse>) {
|
||||||
}
|
}
|
||||||
let savedTable = await api.save(ctx, renaming)
|
let savedTable = await api.save(ctx, renaming)
|
||||||
if (!table._id) {
|
if (!table._id) {
|
||||||
await events.table.created(savedTable)
|
|
||||||
savedTable = sdk.tables.enrichViewSchemas(savedTable)
|
savedTable = sdk.tables.enrichViewSchemas(savedTable)
|
||||||
|
await events.table.created(savedTable)
|
||||||
} else {
|
} else {
|
||||||
await events.table.updated(savedTable)
|
await events.table.updated(savedTable)
|
||||||
}
|
}
|
||||||
|
|
|
@ -6,6 +6,7 @@ import {
|
||||||
UIFieldMetadata,
|
UIFieldMetadata,
|
||||||
UpdateViewRequest,
|
UpdateViewRequest,
|
||||||
ViewResponse,
|
ViewResponse,
|
||||||
|
ViewResponseEnriched,
|
||||||
ViewV2,
|
ViewV2,
|
||||||
} from "@budibase/types"
|
} from "@budibase/types"
|
||||||
import { builderSocket, gridSocket } from "../../../websockets"
|
import { builderSocket, gridSocket } from "../../../websockets"
|
||||||
|
@ -39,9 +40,9 @@ async function parseSchema(view: CreateViewRequest) {
|
||||||
return finalViewSchema
|
return finalViewSchema
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function get(ctx: Ctx<void, ViewResponse>) {
|
export async function get(ctx: Ctx<void, ViewResponseEnriched>) {
|
||||||
ctx.body = {
|
ctx.body = {
|
||||||
data: await sdk.views.get(ctx.params.viewId, { enriched: true }),
|
data: await sdk.views.getEnriched(ctx.params.viewId),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,18 +1,16 @@
|
||||||
jest.mock("pg")
|
|
||||||
import * as setup from "./utilities"
|
import * as setup from "./utilities"
|
||||||
import { checkBuilderEndpoint } from "./utilities/TestFunctions"
|
import { checkBuilderEndpoint } from "./utilities/TestFunctions"
|
||||||
import { checkCacheForDynamicVariable } from "../../../threads/utils"
|
import { getCachedVariable } from "../../../threads/utils"
|
||||||
import { context, events } from "@budibase/backend-core"
|
import { context, events } from "@budibase/backend-core"
|
||||||
import sdk from "../../../sdk"
|
import sdk from "../../../sdk"
|
||||||
|
|
||||||
import tk from "timekeeper"
|
import tk from "timekeeper"
|
||||||
import { mocks } from "@budibase/backend-core/tests"
|
import { mocks } from "@budibase/backend-core/tests"
|
||||||
import { QueryPreview } from "@budibase/types"
|
import { QueryPreview, SourceName } from "@budibase/types"
|
||||||
|
|
||||||
tk.freeze(mocks.date.MOCK_DATE)
|
tk.freeze(mocks.date.MOCK_DATE)
|
||||||
|
|
||||||
let { basicDatasource } = setup.structures
|
let { basicDatasource } = setup.structures
|
||||||
const pg = require("pg")
|
|
||||||
|
|
||||||
describe("/datasources", () => {
|
describe("/datasources", () => {
|
||||||
let request = setup.getRequest()
|
let request = setup.getRequest()
|
||||||
|
@ -42,6 +40,23 @@ describe("/datasources", () => {
|
||||||
expect(res.body.errors).toEqual({})
|
expect(res.body.errors).toEqual({})
|
||||||
expect(events.datasource.created).toHaveBeenCalledTimes(1)
|
expect(events.datasource.created).toHaveBeenCalledTimes(1)
|
||||||
})
|
})
|
||||||
|
|
||||||
|
it("should fail if the datasource is invalid", async () => {
|
||||||
|
await config.api.datasource.create(
|
||||||
|
{
|
||||||
|
name: "Test",
|
||||||
|
type: "test",
|
||||||
|
source: "invalid" as SourceName,
|
||||||
|
config: {},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
status: 500,
|
||||||
|
body: {
|
||||||
|
message: "No datasource implementation found.",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("update", () => {
|
describe("update", () => {
|
||||||
|
@ -74,7 +89,7 @@ describe("/datasources", () => {
|
||||||
schema: {},
|
schema: {},
|
||||||
readable: true,
|
readable: true,
|
||||||
}
|
}
|
||||||
return config.api.query.previewQuery(queryPreview)
|
return config.api.query.preview(queryPreview)
|
||||||
}
|
}
|
||||||
|
|
||||||
it("should invalidate changed or removed variables", async () => {
|
it("should invalidate changed or removed variables", async () => {
|
||||||
|
@ -85,10 +100,7 @@ describe("/datasources", () => {
|
||||||
queryString: "test={{ variable3 }}",
|
queryString: "test={{ variable3 }}",
|
||||||
})
|
})
|
||||||
// check variables in cache
|
// check variables in cache
|
||||||
let contents = await checkCacheForDynamicVariable(
|
let contents = await getCachedVariable(query._id!, "variable3")
|
||||||
query._id!,
|
|
||||||
"variable3"
|
|
||||||
)
|
|
||||||
expect(contents.rows.length).toEqual(1)
|
expect(contents.rows.length).toEqual(1)
|
||||||
|
|
||||||
// update the datasource to remove the variables
|
// update the datasource to remove the variables
|
||||||
|
@ -102,7 +114,7 @@ describe("/datasources", () => {
|
||||||
expect(res.body.errors).toBeUndefined()
|
expect(res.body.errors).toBeUndefined()
|
||||||
|
|
||||||
// check variables no longer in cache
|
// check variables no longer in cache
|
||||||
contents = await checkCacheForDynamicVariable(query._id!, "variable3")
|
contents = await getCachedVariable(query._id!, "variable3")
|
||||||
expect(contents).toBe(null)
|
expect(contents).toBe(null)
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
@ -149,35 +161,6 @@ describe("/datasources", () => {
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("query", () => {
|
|
||||||
it("should be able to query a pg datasource", async () => {
|
|
||||||
const res = await request
|
|
||||||
.post(`/api/datasources/query`)
|
|
||||||
.send({
|
|
||||||
endpoint: {
|
|
||||||
datasourceId: datasource._id,
|
|
||||||
operation: "READ",
|
|
||||||
// table name below
|
|
||||||
entityId: "users",
|
|
||||||
},
|
|
||||||
resource: {
|
|
||||||
fields: ["users.name", "users.age"],
|
|
||||||
},
|
|
||||||
filters: {
|
|
||||||
string: {
|
|
||||||
name: "John",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect(200)
|
|
||||||
// this is mock data, can't test it
|
|
||||||
expect(res.body).toBeDefined()
|
|
||||||
const expSql = `select "users"."name" as "users.name", "users"."age" as "users.age" from (select * from "users" where "users"."name" ilike $1 limit $2) as "users"`
|
|
||||||
expect(pg.queryMock).toHaveBeenCalledWith(expSql, ["John%", 5000])
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
describe("destroy", () => {
|
describe("destroy", () => {
|
||||||
beforeAll(setupTest)
|
beforeAll(setupTest)
|
||||||
|
|
||||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -88,345 +88,491 @@ describe("/queries", () => {
|
||||||
})
|
})
|
||||||
|
|
||||||
afterEach(async () => {
|
afterEach(async () => {
|
||||||
await withCollection(async collection => {
|
await withCollection(collection => collection.drop())
|
||||||
await collection.drop()
|
|
||||||
})
|
|
||||||
})
|
})
|
||||||
|
|
||||||
it("should execute a count query", async () => {
|
describe("preview", () => {
|
||||||
const query = await createQuery({
|
it("should generate a nested schema with an empty array", async () => {
|
||||||
fields: {
|
const name = generator.guid()
|
||||||
json: {},
|
await withCollection(
|
||||||
extra: {
|
async collection => await collection.insertOne({ name, nested: [] })
|
||||||
actionType: "count",
|
)
|
||||||
|
|
||||||
|
const preview = await config.api.query.preview({
|
||||||
|
name: "New Query",
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
fields: {
|
||||||
|
json: {
|
||||||
|
name: { $eq: name },
|
||||||
|
},
|
||||||
|
extra: {
|
||||||
|
collection,
|
||||||
|
actionType: "findOne",
|
||||||
|
},
|
||||||
},
|
},
|
||||||
},
|
schema: {},
|
||||||
|
queryVerb: "read",
|
||||||
|
parameters: [],
|
||||||
|
transformer: "return data",
|
||||||
|
readable: true,
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(preview).toEqual({
|
||||||
|
nestedSchemaFields: {},
|
||||||
|
rows: [{ _id: expect.any(String), name, nested: [] }],
|
||||||
|
schema: {
|
||||||
|
_id: {
|
||||||
|
type: "string",
|
||||||
|
name: "_id",
|
||||||
|
},
|
||||||
|
name: {
|
||||||
|
type: "string",
|
||||||
|
name: "name",
|
||||||
|
},
|
||||||
|
nested: {
|
||||||
|
type: "array",
|
||||||
|
name: "nested",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!)
|
it("should generate a nested schema based on all of the nested items", async () => {
|
||||||
|
const name = generator.guid()
|
||||||
|
const item = {
|
||||||
|
name,
|
||||||
|
contacts: [
|
||||||
|
{
|
||||||
|
address: "123 Lane",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
address: "456 Drive",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
postcode: "BT1 12N",
|
||||||
|
lat: 54.59,
|
||||||
|
long: -5.92,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
city: "Belfast",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
address: "789 Avenue",
|
||||||
|
phoneNumber: "0800-999-5555",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Name",
|
||||||
|
isActive: false,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}
|
||||||
|
|
||||||
expect(result.data).toEqual([{ value: 5 }])
|
await withCollection(collection => collection.insertOne(item))
|
||||||
})
|
|
||||||
|
|
||||||
it("should execute a count query with a transformer", async () => {
|
const preview = await config.api.query.preview({
|
||||||
const query = await createQuery({
|
name: "New Query",
|
||||||
fields: {
|
datasourceId: datasource._id!,
|
||||||
json: {},
|
fields: {
|
||||||
extra: {
|
json: {
|
||||||
actionType: "count",
|
name: { $eq: name },
|
||||||
|
},
|
||||||
|
extra: {
|
||||||
|
collection,
|
||||||
|
actionType: "findOne",
|
||||||
|
},
|
||||||
},
|
},
|
||||||
},
|
schema: {},
|
||||||
transformer: "return data + 1",
|
queryVerb: "read",
|
||||||
})
|
parameters: [],
|
||||||
|
transformer: "return data",
|
||||||
|
readable: true,
|
||||||
|
})
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!)
|
expect(preview).toEqual({
|
||||||
|
nestedSchemaFields: {
|
||||||
expect(result.data).toEqual([{ value: 6 }])
|
contacts: {
|
||||||
})
|
address: {
|
||||||
|
type: "string",
|
||||||
it("should execute a find query", async () => {
|
name: "address",
|
||||||
const query = await createQuery({
|
},
|
||||||
fields: {
|
postcode: {
|
||||||
json: {},
|
type: "string",
|
||||||
extra: {
|
name: "postcode",
|
||||||
actionType: "find",
|
},
|
||||||
|
lat: {
|
||||||
|
type: "number",
|
||||||
|
name: "lat",
|
||||||
|
},
|
||||||
|
long: {
|
||||||
|
type: "number",
|
||||||
|
name: "long",
|
||||||
|
},
|
||||||
|
city: {
|
||||||
|
type: "string",
|
||||||
|
name: "city",
|
||||||
|
},
|
||||||
|
phoneNumber: {
|
||||||
|
type: "string",
|
||||||
|
name: "phoneNumber",
|
||||||
|
},
|
||||||
|
name: {
|
||||||
|
type: "string",
|
||||||
|
name: "name",
|
||||||
|
},
|
||||||
|
isActive: {
|
||||||
|
type: "boolean",
|
||||||
|
name: "isActive",
|
||||||
|
},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
},
|
rows: [{ ...item, _id: expect.any(String) }],
|
||||||
})
|
schema: {
|
||||||
|
_id: { type: "string", name: "_id" },
|
||||||
const result = await config.api.query.execute(query._id!)
|
name: { type: "string", name: "name" },
|
||||||
|
contacts: { type: "json", name: "contacts", subtype: "array" },
|
||||||
expect(result.data).toEqual([
|
|
||||||
{ _id: expectValidId, name: "one" },
|
|
||||||
{ _id: expectValidId, name: "two" },
|
|
||||||
{ _id: expectValidId, name: "three" },
|
|
||||||
{ _id: expectValidId, name: "four" },
|
|
||||||
{ _id: expectValidId, name: "five" },
|
|
||||||
])
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should execute a findOne query", async () => {
|
|
||||||
const query = await createQuery({
|
|
||||||
fields: {
|
|
||||||
json: {},
|
|
||||||
extra: {
|
|
||||||
actionType: "findOne",
|
|
||||||
},
|
},
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!)
|
|
||||||
|
|
||||||
expect(result.data).toEqual([{ _id: expectValidId, name: "one" }])
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should execute a findOneAndUpdate query", async () => {
|
|
||||||
const query = await createQuery({
|
|
||||||
fields: {
|
|
||||||
json: {
|
|
||||||
filter: { name: { $eq: "one" } },
|
|
||||||
update: { $set: { name: "newName" } },
|
|
||||||
},
|
|
||||||
extra: {
|
|
||||||
actionType: "findOneAndUpdate",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!)
|
|
||||||
|
|
||||||
expect(result.data).toEqual([
|
|
||||||
{
|
|
||||||
lastErrorObject: { n: 1, updatedExisting: true },
|
|
||||||
ok: 1,
|
|
||||||
value: { _id: expectValidId, name: "one" },
|
|
||||||
},
|
|
||||||
])
|
|
||||||
|
|
||||||
await withCollection(async collection => {
|
|
||||||
expect(await collection.countDocuments()).toBe(5)
|
|
||||||
|
|
||||||
const doc = await collection.findOne({ name: { $eq: "newName" } })
|
|
||||||
expect(doc).toEqual({
|
|
||||||
_id: expectValidBsonObjectId,
|
|
||||||
name: "newName",
|
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
it("should execute a distinct query", async () => {
|
describe("execute", () => {
|
||||||
const query = await createQuery({
|
it("a count query", async () => {
|
||||||
fields: {
|
const query = await createQuery({
|
||||||
json: "name",
|
fields: {
|
||||||
extra: {
|
json: {},
|
||||||
actionType: "distinct",
|
extra: {
|
||||||
|
actionType: "count",
|
||||||
|
},
|
||||||
},
|
},
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!)
|
|
||||||
const values = result.data.map(o => o.value).sort()
|
|
||||||
expect(values).toEqual(["five", "four", "one", "three", "two"])
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should execute a create query with parameters", async () => {
|
|
||||||
const query = await createQuery({
|
|
||||||
fields: {
|
|
||||||
json: { foo: "{{ foo }}" },
|
|
||||||
extra: {
|
|
||||||
actionType: "insertOne",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
queryVerb: "create",
|
|
||||||
parameters: [
|
|
||||||
{
|
|
||||||
name: "foo",
|
|
||||||
default: "default",
|
|
||||||
},
|
|
||||||
],
|
|
||||||
})
|
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!, {
|
|
||||||
parameters: { foo: "bar" },
|
|
||||||
})
|
|
||||||
|
|
||||||
expect(result.data).toEqual([
|
|
||||||
{
|
|
||||||
acknowledged: true,
|
|
||||||
insertedId: expectValidId,
|
|
||||||
},
|
|
||||||
])
|
|
||||||
|
|
||||||
await withCollection(async collection => {
|
|
||||||
const doc = await collection.findOne({ foo: { $eq: "bar" } })
|
|
||||||
expect(doc).toEqual({
|
|
||||||
_id: expectValidBsonObjectId,
|
|
||||||
foo: "bar",
|
|
||||||
})
|
})
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should execute a delete query with parameters", async () => {
|
const result = await config.api.query.execute(query._id!)
|
||||||
const query = await createQuery({
|
|
||||||
fields: {
|
expect(result.data).toEqual([{ value: 5 }])
|
||||||
json: { name: { $eq: "{{ name }}" } },
|
})
|
||||||
extra: {
|
|
||||||
actionType: "deleteOne",
|
it("should be able to updateOne by ObjectId", async () => {
|
||||||
|
const insertResult = await withCollection(c =>
|
||||||
|
c.insertOne({ name: "one" })
|
||||||
|
)
|
||||||
|
const query = await createQuery({
|
||||||
|
fields: {
|
||||||
|
json: {
|
||||||
|
filter: { _id: { $eq: `ObjectId("${insertResult.insertedId}")` } },
|
||||||
|
update: { $set: { name: "newName" } },
|
||||||
|
},
|
||||||
|
extra: {
|
||||||
|
actionType: "updateOne",
|
||||||
|
},
|
||||||
},
|
},
|
||||||
},
|
queryVerb: "update",
|
||||||
queryVerb: "delete",
|
})
|
||||||
parameters: [
|
|
||||||
|
const result = await config.api.query.execute(query._id!)
|
||||||
|
|
||||||
|
expect(result.data).toEqual([
|
||||||
{
|
{
|
||||||
name: "name",
|
acknowledged: true,
|
||||||
default: "",
|
matchedCount: 1,
|
||||||
|
modifiedCount: 1,
|
||||||
|
upsertedCount: 0,
|
||||||
|
upsertedId: null,
|
||||||
},
|
},
|
||||||
],
|
])
|
||||||
})
|
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!, {
|
await withCollection(async collection => {
|
||||||
parameters: { name: "one" },
|
const doc = await collection.findOne({ name: { $eq: "newName" } })
|
||||||
})
|
expect(doc).toEqual({
|
||||||
|
_id: insertResult.insertedId,
|
||||||
expect(result.data).toEqual([
|
|
||||||
{
|
|
||||||
acknowledged: true,
|
|
||||||
deletedCount: 1,
|
|
||||||
},
|
|
||||||
])
|
|
||||||
|
|
||||||
await withCollection(async collection => {
|
|
||||||
const doc = await collection.findOne({ name: { $eq: "one" } })
|
|
||||||
expect(doc).toBeNull()
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should execute an update query with parameters", async () => {
|
|
||||||
const query = await createQuery({
|
|
||||||
fields: {
|
|
||||||
json: {
|
|
||||||
filter: { name: { $eq: "{{ name }}" } },
|
|
||||||
update: { $set: { name: "{{ newName }}" } },
|
|
||||||
},
|
|
||||||
extra: {
|
|
||||||
actionType: "updateOne",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
queryVerb: "update",
|
|
||||||
parameters: [
|
|
||||||
{
|
|
||||||
name: "name",
|
|
||||||
default: "",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "newName",
|
name: "newName",
|
||||||
default: "",
|
})
|
||||||
},
|
|
||||||
],
|
|
||||||
})
|
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!, {
|
|
||||||
parameters: { name: "one", newName: "newOne" },
|
|
||||||
})
|
|
||||||
|
|
||||||
expect(result.data).toEqual([
|
|
||||||
{
|
|
||||||
acknowledged: true,
|
|
||||||
matchedCount: 1,
|
|
||||||
modifiedCount: 1,
|
|
||||||
upsertedCount: 0,
|
|
||||||
upsertedId: null,
|
|
||||||
},
|
|
||||||
])
|
|
||||||
|
|
||||||
await withCollection(async collection => {
|
|
||||||
const doc = await collection.findOne({ name: { $eq: "newOne" } })
|
|
||||||
expect(doc).toEqual({
|
|
||||||
_id: expectValidBsonObjectId,
|
|
||||||
name: "newOne",
|
|
||||||
})
|
|
||||||
|
|
||||||
const oldDoc = await collection.findOne({ name: { $eq: "one" } })
|
|
||||||
expect(oldDoc).toBeNull()
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should be able to updateOne by ObjectId", async () => {
|
|
||||||
const insertResult = await withCollection(c => c.insertOne({ name: "one" }))
|
|
||||||
const query = await createQuery({
|
|
||||||
fields: {
|
|
||||||
json: {
|
|
||||||
filter: { _id: { $eq: `ObjectId("${insertResult.insertedId}")` } },
|
|
||||||
update: { $set: { name: "newName" } },
|
|
||||||
},
|
|
||||||
extra: {
|
|
||||||
actionType: "updateOne",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
queryVerb: "update",
|
|
||||||
})
|
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!)
|
|
||||||
|
|
||||||
expect(result.data).toEqual([
|
|
||||||
{
|
|
||||||
acknowledged: true,
|
|
||||||
matchedCount: 1,
|
|
||||||
modifiedCount: 1,
|
|
||||||
upsertedCount: 0,
|
|
||||||
upsertedId: null,
|
|
||||||
},
|
|
||||||
])
|
|
||||||
|
|
||||||
await withCollection(async collection => {
|
|
||||||
const doc = await collection.findOne({ name: { $eq: "newName" } })
|
|
||||||
expect(doc).toEqual({
|
|
||||||
_id: insertResult.insertedId,
|
|
||||||
name: "newName",
|
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
})
|
|
||||||
|
|
||||||
it("should be able to delete all records", async () => {
|
it("a count query with a transformer", async () => {
|
||||||
const query = await createQuery({
|
const query = await createQuery({
|
||||||
fields: {
|
fields: {
|
||||||
json: {},
|
json: {},
|
||||||
extra: {
|
extra: {
|
||||||
actionType: "deleteMany",
|
actionType: "count",
|
||||||
|
},
|
||||||
},
|
},
|
||||||
},
|
transformer: "return data + 1",
|
||||||
queryVerb: "delete",
|
})
|
||||||
|
|
||||||
|
const result = await config.api.query.execute(query._id!)
|
||||||
|
|
||||||
|
expect(result.data).toEqual([{ value: 6 }])
|
||||||
})
|
})
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!)
|
it("a find query", async () => {
|
||||||
|
const query = await createQuery({
|
||||||
expect(result.data).toEqual([
|
fields: {
|
||||||
{
|
json: {},
|
||||||
acknowledged: true,
|
extra: {
|
||||||
deletedCount: 5,
|
actionType: "find",
|
||||||
},
|
},
|
||||||
])
|
|
||||||
|
|
||||||
await withCollection(async collection => {
|
|
||||||
const docs = await collection.find().toArray()
|
|
||||||
expect(docs).toHaveLength(0)
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should be able to update all documents", async () => {
|
|
||||||
const query = await createQuery({
|
|
||||||
fields: {
|
|
||||||
json: {
|
|
||||||
filter: {},
|
|
||||||
update: { $set: { name: "newName" } },
|
|
||||||
},
|
},
|
||||||
extra: {
|
})
|
||||||
actionType: "updateMany",
|
|
||||||
},
|
const result = await config.api.query.execute(query._id!)
|
||||||
},
|
|
||||||
queryVerb: "update",
|
expect(result.data).toEqual([
|
||||||
|
{ _id: expectValidId, name: "one" },
|
||||||
|
{ _id: expectValidId, name: "two" },
|
||||||
|
{ _id: expectValidId, name: "three" },
|
||||||
|
{ _id: expectValidId, name: "four" },
|
||||||
|
{ _id: expectValidId, name: "five" },
|
||||||
|
])
|
||||||
})
|
})
|
||||||
|
|
||||||
const result = await config.api.query.execute(query._id!)
|
it("a findOne query", async () => {
|
||||||
|
const query = await createQuery({
|
||||||
|
fields: {
|
||||||
|
json: {},
|
||||||
|
extra: {
|
||||||
|
actionType: "findOne",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
expect(result.data).toEqual([
|
const result = await config.api.query.execute(query._id!)
|
||||||
{
|
|
||||||
acknowledged: true,
|
|
||||||
matchedCount: 5,
|
|
||||||
modifiedCount: 5,
|
|
||||||
upsertedCount: 0,
|
|
||||||
upsertedId: null,
|
|
||||||
},
|
|
||||||
])
|
|
||||||
|
|
||||||
await withCollection(async collection => {
|
expect(result.data).toEqual([{ _id: expectValidId, name: "one" }])
|
||||||
const docs = await collection.find().toArray()
|
})
|
||||||
expect(docs).toHaveLength(5)
|
|
||||||
for (const doc of docs) {
|
it("a findOneAndUpdate query", async () => {
|
||||||
|
const query = await createQuery({
|
||||||
|
fields: {
|
||||||
|
json: {
|
||||||
|
filter: { name: { $eq: "one" } },
|
||||||
|
update: { $set: { name: "newName" } },
|
||||||
|
},
|
||||||
|
extra: {
|
||||||
|
actionType: "findOneAndUpdate",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const result = await config.api.query.execute(query._id!)
|
||||||
|
|
||||||
|
expect(result.data).toEqual([
|
||||||
|
{
|
||||||
|
lastErrorObject: { n: 1, updatedExisting: true },
|
||||||
|
ok: 1,
|
||||||
|
value: { _id: expectValidId, name: "one" },
|
||||||
|
},
|
||||||
|
])
|
||||||
|
|
||||||
|
await withCollection(async collection => {
|
||||||
|
expect(await collection.countDocuments()).toBe(5)
|
||||||
|
|
||||||
|
const doc = await collection.findOne({ name: { $eq: "newName" } })
|
||||||
expect(doc).toEqual({
|
expect(doc).toEqual({
|
||||||
_id: expectValidBsonObjectId,
|
_id: expectValidBsonObjectId,
|
||||||
name: "newName",
|
name: "newName",
|
||||||
})
|
})
|
||||||
}
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
it("a distinct query", async () => {
|
||||||
|
const query = await createQuery({
|
||||||
|
fields: {
|
||||||
|
json: "name",
|
||||||
|
extra: {
|
||||||
|
actionType: "distinct",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const result = await config.api.query.execute(query._id!)
|
||||||
|
const values = result.data.map(o => o.value).sort()
|
||||||
|
expect(values).toEqual(["five", "four", "one", "three", "two"])
|
||||||
|
})
|
||||||
|
|
||||||
|
it("a create query with parameters", async () => {
|
||||||
|
const query = await createQuery({
|
||||||
|
fields: {
|
||||||
|
json: { foo: "{{ foo }}" },
|
||||||
|
extra: {
|
||||||
|
actionType: "insertOne",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
queryVerb: "create",
|
||||||
|
parameters: [
|
||||||
|
{
|
||||||
|
name: "foo",
|
||||||
|
default: "default",
|
||||||
|
},
|
||||||
|
],
|
||||||
|
})
|
||||||
|
|
||||||
|
const result = await config.api.query.execute(query._id!, {
|
||||||
|
parameters: { foo: "bar" },
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(result.data).toEqual([
|
||||||
|
{
|
||||||
|
acknowledged: true,
|
||||||
|
insertedId: expectValidId,
|
||||||
|
},
|
||||||
|
])
|
||||||
|
|
||||||
|
await withCollection(async collection => {
|
||||||
|
const doc = await collection.findOne({ foo: { $eq: "bar" } })
|
||||||
|
expect(doc).toEqual({
|
||||||
|
_id: expectValidBsonObjectId,
|
||||||
|
foo: "bar",
|
||||||
|
})
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
it("a delete query with parameters", async () => {
|
||||||
|
const query = await createQuery({
|
||||||
|
fields: {
|
||||||
|
json: { name: { $eq: "{{ name }}" } },
|
||||||
|
extra: {
|
||||||
|
actionType: "deleteOne",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
queryVerb: "delete",
|
||||||
|
parameters: [
|
||||||
|
{
|
||||||
|
name: "name",
|
||||||
|
default: "",
|
||||||
|
},
|
||||||
|
],
|
||||||
|
})
|
||||||
|
|
||||||
|
const result = await config.api.query.execute(query._id!, {
|
||||||
|
parameters: { name: "one" },
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(result.data).toEqual([
|
||||||
|
{
|
||||||
|
acknowledged: true,
|
||||||
|
deletedCount: 1,
|
||||||
|
},
|
||||||
|
])
|
||||||
|
|
||||||
|
await withCollection(async collection => {
|
||||||
|
const doc = await collection.findOne({ name: { $eq: "one" } })
|
||||||
|
expect(doc).toBeNull()
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
it("an update query with parameters", async () => {
|
||||||
|
const query = await createQuery({
|
||||||
|
fields: {
|
||||||
|
json: {
|
||||||
|
filter: { name: { $eq: "{{ name }}" } },
|
||||||
|
update: { $set: { name: "{{ newName }}" } },
|
||||||
|
},
|
||||||
|
extra: {
|
||||||
|
actionType: "updateOne",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
queryVerb: "update",
|
||||||
|
parameters: [
|
||||||
|
{
|
||||||
|
name: "name",
|
||||||
|
default: "",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "newName",
|
||||||
|
default: "",
|
||||||
|
},
|
||||||
|
],
|
||||||
|
})
|
||||||
|
|
||||||
|
const result = await config.api.query.execute(query._id!, {
|
||||||
|
parameters: { name: "one", newName: "newOne" },
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(result.data).toEqual([
|
||||||
|
{
|
||||||
|
acknowledged: true,
|
||||||
|
matchedCount: 1,
|
||||||
|
modifiedCount: 1,
|
||||||
|
upsertedCount: 0,
|
||||||
|
upsertedId: null,
|
||||||
|
},
|
||||||
|
])
|
||||||
|
|
||||||
|
await withCollection(async collection => {
|
||||||
|
const doc = await collection.findOne({ name: { $eq: "newOne" } })
|
||||||
|
expect(doc).toEqual({
|
||||||
|
_id: expectValidBsonObjectId,
|
||||||
|
name: "newOne",
|
||||||
|
})
|
||||||
|
|
||||||
|
const oldDoc = await collection.findOne({ name: { $eq: "one" } })
|
||||||
|
expect(oldDoc).toBeNull()
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should be able to delete all records", async () => {
|
||||||
|
const query = await createQuery({
|
||||||
|
fields: {
|
||||||
|
json: {},
|
||||||
|
extra: {
|
||||||
|
actionType: "deleteMany",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
queryVerb: "delete",
|
||||||
|
})
|
||||||
|
|
||||||
|
const result = await config.api.query.execute(query._id!)
|
||||||
|
|
||||||
|
expect(result.data).toEqual([
|
||||||
|
{
|
||||||
|
acknowledged: true,
|
||||||
|
deletedCount: 5,
|
||||||
|
},
|
||||||
|
])
|
||||||
|
|
||||||
|
await withCollection(async collection => {
|
||||||
|
const docs = await collection.find().toArray()
|
||||||
|
expect(docs).toHaveLength(0)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should be able to update all documents", async () => {
|
||||||
|
const query = await createQuery({
|
||||||
|
fields: {
|
||||||
|
json: {
|
||||||
|
filter: {},
|
||||||
|
update: { $set: { name: "newName" } },
|
||||||
|
},
|
||||||
|
extra: {
|
||||||
|
actionType: "updateMany",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
queryVerb: "update",
|
||||||
|
})
|
||||||
|
|
||||||
|
const result = await config.api.query.execute(query._id!)
|
||||||
|
|
||||||
|
expect(result.data).toEqual([
|
||||||
|
{
|
||||||
|
acknowledged: true,
|
||||||
|
matchedCount: 5,
|
||||||
|
modifiedCount: 5,
|
||||||
|
upsertedCount: 0,
|
||||||
|
upsertedId: null,
|
||||||
|
},
|
||||||
|
])
|
||||||
|
|
||||||
|
await withCollection(async collection => {
|
||||||
|
const docs = await collection.find().toArray()
|
||||||
|
expect(docs).toHaveLength(5)
|
||||||
|
for (const doc of docs) {
|
||||||
|
expect(doc).toEqual({
|
||||||
|
_id: expectValidBsonObjectId,
|
||||||
|
name: "newName",
|
||||||
|
})
|
||||||
|
}
|
||||||
|
})
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1,47 @@
|
||||||
|
import * as setup from "../utilities"
|
||||||
|
import { checkBuilderEndpoint } from "../utilities/TestFunctions"
|
||||||
|
import TestConfiguration from "../../../../tests/utilities/TestConfiguration"
|
||||||
|
import { Datasource, Query, SourceName } from "@budibase/types"
|
||||||
|
|
||||||
|
describe("query permissions", () => {
|
||||||
|
let config: TestConfiguration
|
||||||
|
let datasource: Datasource
|
||||||
|
let query: Query
|
||||||
|
|
||||||
|
beforeAll(async () => {
|
||||||
|
config = setup.getConfig()
|
||||||
|
await config.init()
|
||||||
|
datasource = await config.api.datasource.create({
|
||||||
|
name: "test datasource",
|
||||||
|
type: "test",
|
||||||
|
source: SourceName.REST,
|
||||||
|
config: {},
|
||||||
|
})
|
||||||
|
query = await config.api.query.save({
|
||||||
|
name: "test query",
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
parameters: [],
|
||||||
|
fields: {},
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
queryVerb: "read",
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
it("delete should require builder", async () => {
|
||||||
|
await checkBuilderEndpoint({
|
||||||
|
config,
|
||||||
|
method: "DELETE",
|
||||||
|
url: `/api/queries/${query._id}/${query._rev}`,
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
it("preview should require builder", async () => {
|
||||||
|
await checkBuilderEndpoint({
|
||||||
|
config,
|
||||||
|
method: "POST",
|
||||||
|
url: `/api/queries/preview`,
|
||||||
|
})
|
||||||
|
})
|
||||||
|
})
|
|
@ -1,774 +0,0 @@
|
||||||
import tk from "timekeeper"
|
|
||||||
|
|
||||||
const pg = require("pg")
|
|
||||||
|
|
||||||
// Mock out postgres for this
|
|
||||||
jest.mock("pg")
|
|
||||||
jest.mock("node-fetch")
|
|
||||||
|
|
||||||
// Mock isProdAppID to we can later mock the implementation and pretend we are
|
|
||||||
// using prod app IDs
|
|
||||||
jest.mock("@budibase/backend-core", () => {
|
|
||||||
const core = jest.requireActual("@budibase/backend-core")
|
|
||||||
return {
|
|
||||||
...core,
|
|
||||||
db: {
|
|
||||||
...core.db,
|
|
||||||
isProdAppID: jest.fn(),
|
|
||||||
},
|
|
||||||
}
|
|
||||||
})
|
|
||||||
import * as setup from "../utilities"
|
|
||||||
import { checkBuilderEndpoint } from "../utilities/TestFunctions"
|
|
||||||
import { checkCacheForDynamicVariable } from "../../../../threads/utils"
|
|
||||||
|
|
||||||
const { basicQuery, basicDatasource } = setup.structures
|
|
||||||
import { events, db as dbCore } from "@budibase/backend-core"
|
|
||||||
import {
|
|
||||||
Datasource,
|
|
||||||
Query,
|
|
||||||
SourceName,
|
|
||||||
QueryPreview,
|
|
||||||
QueryParameter,
|
|
||||||
} from "@budibase/types"
|
|
||||||
|
|
||||||
tk.freeze(Date.now())
|
|
||||||
|
|
||||||
const mockIsProdAppID = dbCore.isProdAppID as jest.MockedFunction<
|
|
||||||
typeof dbCore.isProdAppID
|
|
||||||
>
|
|
||||||
|
|
||||||
describe("/queries", () => {
|
|
||||||
let request = setup.getRequest()
|
|
||||||
let config = setup.getConfig()
|
|
||||||
let datasource: Datasource & Required<Pick<Datasource, "_id">>, query: Query
|
|
||||||
|
|
||||||
afterAll(setup.afterAll)
|
|
||||||
|
|
||||||
const setupTest = async () => {
|
|
||||||
await config.init()
|
|
||||||
datasource = await config.createDatasource()
|
|
||||||
query = await config.createQuery()
|
|
||||||
}
|
|
||||||
|
|
||||||
beforeAll(async () => {
|
|
||||||
await setupTest()
|
|
||||||
})
|
|
||||||
|
|
||||||
const createQuery = async (query: Query) => {
|
|
||||||
return request
|
|
||||||
.post(`/api/queries`)
|
|
||||||
.send(query)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
}
|
|
||||||
|
|
||||||
describe("create", () => {
|
|
||||||
it("should create a new query", async () => {
|
|
||||||
const { _id } = await config.createDatasource()
|
|
||||||
const query = basicQuery(_id)
|
|
||||||
jest.clearAllMocks()
|
|
||||||
const res = await createQuery(query)
|
|
||||||
|
|
||||||
expect((res as any).res.statusMessage).toEqual(
|
|
||||||
`Query ${query.name} saved successfully.`
|
|
||||||
)
|
|
||||||
expect(res.body).toEqual({
|
|
||||||
_rev: res.body._rev,
|
|
||||||
_id: res.body._id,
|
|
||||||
...query,
|
|
||||||
nullDefaultSupport: true,
|
|
||||||
createdAt: new Date().toISOString(),
|
|
||||||
updatedAt: new Date().toISOString(),
|
|
||||||
})
|
|
||||||
expect(events.query.created).toHaveBeenCalledTimes(1)
|
|
||||||
expect(events.query.updated).not.toHaveBeenCalled()
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
describe("update", () => {
|
|
||||||
it("should update query", async () => {
|
|
||||||
const { _id } = await config.createDatasource()
|
|
||||||
const query = basicQuery(_id)
|
|
||||||
const res = await createQuery(query)
|
|
||||||
jest.clearAllMocks()
|
|
||||||
query._id = res.body._id
|
|
||||||
query._rev = res.body._rev
|
|
||||||
await createQuery(query)
|
|
||||||
|
|
||||||
expect((res as any).res.statusMessage).toEqual(
|
|
||||||
`Query ${query.name} saved successfully.`
|
|
||||||
)
|
|
||||||
expect(res.body).toEqual({
|
|
||||||
_rev: res.body._rev,
|
|
||||||
_id: res.body._id,
|
|
||||||
...query,
|
|
||||||
nullDefaultSupport: true,
|
|
||||||
createdAt: new Date().toISOString(),
|
|
||||||
updatedAt: new Date().toISOString(),
|
|
||||||
})
|
|
||||||
expect(events.query.created).not.toHaveBeenCalled()
|
|
||||||
expect(events.query.updated).toHaveBeenCalledTimes(1)
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
describe("fetch", () => {
|
|
||||||
beforeEach(async () => {
|
|
||||||
await setupTest()
|
|
||||||
})
|
|
||||||
|
|
||||||
it("returns all the queries from the server", async () => {
|
|
||||||
const res = await request
|
|
||||||
.get(`/api/queries`)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
const queries = res.body
|
|
||||||
expect(queries).toEqual([
|
|
||||||
{
|
|
||||||
_rev: query._rev,
|
|
||||||
_id: query._id,
|
|
||||||
createdAt: new Date().toISOString(),
|
|
||||||
...basicQuery(datasource._id),
|
|
||||||
nullDefaultSupport: true,
|
|
||||||
updatedAt: new Date().toISOString(),
|
|
||||||
readable: true,
|
|
||||||
},
|
|
||||||
])
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should apply authorization to endpoint", async () => {
|
|
||||||
await checkBuilderEndpoint({
|
|
||||||
config,
|
|
||||||
method: "GET",
|
|
||||||
url: `/api/datasources`,
|
|
||||||
})
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
describe("find", () => {
|
|
||||||
it("should find a query in builder", async () => {
|
|
||||||
const query = await config.createQuery()
|
|
||||||
const res = await request
|
|
||||||
.get(`/api/queries/${query._id}`)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
expect(res.body._id).toEqual(query._id)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should find a query in cloud", async () => {
|
|
||||||
await config.withEnv({ SELF_HOSTED: "true" }, async () => {
|
|
||||||
const query = await config.createQuery()
|
|
||||||
const res = await request
|
|
||||||
.get(`/api/queries/${query._id}`)
|
|
||||||
.set(await config.defaultHeaders())
|
|
||||||
.expect(200)
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
expect(res.body.fields).toBeDefined()
|
|
||||||
expect(res.body.parameters).toBeDefined()
|
|
||||||
expect(res.body.schema).toBeDefined()
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should remove sensitive info for prod apps", async () => {
|
|
||||||
// Mock isProdAppID to pretend we are using a prod app
|
|
||||||
mockIsProdAppID.mockClear()
|
|
||||||
mockIsProdAppID.mockImplementation(() => true)
|
|
||||||
|
|
||||||
const query = await config.createQuery()
|
|
||||||
const res = await request
|
|
||||||
.get(`/api/queries/${query._id}`)
|
|
||||||
.set(await config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
expect(res.body._id).toEqual(query._id)
|
|
||||||
expect(res.body.fields).toBeUndefined()
|
|
||||||
expect(res.body.parameters).toBeUndefined()
|
|
||||||
expect(res.body.schema).toBeDefined()
|
|
||||||
|
|
||||||
// Reset isProdAppID mock
|
|
||||||
expect(dbCore.isProdAppID).toHaveBeenCalledTimes(1)
|
|
||||||
mockIsProdAppID.mockImplementation(() => false)
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
describe("destroy", () => {
|
|
||||||
beforeEach(async () => {
|
|
||||||
await setupTest()
|
|
||||||
})
|
|
||||||
|
|
||||||
it("deletes a query and returns a success message", async () => {
|
|
||||||
await request
|
|
||||||
.delete(`/api/queries/${query._id}/${query._rev}`)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
const res = await request
|
|
||||||
.get(`/api/queries`)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
expect(res.body).toEqual([])
|
|
||||||
expect(events.query.deleted).toHaveBeenCalledTimes(1)
|
|
||||||
expect(events.query.deleted).toHaveBeenCalledWith(datasource, query)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should apply authorization to endpoint", async () => {
|
|
||||||
const query = await config.createQuery()
|
|
||||||
await checkBuilderEndpoint({
|
|
||||||
config,
|
|
||||||
method: "DELETE",
|
|
||||||
url: `/api/queries/${query._id}/${query._rev}`,
|
|
||||||
})
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
describe("preview", () => {
|
|
||||||
it("should be able to preview the query", async () => {
|
|
||||||
const queryPreview: QueryPreview = {
|
|
||||||
datasourceId: datasource._id,
|
|
||||||
queryVerb: "read",
|
|
||||||
fields: {},
|
|
||||||
parameters: [],
|
|
||||||
transformer: "return data",
|
|
||||||
name: datasource.name!,
|
|
||||||
schema: {},
|
|
||||||
readable: true,
|
|
||||||
}
|
|
||||||
const responseBody = await config.api.query.previewQuery(queryPreview)
|
|
||||||
// these responses come from the mock
|
|
||||||
expect(responseBody.schema).toEqual({
|
|
||||||
a: { type: "string", name: "a" },
|
|
||||||
b: { type: "number", name: "b" },
|
|
||||||
})
|
|
||||||
expect(responseBody.rows.length).toEqual(1)
|
|
||||||
expect(events.query.previewed).toHaveBeenCalledTimes(1)
|
|
||||||
delete datasource.config
|
|
||||||
expect(events.query.previewed).toHaveBeenCalledWith(datasource, {
|
|
||||||
...queryPreview,
|
|
||||||
nullDefaultSupport: true,
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should apply authorization to endpoint", async () => {
|
|
||||||
await checkBuilderEndpoint({
|
|
||||||
config,
|
|
||||||
method: "POST",
|
|
||||||
url: `/api/queries/preview`,
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should not error when trying to generate a nested schema for an empty array", async () => {
|
|
||||||
const queryPreview: QueryPreview = {
|
|
||||||
datasourceId: datasource._id,
|
|
||||||
parameters: [],
|
|
||||||
fields: {},
|
|
||||||
queryVerb: "read",
|
|
||||||
name: datasource.name!,
|
|
||||||
transformer: "return data",
|
|
||||||
schema: {},
|
|
||||||
readable: true,
|
|
||||||
}
|
|
||||||
const rows = [
|
|
||||||
{
|
|
||||||
contacts: [],
|
|
||||||
},
|
|
||||||
]
|
|
||||||
pg.queryMock.mockImplementation(() => ({
|
|
||||||
rows,
|
|
||||||
}))
|
|
||||||
|
|
||||||
const responseBody = await config.api.query.previewQuery(queryPreview)
|
|
||||||
expect(responseBody).toEqual({
|
|
||||||
nestedSchemaFields: {},
|
|
||||||
rows,
|
|
||||||
schema: {
|
|
||||||
contacts: { type: "array", name: "contacts" },
|
|
||||||
},
|
|
||||||
})
|
|
||||||
expect(responseBody.rows.length).toEqual(1)
|
|
||||||
delete datasource.config
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should generate a nested schema based on all the nested items", async () => {
|
|
||||||
const queryPreview: QueryPreview = {
|
|
||||||
datasourceId: datasource._id,
|
|
||||||
parameters: [],
|
|
||||||
fields: {},
|
|
||||||
queryVerb: "read",
|
|
||||||
name: datasource.name!,
|
|
||||||
transformer: "return data",
|
|
||||||
schema: {},
|
|
||||||
readable: true,
|
|
||||||
}
|
|
||||||
const rows = [
|
|
||||||
{
|
|
||||||
contacts: [
|
|
||||||
{
|
|
||||||
address: "123 Lane",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
address: "456 Drive",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
postcode: "BT1 12N",
|
|
||||||
lat: 54.59,
|
|
||||||
long: -5.92,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
city: "Belfast",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
address: "789 Avenue",
|
|
||||||
phoneNumber: "0800-999-5555",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "Name",
|
|
||||||
isActive: false,
|
|
||||||
},
|
|
||||||
],
|
|
||||||
},
|
|
||||||
]
|
|
||||||
|
|
||||||
pg.queryMock.mockImplementation(() => ({
|
|
||||||
rows,
|
|
||||||
}))
|
|
||||||
|
|
||||||
const responseBody = await config.api.query.previewQuery(queryPreview)
|
|
||||||
expect(responseBody).toEqual({
|
|
||||||
nestedSchemaFields: {
|
|
||||||
contacts: {
|
|
||||||
address: {
|
|
||||||
type: "string",
|
|
||||||
name: "address",
|
|
||||||
},
|
|
||||||
postcode: {
|
|
||||||
type: "string",
|
|
||||||
name: "postcode",
|
|
||||||
},
|
|
||||||
lat: {
|
|
||||||
type: "number",
|
|
||||||
name: "lat",
|
|
||||||
},
|
|
||||||
long: {
|
|
||||||
type: "number",
|
|
||||||
name: "long",
|
|
||||||
},
|
|
||||||
city: {
|
|
||||||
type: "string",
|
|
||||||
name: "city",
|
|
||||||
},
|
|
||||||
phoneNumber: {
|
|
||||||
type: "string",
|
|
||||||
name: "phoneNumber",
|
|
||||||
},
|
|
||||||
name: {
|
|
||||||
type: "string",
|
|
||||||
name: "name",
|
|
||||||
},
|
|
||||||
isActive: {
|
|
||||||
type: "boolean",
|
|
||||||
name: "isActive",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
rows,
|
|
||||||
schema: {
|
|
||||||
contacts: { type: "json", name: "contacts", subtype: "array" },
|
|
||||||
},
|
|
||||||
})
|
|
||||||
expect(responseBody.rows.length).toEqual(1)
|
|
||||||
delete datasource.config
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
describe("execute", () => {
|
|
||||||
beforeEach(async () => {
|
|
||||||
await setupTest()
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should be able to execute the query", async () => {
|
|
||||||
const res = await request
|
|
||||||
.post(`/api/queries/${query._id}`)
|
|
||||||
.send({
|
|
||||||
parameters: {},
|
|
||||||
})
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
expect(res.body.length).toEqual(1)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should fail with invalid integration type", async () => {
|
|
||||||
const datasource: Datasource = {
|
|
||||||
...basicDatasource().datasource,
|
|
||||||
source: "INVALID_INTEGRATION" as SourceName,
|
|
||||||
}
|
|
||||||
await config.api.datasource.create(datasource, {
|
|
||||||
status: 500,
|
|
||||||
body: {
|
|
||||||
message: "No datasource implementation found.",
|
|
||||||
},
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
it("shouldn't allow handlebars to be passed as parameters", async () => {
|
|
||||||
const res = await request
|
|
||||||
.post(`/api/queries/${query._id}`)
|
|
||||||
.send({
|
|
||||||
parameters: {
|
|
||||||
a: "{{ 'test' }}",
|
|
||||||
},
|
|
||||||
})
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect(400)
|
|
||||||
expect(res.body.message).toEqual(
|
|
||||||
"Parameter 'a' input contains a handlebars binding - this is not allowed."
|
|
||||||
)
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
describe("variables", () => {
|
|
||||||
async function preview(datasource: Datasource, fields: any) {
|
|
||||||
const queryPreview: QueryPreview = {
|
|
||||||
datasourceId: datasource._id!,
|
|
||||||
parameters: [],
|
|
||||||
fields,
|
|
||||||
queryVerb: "read",
|
|
||||||
name: datasource.name!,
|
|
||||||
transformer: "return data",
|
|
||||||
schema: {},
|
|
||||||
readable: true,
|
|
||||||
}
|
|
||||||
return await config.api.query.previewQuery(queryPreview)
|
|
||||||
}
|
|
||||||
|
|
||||||
it("should work with static variables", async () => {
|
|
||||||
const datasource = await config.restDatasource({
|
|
||||||
staticVariables: {
|
|
||||||
variable: "google",
|
|
||||||
variable2: "1",
|
|
||||||
},
|
|
||||||
})
|
|
||||||
const responseBody = await preview(datasource, {
|
|
||||||
path: "www.{{ variable }}.com",
|
|
||||||
queryString: "test={{ variable2 }}",
|
|
||||||
})
|
|
||||||
// these responses come from the mock
|
|
||||||
expect(responseBody.schema).toEqual({
|
|
||||||
opts: { type: "json", name: "opts" },
|
|
||||||
url: { type: "string", name: "url" },
|
|
||||||
value: { type: "string", name: "value" },
|
|
||||||
})
|
|
||||||
expect(responseBody.rows[0].url).toEqual("http://www.google.com?test=1")
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should work with dynamic variables", async () => {
|
|
||||||
const { datasource } = await config.dynamicVariableDatasource()
|
|
||||||
const responseBody = await preview(datasource, {
|
|
||||||
path: "www.google.com",
|
|
||||||
queryString: "test={{ variable3 }}",
|
|
||||||
})
|
|
||||||
expect(responseBody.schema).toEqual({
|
|
||||||
opts: { type: "json", name: "opts" },
|
|
||||||
url: { type: "string", name: "url" },
|
|
||||||
value: { type: "string", name: "value" },
|
|
||||||
})
|
|
||||||
expect(responseBody.rows[0].url).toContain("doctype%20html")
|
|
||||||
})
|
|
||||||
|
|
||||||
it("check that it automatically retries on fail with cached dynamics", async () => {
|
|
||||||
const { datasource, query: base } =
|
|
||||||
await config.dynamicVariableDatasource()
|
|
||||||
// preview once to cache
|
|
||||||
await preview(datasource, {
|
|
||||||
path: "www.google.com",
|
|
||||||
queryString: "test={{ variable3 }}",
|
|
||||||
})
|
|
||||||
// check its in cache
|
|
||||||
const contents = await checkCacheForDynamicVariable(
|
|
||||||
base._id!,
|
|
||||||
"variable3"
|
|
||||||
)
|
|
||||||
expect(contents.rows.length).toEqual(1)
|
|
||||||
const responseBody = await preview(datasource, {
|
|
||||||
path: "www.failonce.com",
|
|
||||||
queryString: "test={{ variable3 }}",
|
|
||||||
})
|
|
||||||
expect(responseBody.schema).toEqual({
|
|
||||||
fails: { type: "number", name: "fails" },
|
|
||||||
opts: { type: "json", name: "opts" },
|
|
||||||
url: { type: "string", name: "url" },
|
|
||||||
})
|
|
||||||
expect(responseBody.rows[0].fails).toEqual(1)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("deletes variables when linked query is deleted", async () => {
|
|
||||||
const { datasource, query: base } =
|
|
||||||
await config.dynamicVariableDatasource()
|
|
||||||
// preview once to cache
|
|
||||||
await preview(datasource, {
|
|
||||||
path: "www.google.com",
|
|
||||||
queryString: "test={{ variable3 }}",
|
|
||||||
})
|
|
||||||
// check its in cache
|
|
||||||
let contents = await checkCacheForDynamicVariable(base._id!, "variable3")
|
|
||||||
expect(contents.rows.length).toEqual(1)
|
|
||||||
|
|
||||||
// delete the query
|
|
||||||
await request
|
|
||||||
.delete(`/api/queries/${base._id}/${base._rev}`)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
// check variables no longer in cache
|
|
||||||
contents = await checkCacheForDynamicVariable(base._id!, "variable3")
|
|
||||||
expect(contents).toBe(null)
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
describe("Current User Request Mapping", () => {
|
|
||||||
async function previewGet(
|
|
||||||
datasource: Datasource,
|
|
||||||
fields: any,
|
|
||||||
params: QueryParameter[]
|
|
||||||
) {
|
|
||||||
const queryPreview: QueryPreview = {
|
|
||||||
datasourceId: datasource._id!,
|
|
||||||
parameters: params,
|
|
||||||
fields,
|
|
||||||
queryVerb: "read",
|
|
||||||
name: datasource.name!,
|
|
||||||
transformer: "return data",
|
|
||||||
schema: {},
|
|
||||||
readable: true,
|
|
||||||
}
|
|
||||||
return await config.api.query.previewQuery(queryPreview)
|
|
||||||
}
|
|
||||||
|
|
||||||
async function previewPost(
|
|
||||||
datasource: Datasource,
|
|
||||||
fields: any,
|
|
||||||
params: QueryParameter[]
|
|
||||||
) {
|
|
||||||
const queryPreview: QueryPreview = {
|
|
||||||
datasourceId: datasource._id!,
|
|
||||||
parameters: params,
|
|
||||||
fields,
|
|
||||||
queryVerb: "create",
|
|
||||||
name: datasource.name!,
|
|
||||||
transformer: null,
|
|
||||||
schema: {},
|
|
||||||
readable: false,
|
|
||||||
}
|
|
||||||
return await config.api.query.previewQuery(queryPreview)
|
|
||||||
}
|
|
||||||
|
|
||||||
it("should parse global and query level header mappings", async () => {
|
|
||||||
const userDetails = config.getUserDetails()
|
|
||||||
|
|
||||||
const datasource = await config.restDatasource({
|
|
||||||
defaultHeaders: {
|
|
||||||
test: "headerVal",
|
|
||||||
emailHdr: "{{[user].[email]}}",
|
|
||||||
},
|
|
||||||
})
|
|
||||||
const responseBody = await previewGet(
|
|
||||||
datasource,
|
|
||||||
{
|
|
||||||
path: "www.google.com",
|
|
||||||
queryString: "email={{[user].[email]}}",
|
|
||||||
headers: {
|
|
||||||
queryHdr: "{{[user].[firstName]}}",
|
|
||||||
secondHdr: "1234",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
[]
|
|
||||||
)
|
|
||||||
|
|
||||||
const parsedRequest = JSON.parse(responseBody.extra.raw)
|
|
||||||
expect(parsedRequest.opts.headers).toEqual({
|
|
||||||
test: "headerVal",
|
|
||||||
emailHdr: userDetails.email,
|
|
||||||
queryHdr: userDetails.firstName,
|
|
||||||
secondHdr: "1234",
|
|
||||||
})
|
|
||||||
expect(responseBody.rows[0].url).toEqual(
|
|
||||||
"http://www.google.com?email=" + userDetails.email.replace("@", "%40")
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should bind the current user to query parameters", async () => {
|
|
||||||
const userDetails = config.getUserDetails()
|
|
||||||
|
|
||||||
const datasource = await config.restDatasource()
|
|
||||||
|
|
||||||
const responseBody = await previewGet(
|
|
||||||
datasource,
|
|
||||||
{
|
|
||||||
path: "www.google.com",
|
|
||||||
queryString:
|
|
||||||
"test={{myEmail}}&testName={{myName}}&testParam={{testParam}}",
|
|
||||||
},
|
|
||||||
[
|
|
||||||
{ name: "myEmail", default: "{{[user].[email]}}" },
|
|
||||||
{ name: "myName", default: "{{[user].[firstName]}}" },
|
|
||||||
{ name: "testParam", default: "1234" },
|
|
||||||
]
|
|
||||||
)
|
|
||||||
|
|
||||||
expect(responseBody.rows[0].url).toEqual(
|
|
||||||
"http://www.google.com?test=" +
|
|
||||||
userDetails.email.replace("@", "%40") +
|
|
||||||
"&testName=" +
|
|
||||||
userDetails.firstName +
|
|
||||||
"&testParam=1234"
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should bind the current user the request body - plain text", async () => {
|
|
||||||
const userDetails = config.getUserDetails()
|
|
||||||
const datasource = await config.restDatasource()
|
|
||||||
|
|
||||||
const responseBody = await previewPost(
|
|
||||||
datasource,
|
|
||||||
{
|
|
||||||
path: "www.google.com",
|
|
||||||
queryString: "testParam={{testParam}}",
|
|
||||||
requestBody:
|
|
||||||
"This is plain text and this is my email: {{[user].[email]}}. This is a test param: {{testParam}}",
|
|
||||||
bodyType: "text",
|
|
||||||
},
|
|
||||||
[{ name: "testParam", default: "1234" }]
|
|
||||||
)
|
|
||||||
|
|
||||||
const parsedRequest = JSON.parse(responseBody.extra.raw)
|
|
||||||
expect(parsedRequest.opts.body).toEqual(
|
|
||||||
`This is plain text and this is my email: ${userDetails.email}. This is a test param: 1234`
|
|
||||||
)
|
|
||||||
expect(responseBody.rows[0].url).toEqual(
|
|
||||||
"http://www.google.com?testParam=1234"
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should bind the current user the request body - json", async () => {
|
|
||||||
const userDetails = config.getUserDetails()
|
|
||||||
const datasource = await config.restDatasource()
|
|
||||||
|
|
||||||
const responseBody = await previewPost(
|
|
||||||
datasource,
|
|
||||||
{
|
|
||||||
path: "www.google.com",
|
|
||||||
queryString: "testParam={{testParam}}",
|
|
||||||
requestBody:
|
|
||||||
'{"email":"{{[user].[email]}}","queryCode":{{testParam}},"userRef":"{{userRef}}"}',
|
|
||||||
bodyType: "json",
|
|
||||||
},
|
|
||||||
[
|
|
||||||
{ name: "testParam", default: "1234" },
|
|
||||||
{ name: "userRef", default: "{{[user].[firstName]}}" },
|
|
||||||
]
|
|
||||||
)
|
|
||||||
|
|
||||||
const parsedRequest = JSON.parse(responseBody.extra.raw)
|
|
||||||
const test = `{"email":"${userDetails.email}","queryCode":1234,"userRef":"${userDetails.firstName}"}`
|
|
||||||
expect(parsedRequest.opts.body).toEqual(test)
|
|
||||||
expect(responseBody.rows[0].url).toEqual(
|
|
||||||
"http://www.google.com?testParam=1234"
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should bind the current user the request body - xml", async () => {
|
|
||||||
const userDetails = config.getUserDetails()
|
|
||||||
const datasource = await config.restDatasource()
|
|
||||||
|
|
||||||
const responseBody = await previewPost(
|
|
||||||
datasource,
|
|
||||||
{
|
|
||||||
path: "www.google.com",
|
|
||||||
queryString: "testParam={{testParam}}",
|
|
||||||
requestBody:
|
|
||||||
"<note> <email>{{[user].[email]}}</email> <code>{{testParam}}</code> " +
|
|
||||||
"<ref>{{userId}}</ref> <somestring>testing</somestring> </note>",
|
|
||||||
bodyType: "xml",
|
|
||||||
},
|
|
||||||
[
|
|
||||||
{ name: "testParam", default: "1234" },
|
|
||||||
{ name: "userId", default: "{{[user].[firstName]}}" },
|
|
||||||
]
|
|
||||||
)
|
|
||||||
|
|
||||||
const parsedRequest = JSON.parse(responseBody.extra.raw)
|
|
||||||
const test = `<note> <email>${userDetails.email}</email> <code>1234</code> <ref>${userDetails.firstName}</ref> <somestring>testing</somestring> </note>`
|
|
||||||
|
|
||||||
expect(parsedRequest.opts.body).toEqual(test)
|
|
||||||
expect(responseBody.rows[0].url).toEqual(
|
|
||||||
"http://www.google.com?testParam=1234"
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should bind the current user the request body - form-data", async () => {
|
|
||||||
const userDetails = config.getUserDetails()
|
|
||||||
const datasource = await config.restDatasource()
|
|
||||||
|
|
||||||
const responseBody = await previewPost(
|
|
||||||
datasource,
|
|
||||||
{
|
|
||||||
path: "www.google.com",
|
|
||||||
queryString: "testParam={{testParam}}",
|
|
||||||
requestBody:
|
|
||||||
'{"email":"{{[user].[email]}}","queryCode":{{testParam}},"userRef":"{{userRef}}"}',
|
|
||||||
bodyType: "form",
|
|
||||||
},
|
|
||||||
[
|
|
||||||
{ name: "testParam", default: "1234" },
|
|
||||||
{ name: "userRef", default: "{{[user].[firstName]}}" },
|
|
||||||
]
|
|
||||||
)
|
|
||||||
|
|
||||||
const parsedRequest = JSON.parse(responseBody.extra.raw)
|
|
||||||
|
|
||||||
const emailData = parsedRequest.opts.body._streams[1]
|
|
||||||
expect(emailData).toEqual(userDetails.email)
|
|
||||||
|
|
||||||
const queryCodeData = parsedRequest.opts.body._streams[4]
|
|
||||||
expect(queryCodeData).toEqual("1234")
|
|
||||||
|
|
||||||
const userRef = parsedRequest.opts.body._streams[7]
|
|
||||||
expect(userRef).toEqual(userDetails.firstName)
|
|
||||||
|
|
||||||
expect(responseBody.rows[0].url).toEqual(
|
|
||||||
"http://www.google.com?testParam=1234"
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should bind the current user the request body - encoded", async () => {
|
|
||||||
const userDetails = config.getUserDetails()
|
|
||||||
const datasource = await config.restDatasource()
|
|
||||||
|
|
||||||
const responseBody = await previewPost(
|
|
||||||
datasource,
|
|
||||||
{
|
|
||||||
path: "www.google.com",
|
|
||||||
queryString: "testParam={{testParam}}",
|
|
||||||
requestBody:
|
|
||||||
'{"email":"{{[user].[email]}}","queryCode":{{testParam}},"userRef":"{{userRef}}"}',
|
|
||||||
bodyType: "encoded",
|
|
||||||
},
|
|
||||||
[
|
|
||||||
{ name: "testParam", default: "1234" },
|
|
||||||
{ name: "userRef", default: "{{[user].[firstName]}}" },
|
|
||||||
]
|
|
||||||
)
|
|
||||||
const parsedRequest = JSON.parse(responseBody.extra.raw)
|
|
||||||
|
|
||||||
expect(parsedRequest.opts.body.email).toEqual(userDetails.email)
|
|
||||||
expect(parsedRequest.opts.body.queryCode).toEqual("1234")
|
|
||||||
expect(parsedRequest.opts.body.userRef).toEqual(userDetails.firstName)
|
|
||||||
})
|
|
||||||
})
|
|
||||||
})
|
|
|
@ -0,0 +1,406 @@
|
||||||
|
import * as setup from "../utilities"
|
||||||
|
import TestConfiguration from "../../../../tests/utilities/TestConfiguration"
|
||||||
|
import { Datasource, SourceName } from "@budibase/types"
|
||||||
|
import { getCachedVariable } from "../../../../threads/utils"
|
||||||
|
import nock from "nock"
|
||||||
|
import { generator } from "@budibase/backend-core/tests"
|
||||||
|
|
||||||
|
jest.unmock("node-fetch")
|
||||||
|
|
||||||
|
describe("rest", () => {
|
||||||
|
let config: TestConfiguration
|
||||||
|
let datasource: Datasource
|
||||||
|
|
||||||
|
async function createQuery(fields: any) {
|
||||||
|
return await config.api.query.save({
|
||||||
|
name: "test query",
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
parameters: [],
|
||||||
|
fields,
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
queryVerb: "read",
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
beforeAll(async () => {
|
||||||
|
config = setup.getConfig()
|
||||||
|
await config.init()
|
||||||
|
datasource = await config.api.datasource.create({
|
||||||
|
name: generator.guid(),
|
||||||
|
type: "test",
|
||||||
|
source: SourceName.REST,
|
||||||
|
config: {},
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
afterEach(() => {
|
||||||
|
nock.cleanAll()
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should automatically retry on fail with cached dynamics", async () => {
|
||||||
|
const basedOnQuery = await createQuery({
|
||||||
|
path: "one.example.com",
|
||||||
|
})
|
||||||
|
|
||||||
|
let cached = await getCachedVariable(basedOnQuery._id!, "foo")
|
||||||
|
expect(cached).toBeNull()
|
||||||
|
|
||||||
|
await config.api.datasource.update({
|
||||||
|
...datasource,
|
||||||
|
config: {
|
||||||
|
...datasource.config,
|
||||||
|
dynamicVariables: [
|
||||||
|
{
|
||||||
|
queryId: basedOnQuery._id!,
|
||||||
|
name: "foo",
|
||||||
|
value: "{{ data[0].name }}",
|
||||||
|
},
|
||||||
|
],
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
cached = await getCachedVariable(basedOnQuery._id!, "foo")
|
||||||
|
expect(cached).toBeNull()
|
||||||
|
|
||||||
|
nock("http://one.example.com")
|
||||||
|
.get("/")
|
||||||
|
.reply(200, [{ name: "one" }])
|
||||||
|
nock("http://two.example.com").get("/?test=one").reply(500)
|
||||||
|
nock("http://two.example.com")
|
||||||
|
.get("/?test=one")
|
||||||
|
.reply(200, [{ name: "two" }])
|
||||||
|
|
||||||
|
const res = await config.api.query.preview({
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
name: "test query",
|
||||||
|
parameters: [],
|
||||||
|
queryVerb: "read",
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
fields: {
|
||||||
|
path: "two.example.com",
|
||||||
|
queryString: "test={{ foo }}",
|
||||||
|
},
|
||||||
|
})
|
||||||
|
expect(res.schema).toEqual({
|
||||||
|
name: { type: "string", name: "name" },
|
||||||
|
})
|
||||||
|
|
||||||
|
cached = await getCachedVariable(basedOnQuery._id!, "foo")
|
||||||
|
expect(cached.rows.length).toEqual(1)
|
||||||
|
expect(cached.rows[0].name).toEqual("one")
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should parse global and query level header mappings", async () => {
|
||||||
|
const datasource = await config.api.datasource.create({
|
||||||
|
name: generator.guid(),
|
||||||
|
type: "test",
|
||||||
|
source: SourceName.REST,
|
||||||
|
config: {
|
||||||
|
defaultHeaders: {
|
||||||
|
test: "headerVal",
|
||||||
|
emailHdr: "{{[user].[email]}}",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const user = config.getUserDetails()
|
||||||
|
const mock = nock("http://www.example.com", {
|
||||||
|
reqheaders: {
|
||||||
|
test: "headerVal",
|
||||||
|
emailhdr: user.email,
|
||||||
|
queryhdr: user.firstName!,
|
||||||
|
secondhdr: "1234",
|
||||||
|
},
|
||||||
|
})
|
||||||
|
.get("/?email=" + user.email.replace("@", "%40"))
|
||||||
|
.reply(200, {})
|
||||||
|
|
||||||
|
await config.api.query.preview({
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
name: generator.guid(),
|
||||||
|
parameters: [],
|
||||||
|
queryVerb: "read",
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
fields: {
|
||||||
|
path: "www.example.com",
|
||||||
|
queryString: "email={{[user].[email]}}",
|
||||||
|
headers: {
|
||||||
|
queryHdr: "{{[user].[firstName]}}",
|
||||||
|
secondHdr: "1234",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(mock.isDone()).toEqual(true)
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should bind the current user to query params", async () => {
|
||||||
|
const user = config.getUserDetails()
|
||||||
|
const mock = nock("http://www.example.com")
|
||||||
|
.get(
|
||||||
|
"/?test=" +
|
||||||
|
user.email.replace("@", "%40") +
|
||||||
|
"&testName=" +
|
||||||
|
user.firstName +
|
||||||
|
"&testParam=1234"
|
||||||
|
)
|
||||||
|
.reply(200, {})
|
||||||
|
|
||||||
|
await config.api.query.preview({
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
name: generator.guid(),
|
||||||
|
parameters: [
|
||||||
|
{ name: "myEmail", default: "{{[user].[email]}}" },
|
||||||
|
{ name: "myName", default: "{{[user].[firstName]}}" },
|
||||||
|
{ name: "testParam", default: "1234" },
|
||||||
|
],
|
||||||
|
queryVerb: "read",
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
fields: {
|
||||||
|
path: "www.example.com",
|
||||||
|
queryString:
|
||||||
|
"test={{myEmail}}&testName={{myName}}&testParam={{testParam}}",
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(mock.isDone()).toEqual(true)
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should bind the current user to the request body - plain text", async () => {
|
||||||
|
const datasource = await config.api.datasource.create({
|
||||||
|
name: generator.guid(),
|
||||||
|
type: "test",
|
||||||
|
source: SourceName.REST,
|
||||||
|
config: {
|
||||||
|
method: "POST",
|
||||||
|
defaultHeaders: {
|
||||||
|
test: "headerVal",
|
||||||
|
emailHdr: "{{[user].[email]}}",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const user = config.getUserDetails()
|
||||||
|
const mock = nock("http://www.example.com")
|
||||||
|
.post(
|
||||||
|
"/?testParam=1234",
|
||||||
|
"This is plain text and this is my email: " +
|
||||||
|
user.email +
|
||||||
|
". This is a test param: 1234"
|
||||||
|
)
|
||||||
|
.reply(200, {})
|
||||||
|
|
||||||
|
await config.api.query.preview({
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
name: generator.guid(),
|
||||||
|
parameters: [{ name: "testParam", default: "1234" }],
|
||||||
|
queryVerb: "create",
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
fields: {
|
||||||
|
path: "www.example.com",
|
||||||
|
bodyType: "text",
|
||||||
|
queryString: "&testParam={{testParam}}",
|
||||||
|
requestBody:
|
||||||
|
"This is plain text and this is my email: {{[user].[email]}}. This is a test param: {{testParam}}",
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(mock.isDone()).toEqual(true)
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should bind the current user to the request body - json", async () => {
|
||||||
|
const datasource = await config.api.datasource.create({
|
||||||
|
name: generator.guid(),
|
||||||
|
type: "test",
|
||||||
|
source: SourceName.REST,
|
||||||
|
config: {
|
||||||
|
method: "POST",
|
||||||
|
defaultHeaders: {
|
||||||
|
test: "headerVal",
|
||||||
|
emailHdr: "{{[user].[email]}}",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const user = config.getUserDetails()
|
||||||
|
const mock = nock("http://www.example.com")
|
||||||
|
.post("/?testParam=1234", {
|
||||||
|
email: user.email,
|
||||||
|
queryCode: 1234,
|
||||||
|
userRef: user.firstName,
|
||||||
|
})
|
||||||
|
.reply(200, {})
|
||||||
|
|
||||||
|
await config.api.query.preview({
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
name: generator.guid(),
|
||||||
|
parameters: [
|
||||||
|
{ name: "testParam", default: "1234" },
|
||||||
|
{ name: "userRef", default: "{{[user].[firstName]}}" },
|
||||||
|
],
|
||||||
|
queryVerb: "create",
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
fields: {
|
||||||
|
path: "www.example.com",
|
||||||
|
bodyType: "json",
|
||||||
|
queryString: "&testParam={{testParam}}",
|
||||||
|
requestBody:
|
||||||
|
'{"email":"{{[user].[email]}}","queryCode":{{testParam}},"userRef":"{{userRef}}"}',
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(mock.isDone()).toEqual(true)
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should bind the current user to the request body - xml", async () => {
|
||||||
|
const datasource = await config.api.datasource.create({
|
||||||
|
name: generator.guid(),
|
||||||
|
type: "test",
|
||||||
|
source: SourceName.REST,
|
||||||
|
config: {
|
||||||
|
method: "POST",
|
||||||
|
defaultHeaders: {
|
||||||
|
test: "headerVal",
|
||||||
|
emailHdr: "{{[user].[email]}}",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const user = config.getUserDetails()
|
||||||
|
const mock = nock("http://www.example.com")
|
||||||
|
.post(
|
||||||
|
"/?testParam=1234",
|
||||||
|
`<note> <email>${user.email}</email> <code>1234</code> <ref>${user.firstName}</ref> <somestring>testing</somestring> </note>`
|
||||||
|
)
|
||||||
|
.reply(200, {})
|
||||||
|
|
||||||
|
await config.api.query.preview({
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
name: generator.guid(),
|
||||||
|
parameters: [
|
||||||
|
{ name: "testParam", default: "1234" },
|
||||||
|
{ name: "userId", default: "{{[user].[firstName]}}" },
|
||||||
|
],
|
||||||
|
queryVerb: "create",
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
fields: {
|
||||||
|
path: "www.example.com",
|
||||||
|
bodyType: "xml",
|
||||||
|
queryString: "&testParam={{testParam}}",
|
||||||
|
requestBody:
|
||||||
|
"<note> <email>{{[user].[email]}}</email> <code>{{testParam}}</code> " +
|
||||||
|
"<ref>{{userId}}</ref> <somestring>testing</somestring> </note>",
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(mock.isDone()).toEqual(true)
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should bind the current user to the request body - form-data", async () => {
|
||||||
|
const datasource = await config.api.datasource.create({
|
||||||
|
name: generator.guid(),
|
||||||
|
type: "test",
|
||||||
|
source: SourceName.REST,
|
||||||
|
config: {
|
||||||
|
method: "POST",
|
||||||
|
defaultHeaders: {
|
||||||
|
test: "headerVal",
|
||||||
|
emailHdr: "{{[user].[email]}}",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const user = config.getUserDetails()
|
||||||
|
const mock = nock("http://www.example.com")
|
||||||
|
.post("/?testParam=1234", body => {
|
||||||
|
return (
|
||||||
|
body.includes('name="email"\r\n\r\n' + user.email + "\r\n") &&
|
||||||
|
body.includes('name="queryCode"\r\n\r\n1234\r\n') &&
|
||||||
|
body.includes('name="userRef"\r\n\r\n' + user.firstName + "\r\n")
|
||||||
|
)
|
||||||
|
})
|
||||||
|
.reply(200, {})
|
||||||
|
|
||||||
|
await config.api.query.preview({
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
name: generator.guid(),
|
||||||
|
parameters: [
|
||||||
|
{ name: "testParam", default: "1234" },
|
||||||
|
{ name: "userRef", default: "{{[user].[firstName]}}" },
|
||||||
|
],
|
||||||
|
queryVerb: "create",
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
fields: {
|
||||||
|
path: "www.example.com",
|
||||||
|
bodyType: "form",
|
||||||
|
queryString: "&testParam={{testParam}}",
|
||||||
|
requestBody:
|
||||||
|
'{"email":"{{[user].[email]}}","queryCode":{{testParam}},"userRef":"{{userRef}}"}',
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(mock.isDone()).toEqual(true)
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should bind the current user to the request body - encoded", async () => {
|
||||||
|
const datasource = await config.api.datasource.create({
|
||||||
|
name: generator.guid(),
|
||||||
|
type: "test",
|
||||||
|
source: SourceName.REST,
|
||||||
|
config: {
|
||||||
|
method: "POST",
|
||||||
|
defaultHeaders: {
|
||||||
|
test: "headerVal",
|
||||||
|
emailHdr: "{{[user].[email]}}",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
const user = config.getUserDetails()
|
||||||
|
const mock = nock("http://www.example.com")
|
||||||
|
.post("/?testParam=1234", {
|
||||||
|
email: user.email,
|
||||||
|
queryCode: 1234,
|
||||||
|
userRef: user.firstName,
|
||||||
|
})
|
||||||
|
.reply(200, {})
|
||||||
|
|
||||||
|
await config.api.query.preview({
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
name: generator.guid(),
|
||||||
|
parameters: [
|
||||||
|
{ name: "testParam", default: "1234" },
|
||||||
|
{ name: "userRef", default: "{{[user].[firstName]}}" },
|
||||||
|
],
|
||||||
|
queryVerb: "create",
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
fields: {
|
||||||
|
path: "www.example.com",
|
||||||
|
bodyType: "encoded",
|
||||||
|
queryString: "&testParam={{testParam}}",
|
||||||
|
requestBody:
|
||||||
|
'{"email":"{{[user].[email]}}","queryCode":{{testParam}},"userRef":"{{userRef}}"}',
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expect(mock.isDone()).toEqual(true)
|
||||||
|
})
|
||||||
|
})
|
|
@ -30,7 +30,6 @@ const timestamp = new Date("2023-01-26T11:48:57.597Z").toISOString()
|
||||||
tk.freeze(timestamp)
|
tk.freeze(timestamp)
|
||||||
|
|
||||||
jest.unmock("mssql")
|
jest.unmock("mssql")
|
||||||
jest.unmock("pg")
|
|
||||||
|
|
||||||
describe.each([
|
describe.each([
|
||||||
["internal", undefined],
|
["internal", undefined],
|
||||||
|
@ -723,6 +722,39 @@ describe.each([
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
|
describe("bulkImport", () => {
|
||||||
|
isInternal &&
|
||||||
|
it("should update Auto ID field after bulk import", async () => {
|
||||||
|
const table = await config.api.table.save(
|
||||||
|
saveTableRequest({
|
||||||
|
primary: ["autoId"],
|
||||||
|
schema: {
|
||||||
|
autoId: {
|
||||||
|
name: "autoId",
|
||||||
|
type: FieldType.NUMBER,
|
||||||
|
subtype: AutoFieldSubType.AUTO_ID,
|
||||||
|
autocolumn: true,
|
||||||
|
constraints: {
|
||||||
|
type: "number",
|
||||||
|
presence: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
)
|
||||||
|
|
||||||
|
let row = await config.api.row.save(table._id!, {})
|
||||||
|
expect(row.autoId).toEqual(1)
|
||||||
|
|
||||||
|
await config.api.row.bulkImport(table._id!, {
|
||||||
|
rows: [{ autoId: 2 }],
|
||||||
|
})
|
||||||
|
|
||||||
|
row = await config.api.row.save(table._id!, {})
|
||||||
|
expect(row.autoId).toEqual(3)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
describe("enrich", () => {
|
describe("enrich", () => {
|
||||||
beforeAll(async () => {
|
beforeAll(async () => {
|
||||||
table = await config.api.table.save(defaultTable())
|
table = await config.api.table.save(defaultTable())
|
||||||
|
@ -1296,7 +1328,7 @@ describe.each([
|
||||||
|
|
||||||
describe("Formula JS protection", () => {
|
describe("Formula JS protection", () => {
|
||||||
it("should time out JS execution if a single cell takes too long", async () => {
|
it("should time out JS execution if a single cell takes too long", async () => {
|
||||||
await config.withEnv({ JS_PER_INVOCATION_TIMEOUT_MS: 20 }, async () => {
|
await config.withEnv({ JS_PER_INVOCATION_TIMEOUT_MS: 40 }, async () => {
|
||||||
const js = Buffer.from(
|
const js = Buffer.from(
|
||||||
`
|
`
|
||||||
let i = 0;
|
let i = 0;
|
||||||
|
@ -1336,8 +1368,8 @@ describe.each([
|
||||||
it("should time out JS execution if a multiple cells take too long", async () => {
|
it("should time out JS execution if a multiple cells take too long", async () => {
|
||||||
await config.withEnv(
|
await config.withEnv(
|
||||||
{
|
{
|
||||||
JS_PER_INVOCATION_TIMEOUT_MS: 20,
|
JS_PER_INVOCATION_TIMEOUT_MS: 40,
|
||||||
JS_PER_REQUEST_TIMEOUT_MS: 40,
|
JS_PER_REQUEST_TIMEOUT_MS: 80,
|
||||||
},
|
},
|
||||||
async () => {
|
async () => {
|
||||||
const js = Buffer.from(
|
const js = Buffer.from(
|
||||||
|
|
|
@ -1,11 +1,11 @@
|
||||||
import { context, events } from "@budibase/backend-core"
|
import { context, events } from "@budibase/backend-core"
|
||||||
import {
|
import {
|
||||||
AutoFieldSubType,
|
AutoFieldSubType,
|
||||||
|
Datasource,
|
||||||
FieldSubtype,
|
FieldSubtype,
|
||||||
FieldType,
|
FieldType,
|
||||||
INTERNAL_TABLE_SOURCE_ID,
|
INTERNAL_TABLE_SOURCE_ID,
|
||||||
InternalTable,
|
InternalTable,
|
||||||
NumberFieldMetadata,
|
|
||||||
RelationshipType,
|
RelationshipType,
|
||||||
Row,
|
Row,
|
||||||
SaveTableRequest,
|
SaveTableRequest,
|
||||||
|
@ -13,31 +13,41 @@ import {
|
||||||
TableSourceType,
|
TableSourceType,
|
||||||
User,
|
User,
|
||||||
ViewCalculation,
|
ViewCalculation,
|
||||||
|
ViewV2Enriched,
|
||||||
} from "@budibase/types"
|
} from "@budibase/types"
|
||||||
import { checkBuilderEndpoint } from "./utilities/TestFunctions"
|
import { checkBuilderEndpoint } from "./utilities/TestFunctions"
|
||||||
import * as setup from "./utilities"
|
import * as setup from "./utilities"
|
||||||
import sdk from "../../../sdk"
|
|
||||||
import * as uuid from "uuid"
|
import * as uuid from "uuid"
|
||||||
|
|
||||||
import tk from "timekeeper"
|
import { generator } from "@budibase/backend-core/tests"
|
||||||
import { generator, mocks } from "@budibase/backend-core/tests"
|
import { DatabaseName, getDatasource } from "../../../integrations/tests/utils"
|
||||||
import { TableToBuild } from "../../../tests/utilities/TestConfiguration"
|
import { tableForDatasource } from "../../../tests/utilities/structures"
|
||||||
|
import timekeeper from "timekeeper"
|
||||||
tk.freeze(mocks.date.MOCK_DATE)
|
|
||||||
|
|
||||||
const { basicTable } = setup.structures
|
const { basicTable } = setup.structures
|
||||||
const ISO_REGEX_PATTERN = /^\d{4}-\d{2}-\d{2}T\d{2}:\d{2}:\d{2}\.\d{3}Z$/
|
const ISO_REGEX_PATTERN = /^\d{4}-\d{2}-\d{2}T\d{2}:\d{2}:\d{2}\.\d{3}Z$/
|
||||||
|
|
||||||
describe("/tables", () => {
|
describe.each([
|
||||||
let request = setup.getRequest()
|
["internal", undefined],
|
||||||
|
[DatabaseName.POSTGRES, getDatasource(DatabaseName.POSTGRES)],
|
||||||
|
[DatabaseName.MYSQL, getDatasource(DatabaseName.MYSQL)],
|
||||||
|
[DatabaseName.SQL_SERVER, getDatasource(DatabaseName.SQL_SERVER)],
|
||||||
|
[DatabaseName.MARIADB, getDatasource(DatabaseName.MARIADB)],
|
||||||
|
])("/tables (%s)", (_, dsProvider) => {
|
||||||
|
let isInternal: boolean
|
||||||
|
let datasource: Datasource | undefined
|
||||||
let config = setup.getConfig()
|
let config = setup.getConfig()
|
||||||
let appId: string
|
|
||||||
|
|
||||||
afterAll(setup.afterAll)
|
afterAll(setup.afterAll)
|
||||||
|
|
||||||
beforeAll(async () => {
|
beforeAll(async () => {
|
||||||
const app = await config.init()
|
await config.init()
|
||||||
appId = app.appId
|
if (dsProvider) {
|
||||||
|
datasource = await config.api.datasource.create(await dsProvider)
|
||||||
|
isInternal = false
|
||||||
|
} else {
|
||||||
|
isInternal = true
|
||||||
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("create", () => {
|
describe("create", () => {
|
||||||
|
@ -45,102 +55,28 @@ describe("/tables", () => {
|
||||||
jest.clearAllMocks()
|
jest.clearAllMocks()
|
||||||
})
|
})
|
||||||
|
|
||||||
const createTable = (table?: Table) => {
|
it("creates a table successfully", async () => {
|
||||||
if (!table) {
|
const name = generator.guid()
|
||||||
table = basicTable()
|
const table = await config.api.table.save(
|
||||||
}
|
tableForDatasource(datasource, { name })
|
||||||
return request
|
|
||||||
.post(`/api/tables`)
|
|
||||||
.send(table)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
}
|
|
||||||
|
|
||||||
it("returns a success message when the table is successfully created", async () => {
|
|
||||||
const res = await createTable()
|
|
||||||
|
|
||||||
expect((res as any).res.statusMessage).toEqual(
|
|
||||||
"Table TestTable saved successfully."
|
|
||||||
)
|
)
|
||||||
expect(res.body.name).toEqual("TestTable")
|
expect(table.name).toEqual(name)
|
||||||
expect(events.table.created).toHaveBeenCalledTimes(1)
|
expect(events.table.created).toHaveBeenCalledTimes(1)
|
||||||
expect(events.table.created).toHaveBeenCalledWith(res.body)
|
expect(events.table.created).toHaveBeenCalledWith(table)
|
||||||
})
|
|
||||||
|
|
||||||
it("creates all the passed fields", async () => {
|
|
||||||
const tableData: TableToBuild = {
|
|
||||||
name: "TestTable",
|
|
||||||
type: "table",
|
|
||||||
schema: {
|
|
||||||
autoId: {
|
|
||||||
name: "id",
|
|
||||||
type: FieldType.NUMBER,
|
|
||||||
subtype: AutoFieldSubType.AUTO_ID,
|
|
||||||
autocolumn: true,
|
|
||||||
constraints: {
|
|
||||||
type: "number",
|
|
||||||
presence: false,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
views: {
|
|
||||||
"table view": {
|
|
||||||
id: "viewId",
|
|
||||||
version: 2,
|
|
||||||
name: "table view",
|
|
||||||
tableId: "tableId",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
const testTable = await config.createTable(tableData)
|
|
||||||
|
|
||||||
const expected: Table = {
|
|
||||||
...tableData,
|
|
||||||
type: "table",
|
|
||||||
views: {
|
|
||||||
"table view": {
|
|
||||||
...tableData.views!["table view"],
|
|
||||||
schema: {
|
|
||||||
autoId: {
|
|
||||||
autocolumn: true,
|
|
||||||
constraints: {
|
|
||||||
presence: false,
|
|
||||||
type: "number",
|
|
||||||
},
|
|
||||||
name: "id",
|
|
||||||
type: FieldType.NUMBER,
|
|
||||||
subtype: AutoFieldSubType.AUTO_ID,
|
|
||||||
visible: false,
|
|
||||||
} as NumberFieldMetadata,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
sourceType: TableSourceType.INTERNAL,
|
|
||||||
sourceId: expect.any(String),
|
|
||||||
_rev: expect.stringMatching(/^1-.+/),
|
|
||||||
_id: expect.any(String),
|
|
||||||
createdAt: mocks.date.MOCK_DATE.toISOString(),
|
|
||||||
updatedAt: mocks.date.MOCK_DATE.toISOString(),
|
|
||||||
}
|
|
||||||
expect(testTable).toEqual(expected)
|
|
||||||
|
|
||||||
const persistedTable = await config.api.table.get(testTable._id!)
|
|
||||||
expect(persistedTable).toEqual(expected)
|
|
||||||
})
|
})
|
||||||
|
|
||||||
it("creates a table via data import", async () => {
|
it("creates a table via data import", async () => {
|
||||||
const table: SaveTableRequest = basicTable()
|
const table: SaveTableRequest = basicTable()
|
||||||
table.rows = [{ name: "test-name", description: "test-desc" }]
|
table.rows = [{ name: "test-name", description: "test-desc" }]
|
||||||
|
|
||||||
const res = await createTable(table)
|
const res = await config.api.table.save(table)
|
||||||
|
|
||||||
expect(events.table.created).toHaveBeenCalledTimes(1)
|
expect(events.table.created).toHaveBeenCalledTimes(1)
|
||||||
expect(events.table.created).toHaveBeenCalledWith(res.body)
|
expect(events.table.created).toHaveBeenCalledWith(res)
|
||||||
expect(events.table.imported).toHaveBeenCalledTimes(1)
|
expect(events.table.imported).toHaveBeenCalledTimes(1)
|
||||||
expect(events.table.imported).toHaveBeenCalledWith(res.body)
|
expect(events.table.imported).toHaveBeenCalledWith(res)
|
||||||
expect(events.rows.imported).toHaveBeenCalledTimes(1)
|
expect(events.rows.imported).toHaveBeenCalledTimes(1)
|
||||||
expect(events.rows.imported).toHaveBeenCalledWith(res.body, 1)
|
expect(events.rows.imported).toHaveBeenCalledWith(res, 1)
|
||||||
})
|
})
|
||||||
|
|
||||||
it("should apply authorization to endpoint", async () => {
|
it("should apply authorization to endpoint", async () => {
|
||||||
|
@ -155,21 +91,31 @@ describe("/tables", () => {
|
||||||
|
|
||||||
describe("update", () => {
|
describe("update", () => {
|
||||||
it("updates a table", async () => {
|
it("updates a table", async () => {
|
||||||
const testTable = await config.createTable()
|
const table = await config.api.table.save(
|
||||||
|
tableForDatasource(datasource, {
|
||||||
|
schema: {
|
||||||
|
name: {
|
||||||
|
type: FieldType.STRING,
|
||||||
|
name: "name",
|
||||||
|
constraints: {
|
||||||
|
type: "string",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
)
|
||||||
|
|
||||||
const res = await request
|
const updatedTable = await config.api.table.save({
|
||||||
.post(`/api/tables`)
|
...table,
|
||||||
.send(testTable)
|
name: generator.guid(),
|
||||||
.set(config.defaultHeaders())
|
})
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
expect(events.table.updated).toHaveBeenCalledTimes(1)
|
expect(events.table.updated).toHaveBeenCalledTimes(1)
|
||||||
expect(events.table.updated).toHaveBeenCalledWith(res.body)
|
expect(events.table.updated).toHaveBeenCalledWith(updatedTable)
|
||||||
})
|
})
|
||||||
|
|
||||||
it("updates all the row fields for a table when a schema key is renamed", async () => {
|
it("updates all the row fields for a table when a schema key is renamed", async () => {
|
||||||
const testTable = await config.createTable()
|
const testTable = await config.api.table.save(basicTable(datasource))
|
||||||
await config.createLegacyView({
|
await config.createLegacyView({
|
||||||
name: "TestView",
|
name: "TestView",
|
||||||
field: "Price",
|
field: "Price",
|
||||||
|
@ -179,112 +125,96 @@ describe("/tables", () => {
|
||||||
filters: [],
|
filters: [],
|
||||||
})
|
})
|
||||||
|
|
||||||
const testRow = await request
|
const testRow = await config.api.row.save(testTable._id!, {
|
||||||
.post(`/api/${testTable._id}/rows`)
|
name: "test",
|
||||||
.send({
|
})
|
||||||
name: "test",
|
|
||||||
})
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
const updatedTable = await request
|
const { name, ...otherColumns } = testTable.schema
|
||||||
.post(`/api/tables`)
|
const updatedTable = await config.api.table.save({
|
||||||
.send({
|
...testTable,
|
||||||
_id: testTable._id,
|
_rename: {
|
||||||
_rev: testTable._rev,
|
old: "name",
|
||||||
name: "TestTable",
|
updated: "updatedName",
|
||||||
key: "name",
|
},
|
||||||
_rename: {
|
schema: {
|
||||||
old: "name",
|
...otherColumns,
|
||||||
updated: "updatedName",
|
updatedName: {
|
||||||
|
...name,
|
||||||
|
name: "updatedName",
|
||||||
},
|
},
|
||||||
schema: {
|
},
|
||||||
updatedName: { type: "string" },
|
})
|
||||||
},
|
|
||||||
})
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
expect((updatedTable as any).res.statusMessage).toEqual(
|
|
||||||
"Table TestTable saved successfully."
|
|
||||||
)
|
|
||||||
expect(updatedTable.body.name).toEqual("TestTable")
|
|
||||||
|
|
||||||
const res = await request
|
expect(updatedTable.name).toEqual(testTable.name)
|
||||||
.get(`/api/${testTable._id}/rows/${testRow.body._id}`)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
expect(res.body.updatedName).toEqual("test")
|
const res = await config.api.row.get(testTable._id!, testRow._id!)
|
||||||
expect(res.body.name).toBeUndefined()
|
expect(res.updatedName).toEqual("test")
|
||||||
|
expect(res.name).toBeUndefined()
|
||||||
})
|
})
|
||||||
|
|
||||||
it("updates only the passed fields", async () => {
|
it("updates only the passed fields", async () => {
|
||||||
const testTable = await config.createTable({
|
await timekeeper.withFreeze(new Date(2021, 1, 1), async () => {
|
||||||
name: "TestTable",
|
const table = await config.api.table.save(
|
||||||
type: "table",
|
tableForDatasource(datasource, {
|
||||||
schema: {
|
schema: {
|
||||||
autoId: {
|
autoId: {
|
||||||
name: "id",
|
name: "id",
|
||||||
type: FieldType.NUMBER,
|
type: FieldType.NUMBER,
|
||||||
subtype: AutoFieldSubType.AUTO_ID,
|
subtype: AutoFieldSubType.AUTO_ID,
|
||||||
autocolumn: true,
|
autocolumn: true,
|
||||||
constraints: {
|
constraints: {
|
||||||
type: "number",
|
type: "number",
|
||||||
presence: false,
|
presence: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
},
|
})
|
||||||
},
|
)
|
||||||
views: {
|
|
||||||
view1: {
|
|
||||||
id: "viewId",
|
|
||||||
version: 2,
|
|
||||||
name: "table view",
|
|
||||||
tableId: "tableId",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
const response = await request
|
const newName = generator.guid()
|
||||||
.post(`/api/tables`)
|
|
||||||
.send({
|
const updatedTable = await config.api.table.save({
|
||||||
...testTable,
|
...table,
|
||||||
name: "UpdatedName",
|
name: newName,
|
||||||
})
|
})
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
expect(response.body).toEqual({
|
let expected: Table = {
|
||||||
...testTable,
|
...table,
|
||||||
name: "UpdatedName",
|
name: newName,
|
||||||
_rev: expect.stringMatching(/^2-.+/),
|
_id: expect.any(String),
|
||||||
})
|
}
|
||||||
|
if (isInternal) {
|
||||||
|
expected._rev = expect.stringMatching(/^2-.+/)
|
||||||
|
}
|
||||||
|
|
||||||
const persistedTable = await config.api.table.get(testTable._id!)
|
expect(updatedTable).toEqual(expected)
|
||||||
expect(persistedTable).toEqual({
|
|
||||||
...testTable,
|
const persistedTable = await config.api.table.get(updatedTable._id!)
|
||||||
name: "UpdatedName",
|
expected = {
|
||||||
_rev: expect.stringMatching(/^2-.+/),
|
...table,
|
||||||
|
name: newName,
|
||||||
|
_id: updatedTable._id,
|
||||||
|
}
|
||||||
|
if (datasource?.isSQL) {
|
||||||
|
expected.sql = true
|
||||||
|
}
|
||||||
|
if (isInternal) {
|
||||||
|
expected._rev = expect.stringMatching(/^2-.+/)
|
||||||
|
}
|
||||||
|
expect(persistedTable).toEqual(expected)
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("user table", () => {
|
describe("user table", () => {
|
||||||
it("should add roleId and email field when adjusting user table schema", async () => {
|
isInternal &&
|
||||||
const res = await request
|
it("should add roleId and email field when adjusting user table schema", async () => {
|
||||||
.post(`/api/tables`)
|
const table = await config.api.table.save({
|
||||||
.send({
|
...basicTable(datasource),
|
||||||
...basicTable(),
|
|
||||||
_id: "ta_users",
|
_id: "ta_users",
|
||||||
})
|
})
|
||||||
.set(config.defaultHeaders())
|
expect(table.schema.email).toBeDefined()
|
||||||
.expect("Content-Type", /json/)
|
expect(table.schema.roleId).toBeDefined()
|
||||||
.expect(200)
|
})
|
||||||
expect(res.body.schema.email).toBeDefined()
|
|
||||||
expect(res.body.schema.roleId).toBeDefined()
|
|
||||||
})
|
|
||||||
})
|
})
|
||||||
|
|
||||||
it("should add a new column for an internal DB table", async () => {
|
it("should add a new column for an internal DB table", async () => {
|
||||||
|
@ -295,12 +225,7 @@ describe("/tables", () => {
|
||||||
...basicTable(),
|
...basicTable(),
|
||||||
}
|
}
|
||||||
|
|
||||||
const response = await request
|
const response = await config.api.table.save(saveTableRequest)
|
||||||
.post(`/api/tables`)
|
|
||||||
.send(saveTableRequest)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
const expectedResponse = {
|
const expectedResponse = {
|
||||||
...saveTableRequest,
|
...saveTableRequest,
|
||||||
|
@ -311,15 +236,16 @@ describe("/tables", () => {
|
||||||
views: {},
|
views: {},
|
||||||
}
|
}
|
||||||
delete expectedResponse._add
|
delete expectedResponse._add
|
||||||
|
expect(response).toEqual(expectedResponse)
|
||||||
expect(response.status).toBe(200)
|
|
||||||
expect(response.body).toEqual(expectedResponse)
|
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("import", () => {
|
describe("import", () => {
|
||||||
it("imports rows successfully", async () => {
|
it("imports rows successfully", async () => {
|
||||||
const table = await config.createTable()
|
const name = generator.guid()
|
||||||
|
const table = await config.api.table.save(
|
||||||
|
basicTable(datasource, { name })
|
||||||
|
)
|
||||||
const importRequest = {
|
const importRequest = {
|
||||||
schema: table.schema,
|
schema: table.schema,
|
||||||
rows: [{ name: "test-name", description: "test-desc" }],
|
rows: [{ name: "test-name", description: "test-desc" }],
|
||||||
|
@ -327,83 +253,36 @@ describe("/tables", () => {
|
||||||
|
|
||||||
jest.clearAllMocks()
|
jest.clearAllMocks()
|
||||||
|
|
||||||
await request
|
await config.api.table.import(table._id!, importRequest)
|
||||||
.post(`/api/tables/${table._id}/import`)
|
|
||||||
.send(importRequest)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
expect(events.table.created).not.toHaveBeenCalled()
|
expect(events.table.created).not.toHaveBeenCalled()
|
||||||
expect(events.rows.imported).toHaveBeenCalledTimes(1)
|
expect(events.rows.imported).toHaveBeenCalledTimes(1)
|
||||||
expect(events.rows.imported).toHaveBeenCalledWith(
|
expect(events.rows.imported).toHaveBeenCalledWith(
|
||||||
expect.objectContaining({
|
expect.objectContaining({
|
||||||
name: "TestTable",
|
name,
|
||||||
_id: table._id,
|
_id: table._id,
|
||||||
}),
|
}),
|
||||||
1
|
1
|
||||||
)
|
)
|
||||||
})
|
})
|
||||||
|
|
||||||
it("should update Auto ID field after bulk import", async () => {
|
|
||||||
const table = await config.createTable({
|
|
||||||
name: "TestTable",
|
|
||||||
type: "table",
|
|
||||||
schema: {
|
|
||||||
autoId: {
|
|
||||||
name: "id",
|
|
||||||
type: FieldType.NUMBER,
|
|
||||||
subtype: AutoFieldSubType.AUTO_ID,
|
|
||||||
autocolumn: true,
|
|
||||||
constraints: {
|
|
||||||
type: "number",
|
|
||||||
presence: false,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
let row = await config.api.row.save(table._id!, {})
|
|
||||||
expect(row.autoId).toEqual(1)
|
|
||||||
|
|
||||||
await config.api.row.bulkImport(table._id!, {
|
|
||||||
rows: [{ autoId: 2 }],
|
|
||||||
identifierFields: [],
|
|
||||||
})
|
|
||||||
|
|
||||||
row = await config.api.row.save(table._id!, {})
|
|
||||||
expect(row.autoId).toEqual(3)
|
|
||||||
})
|
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("fetch", () => {
|
describe("fetch", () => {
|
||||||
let testTable: Table
|
let testTable: Table
|
||||||
const enrichViewSchemasMock = jest.spyOn(sdk.tables, "enrichViewSchemas")
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
beforeEach(async () => {
|
||||||
testTable = await config.createTable(testTable)
|
testTable = await config.api.table.save(
|
||||||
|
basicTable(datasource, { name: generator.guid() })
|
||||||
|
)
|
||||||
})
|
})
|
||||||
|
|
||||||
afterEach(() => {
|
it("returns all tables", async () => {
|
||||||
delete testTable._rev
|
const res = await config.api.table.fetch()
|
||||||
})
|
const table = res.find(t => t._id === testTable._id)
|
||||||
|
|
||||||
afterAll(() => {
|
|
||||||
enrichViewSchemasMock.mockRestore()
|
|
||||||
})
|
|
||||||
|
|
||||||
it("returns all the tables for that instance in the response body", async () => {
|
|
||||||
const res = await request
|
|
||||||
.get(`/api/tables`)
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
const table = res.body.find((t: Table) => t._id === testTable._id)
|
|
||||||
expect(table).toBeDefined()
|
expect(table).toBeDefined()
|
||||||
expect(table.name).toEqual(testTable.name)
|
expect(table!.name).toEqual(testTable.name)
|
||||||
expect(table.type).toEqual("table")
|
expect(table!.type).toEqual("table")
|
||||||
expect(table.sourceType).toEqual("internal")
|
expect(table!.sourceType).toEqual(testTable.sourceType)
|
||||||
})
|
})
|
||||||
|
|
||||||
it("should apply authorization to endpoint", async () => {
|
it("should apply authorization to endpoint", async () => {
|
||||||
|
@ -414,99 +293,110 @@ describe("/tables", () => {
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
it("should fetch views", async () => {
|
it("should enrich the view schemas", async () => {
|
||||||
const tableId = config.table!._id!
|
const viewV2 = await config.api.viewV2.create({
|
||||||
const views = [
|
tableId: testTable._id!,
|
||||||
await config.api.viewV2.create({ tableId, name: generator.guid() }),
|
name: generator.guid(),
|
||||||
await config.api.viewV2.create({ tableId, name: generator.guid() }),
|
})
|
||||||
]
|
const legacyView = await config.api.legacyView.save({
|
||||||
|
tableId: testTable._id!,
|
||||||
const res = await request
|
name: generator.guid(),
|
||||||
.get(`/api/tables`)
|
filters: [],
|
||||||
.set(config.defaultHeaders())
|
schema: {},
|
||||||
.expect("Content-Type", /json/)
|
})
|
||||||
.expect(200)
|
|
||||||
|
|
||||||
expect(res.body).toEqual(
|
|
||||||
expect.arrayContaining([
|
|
||||||
expect.objectContaining({
|
|
||||||
_id: tableId,
|
|
||||||
views: views.reduce((p, c) => {
|
|
||||||
p[c.name] = { ...c, schema: expect.anything() }
|
|
||||||
return p
|
|
||||||
}, {} as any),
|
|
||||||
}),
|
|
||||||
])
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should enrich the view schemas for viewsV2", async () => {
|
|
||||||
const tableId = config.table!._id!
|
|
||||||
enrichViewSchemasMock.mockImplementation(t => ({
|
|
||||||
...t,
|
|
||||||
views: {
|
|
||||||
view1: {
|
|
||||||
version: 2,
|
|
||||||
name: "view1",
|
|
||||||
schema: {},
|
|
||||||
id: "new_view_id",
|
|
||||||
tableId: t._id!,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}))
|
|
||||||
|
|
||||||
await config.api.viewV2.create({ tableId, name: generator.guid() })
|
|
||||||
await config.createLegacyView()
|
|
||||||
|
|
||||||
const res = await config.api.table.fetch()
|
const res = await config.api.table.fetch()
|
||||||
|
|
||||||
expect(res).toEqual(
|
const table = res.find(t => t._id === testTable._id)
|
||||||
expect.arrayContaining([
|
expect(table).toBeDefined()
|
||||||
expect.objectContaining({
|
expect(table!.views![viewV2.name]).toBeDefined()
|
||||||
_id: tableId,
|
|
||||||
views: {
|
const expectedViewV2: ViewV2Enriched = {
|
||||||
view1: {
|
...viewV2,
|
||||||
version: 2,
|
schema: {
|
||||||
name: "view1",
|
description: {
|
||||||
schema: {},
|
constraints: {
|
||||||
id: "new_view_id",
|
type: "string",
|
||||||
tableId,
|
|
||||||
},
|
|
||||||
},
|
},
|
||||||
}),
|
name: "description",
|
||||||
])
|
type: FieldType.STRING,
|
||||||
|
visible: false,
|
||||||
|
},
|
||||||
|
name: {
|
||||||
|
constraints: {
|
||||||
|
type: "string",
|
||||||
|
},
|
||||||
|
name: "name",
|
||||||
|
type: FieldType.STRING,
|
||||||
|
visible: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!isInternal) {
|
||||||
|
expectedViewV2.schema!.id = {
|
||||||
|
name: "id",
|
||||||
|
type: FieldType.NUMBER,
|
||||||
|
visible: false,
|
||||||
|
autocolumn: true,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
expect(table!.views![viewV2.name!]).toEqual(expectedViewV2)
|
||||||
|
|
||||||
|
if (isInternal) {
|
||||||
|
expect(table!.views![legacyView.name!]).toBeDefined()
|
||||||
|
expect(table!.views![legacyView.name!]).toEqual({
|
||||||
|
...legacyView,
|
||||||
|
schema: {
|
||||||
|
description: {
|
||||||
|
constraints: {
|
||||||
|
type: "string",
|
||||||
|
},
|
||||||
|
name: "description",
|
||||||
|
type: "string",
|
||||||
|
},
|
||||||
|
name: {
|
||||||
|
constraints: {
|
||||||
|
type: "string",
|
||||||
|
},
|
||||||
|
name: "name",
|
||||||
|
type: "string",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
}
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe("get", () => {
|
||||||
|
it("returns a table", async () => {
|
||||||
|
const table = await config.api.table.save(
|
||||||
|
basicTable(datasource, { name: generator.guid() })
|
||||||
)
|
)
|
||||||
|
const res = await config.api.table.get(table._id!)
|
||||||
|
expect(res).toEqual(table)
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("indexing", () => {
|
describe("indexing", () => {
|
||||||
it("should be able to create a table with indexes", async () => {
|
it("should be able to create a table with indexes", async () => {
|
||||||
await context.doInAppContext(appId, async () => {
|
await context.doInAppContext(config.getAppId(), async () => {
|
||||||
const db = context.getAppDB()
|
const db = context.getAppDB()
|
||||||
const indexCount = (await db.getIndexes()).total_rows
|
const indexCount = (await db.getIndexes()).total_rows
|
||||||
const table = basicTable()
|
const table = basicTable()
|
||||||
table.indexes = ["name"]
|
table.indexes = ["name"]
|
||||||
const res = await request
|
const res = await config.api.table.save(table)
|
||||||
.post(`/api/tables`)
|
expect(res._id).toBeDefined()
|
||||||
.send(table)
|
expect(res._rev).toBeDefined()
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
expect(res.body._id).toBeDefined()
|
|
||||||
expect(res.body._rev).toBeDefined()
|
|
||||||
expect((await db.getIndexes()).total_rows).toEqual(indexCount + 1)
|
expect((await db.getIndexes()).total_rows).toEqual(indexCount + 1)
|
||||||
// update index to see what happens
|
// update index to see what happens
|
||||||
table.indexes = ["name", "description"]
|
table.indexes = ["name", "description"]
|
||||||
await request
|
await config.api.table.save({
|
||||||
.post(`/api/tables`)
|
...table,
|
||||||
.send({
|
_id: res._id,
|
||||||
...table,
|
_rev: res._rev,
|
||||||
_id: res.body._id,
|
})
|
||||||
_rev: res.body._rev,
|
|
||||||
})
|
|
||||||
.set(config.defaultHeaders())
|
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
// shouldn't have created a new index
|
// shouldn't have created a new index
|
||||||
expect((await db.getIndexes()).total_rows).toEqual(indexCount + 1)
|
expect((await db.getIndexes()).total_rows).toEqual(indexCount + 1)
|
||||||
})
|
})
|
||||||
|
@ -521,12 +411,9 @@ describe("/tables", () => {
|
||||||
})
|
})
|
||||||
|
|
||||||
it("returns a success response when a table is deleted.", async () => {
|
it("returns a success response when a table is deleted.", async () => {
|
||||||
const res = await request
|
await config.api.table.destroy(testTable._id!, testTable._rev!, {
|
||||||
.delete(`/api/tables/${testTable._id}/${testTable._rev}`)
|
body: { message: `Table ${testTable._id} deleted.` },
|
||||||
.set(config.defaultHeaders())
|
})
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
expect(res.body.message).toEqual(`Table ${testTable._id} deleted.`)
|
|
||||||
expect(events.table.deleted).toHaveBeenCalledTimes(1)
|
expect(events.table.deleted).toHaveBeenCalledTimes(1)
|
||||||
expect(events.table.deleted).toHaveBeenCalledWith({
|
expect(events.table.deleted).toHaveBeenCalledWith({
|
||||||
...testTable,
|
...testTable,
|
||||||
|
@ -559,12 +446,9 @@ describe("/tables", () => {
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
|
||||||
const res = await request
|
await config.api.table.destroy(testTable._id!, testTable._rev!, {
|
||||||
.delete(`/api/tables/${testTable._id}/${testTable._rev}`)
|
body: { message: `Table ${testTable._id} deleted.` },
|
||||||
.set(config.defaultHeaders())
|
})
|
||||||
.expect("Content-Type", /json/)
|
|
||||||
.expect(200)
|
|
||||||
expect(res.body.message).toEqual(`Table ${testTable._id} deleted.`)
|
|
||||||
const dependentTable = await config.api.table.get(linkedTable._id!)
|
const dependentTable = await config.api.table.get(linkedTable._id!)
|
||||||
expect(dependentTable.schema.TestTable).not.toBeDefined()
|
expect(dependentTable.schema.TestTable).not.toBeDefined()
|
||||||
})
|
})
|
||||||
|
@ -816,33 +700,31 @@ describe("/tables", () => {
|
||||||
describe("unhappy paths", () => {
|
describe("unhappy paths", () => {
|
||||||
let table: Table
|
let table: Table
|
||||||
beforeAll(async () => {
|
beforeAll(async () => {
|
||||||
table = await config.api.table.save({
|
table = await config.api.table.save(
|
||||||
name: "table",
|
tableForDatasource(datasource, {
|
||||||
type: "table",
|
schema: {
|
||||||
sourceId: INTERNAL_TABLE_SOURCE_ID,
|
"user relationship": {
|
||||||
sourceType: TableSourceType.INTERNAL,
|
type: FieldType.LINK,
|
||||||
schema: {
|
fieldName: "test",
|
||||||
"user relationship": {
|
name: "user relationship",
|
||||||
type: FieldType.LINK,
|
constraints: {
|
||||||
fieldName: "test",
|
type: "array",
|
||||||
name: "user relationship",
|
presence: false,
|
||||||
constraints: {
|
},
|
||||||
type: "array",
|
relationshipType: RelationshipType.MANY_TO_ONE,
|
||||||
presence: false,
|
tableId: InternalTable.USER_METADATA,
|
||||||
},
|
},
|
||||||
relationshipType: RelationshipType.MANY_TO_ONE,
|
num: {
|
||||||
tableId: InternalTable.USER_METADATA,
|
type: FieldType.NUMBER,
|
||||||
},
|
name: "num",
|
||||||
num: {
|
constraints: {
|
||||||
type: FieldType.NUMBER,
|
type: "number",
|
||||||
name: "num",
|
presence: false,
|
||||||
constraints: {
|
},
|
||||||
type: "number",
|
|
||||||
presence: false,
|
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
})
|
||||||
})
|
)
|
||||||
})
|
})
|
||||||
|
|
||||||
it("should fail if the new column name is blank", async () => {
|
it("should fail if the new column name is blank", async () => {
|
||||||
|
|
|
@ -25,7 +25,6 @@ import { quotas } from "@budibase/pro"
|
||||||
import { roles } from "@budibase/backend-core"
|
import { roles } from "@budibase/backend-core"
|
||||||
|
|
||||||
jest.unmock("mssql")
|
jest.unmock("mssql")
|
||||||
jest.unmock("pg")
|
|
||||||
|
|
||||||
describe.each([
|
describe.each([
|
||||||
["internal", undefined],
|
["internal", undefined],
|
||||||
|
@ -182,7 +181,7 @@ describe.each([
|
||||||
|
|
||||||
const createdView = await config.api.viewV2.create(newView)
|
const createdView = await config.api.viewV2.create(newView)
|
||||||
|
|
||||||
expect(await config.api.viewV2.get(createdView.id)).toEqual({
|
expect(createdView).toEqual({
|
||||||
...newView,
|
...newView,
|
||||||
schema: {
|
schema: {
|
||||||
Price: {
|
Price: {
|
||||||
|
@ -399,7 +398,7 @@ describe.each([
|
||||||
})
|
})
|
||||||
|
|
||||||
it("updates only UI schema overrides", async () => {
|
it("updates only UI schema overrides", async () => {
|
||||||
await config.api.viewV2.update({
|
const updatedView = await config.api.viewV2.update({
|
||||||
...view,
|
...view,
|
||||||
schema: {
|
schema: {
|
||||||
Price: {
|
Price: {
|
||||||
|
@ -418,7 +417,7 @@ describe.each([
|
||||||
} as Record<string, FieldSchema>,
|
} as Record<string, FieldSchema>,
|
||||||
})
|
})
|
||||||
|
|
||||||
expect(await config.api.viewV2.get(view.id)).toEqual({
|
expect(updatedView).toEqual({
|
||||||
...view,
|
...view,
|
||||||
schema: {
|
schema: {
|
||||||
Price: {
|
Price: {
|
||||||
|
@ -480,17 +479,17 @@ describe.each([
|
||||||
|
|
||||||
describe("fetch view (through table)", () => {
|
describe("fetch view (through table)", () => {
|
||||||
it("should be able to fetch a view V2", async () => {
|
it("should be able to fetch a view V2", async () => {
|
||||||
const newView: CreateViewRequest = {
|
const res = await config.api.viewV2.create({
|
||||||
name: generator.name(),
|
name: generator.name(),
|
||||||
tableId: table._id!,
|
tableId: table._id!,
|
||||||
schema: {
|
schema: {
|
||||||
Price: { visible: false },
|
Price: { visible: false },
|
||||||
Category: { visible: true },
|
Category: { visible: true },
|
||||||
},
|
},
|
||||||
}
|
})
|
||||||
const res = await config.api.viewV2.create(newView)
|
expect(res.schema?.Price).toBeUndefined()
|
||||||
|
|
||||||
const view = await config.api.viewV2.get(res.id)
|
const view = await config.api.viewV2.get(res.id)
|
||||||
expect(view!.schema?.Price).toBeUndefined()
|
|
||||||
const updatedTable = await config.api.table.get(table._id!)
|
const updatedTable = await config.api.table.get(table._id!)
|
||||||
const viewSchema = updatedTable.views![view!.name!].schema as Record<
|
const viewSchema = updatedTable.views![view!.name!].schema as Record<
|
||||||
string,
|
string,
|
||||||
|
|
|
@ -21,7 +21,7 @@ async function start() {
|
||||||
app = koa.app
|
app = koa.app
|
||||||
server = koa.server
|
server = koa.server
|
||||||
// startup includes automation runner - if enabled
|
// startup includes automation runner - if enabled
|
||||||
await startup(app, server)
|
await startup({ app, server })
|
||||||
}
|
}
|
||||||
|
|
||||||
start().catch(err => {
|
start().catch(err => {
|
||||||
|
|
|
@ -1,39 +0,0 @@
|
||||||
const setup = require("./utilities")
|
|
||||||
|
|
||||||
describe("test the execute query action", () => {
|
|
||||||
let query
|
|
||||||
let config = setup.getConfig()
|
|
||||||
|
|
||||||
beforeAll(async () => {
|
|
||||||
await config.init()
|
|
||||||
|
|
||||||
await config.createDatasource()
|
|
||||||
query = await config.createQuery()
|
|
||||||
})
|
|
||||||
|
|
||||||
afterAll(setup.afterAll)
|
|
||||||
|
|
||||||
it("should be able to execute a query", async () => {
|
|
||||||
let res = await setup.runStep(setup.actions.EXECUTE_QUERY.stepId, {
|
|
||||||
query: { queryId: query._id },
|
|
||||||
})
|
|
||||||
expect(res.response).toEqual([{ a: "string", b: 1 }])
|
|
||||||
expect(res.success).toEqual(true)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should handle a null query value", async () => {
|
|
||||||
let res = await setup.runStep(setup.actions.EXECUTE_QUERY.stepId, {
|
|
||||||
query: null,
|
|
||||||
})
|
|
||||||
expect(res.response.message).toEqual("Invalid inputs")
|
|
||||||
expect(res.success).toEqual(false)
|
|
||||||
})
|
|
||||||
|
|
||||||
it("should handle an error executing a query", async () => {
|
|
||||||
let res = await setup.runStep(setup.actions.EXECUTE_QUERY.stepId, {
|
|
||||||
query: { queryId: "wrong_id" },
|
|
||||||
})
|
|
||||||
expect(res.response).toEqual("Error: missing")
|
|
||||||
expect(res.success).toEqual(false)
|
|
||||||
})
|
|
||||||
})
|
|
|
@ -0,0 +1,94 @@
|
||||||
|
import { Datasource, Query, SourceName } from "@budibase/types"
|
||||||
|
import * as setup from "./utilities"
|
||||||
|
import { DatabaseName, getDatasource } from "../../integrations/tests/utils"
|
||||||
|
import knex, { Knex } from "knex"
|
||||||
|
import { generator } from "@budibase/backend-core/tests"
|
||||||
|
|
||||||
|
function getKnexClientName(source: SourceName) {
|
||||||
|
switch (source) {
|
||||||
|
case SourceName.MYSQL:
|
||||||
|
return "mysql2"
|
||||||
|
case SourceName.SQL_SERVER:
|
||||||
|
return "mssql"
|
||||||
|
case SourceName.POSTGRES:
|
||||||
|
return "pg"
|
||||||
|
}
|
||||||
|
throw new Error(`Unsupported source: ${source}`)
|
||||||
|
}
|
||||||
|
|
||||||
|
describe.each(
|
||||||
|
[
|
||||||
|
DatabaseName.POSTGRES,
|
||||||
|
DatabaseName.MYSQL,
|
||||||
|
DatabaseName.SQL_SERVER,
|
||||||
|
DatabaseName.MARIADB,
|
||||||
|
].map(name => [name, getDatasource(name)])
|
||||||
|
)("execute query action (%s)", (_, dsProvider) => {
|
||||||
|
let tableName: string
|
||||||
|
let client: Knex
|
||||||
|
let datasource: Datasource
|
||||||
|
let query: Query
|
||||||
|
let config = setup.getConfig()
|
||||||
|
|
||||||
|
beforeAll(async () => {
|
||||||
|
await config.init()
|
||||||
|
|
||||||
|
const ds = await dsProvider
|
||||||
|
datasource = await config.api.datasource.create(ds)
|
||||||
|
client = knex({
|
||||||
|
client: getKnexClientName(ds.source),
|
||||||
|
connection: ds.config,
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
beforeEach(async () => {
|
||||||
|
tableName = generator.guid()
|
||||||
|
await client.schema.createTable(tableName, table => {
|
||||||
|
table.string("a")
|
||||||
|
table.integer("b")
|
||||||
|
})
|
||||||
|
await client(tableName).insert({ a: "string", b: 1 })
|
||||||
|
query = await config.api.query.save({
|
||||||
|
name: "test query",
|
||||||
|
datasourceId: datasource._id!,
|
||||||
|
parameters: [],
|
||||||
|
fields: {
|
||||||
|
sql: client(tableName).select("*").toSQL().toNative().sql,
|
||||||
|
},
|
||||||
|
transformer: "",
|
||||||
|
schema: {},
|
||||||
|
readable: true,
|
||||||
|
queryVerb: "read",
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
afterEach(async () => {
|
||||||
|
await client.schema.dropTable(tableName)
|
||||||
|
})
|
||||||
|
|
||||||
|
afterAll(setup.afterAll)
|
||||||
|
|
||||||
|
it("should be able to execute a query", async () => {
|
||||||
|
let res = await setup.runStep(setup.actions.EXECUTE_QUERY.stepId, {
|
||||||
|
query: { queryId: query._id },
|
||||||
|
})
|
||||||
|
expect(res.response).toEqual([{ a: "string", b: 1 }])
|
||||||
|
expect(res.success).toEqual(true)
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should handle a null query value", async () => {
|
||||||
|
let res = await setup.runStep(setup.actions.EXECUTE_QUERY.stepId, {
|
||||||
|
query: null,
|
||||||
|
})
|
||||||
|
expect(res.response.message).toEqual("Invalid inputs")
|
||||||
|
expect(res.success).toEqual(false)
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should handle an error executing a query", async () => {
|
||||||
|
let res = await setup.runStep(setup.actions.EXECUTE_QUERY.stepId, {
|
||||||
|
query: { queryId: "wrong_id" },
|
||||||
|
})
|
||||||
|
expect(res.response).toEqual("Error: missing")
|
||||||
|
expect(res.success).toEqual(false)
|
||||||
|
})
|
||||||
|
})
|
|
@ -27,7 +27,6 @@ fetch.mockSearch()
|
||||||
|
|
||||||
const config = setup.getConfig()!
|
const config = setup.getConfig()!
|
||||||
|
|
||||||
jest.unmock("pg")
|
|
||||||
jest.mock("../websockets")
|
jest.mock("../websockets")
|
||||||
|
|
||||||
describe("postgres integrations", () => {
|
describe("postgres integrations", () => {
|
||||||
|
|
|
@ -224,12 +224,12 @@ class SqlTableQueryBuilder {
|
||||||
const tableName = schemaName
|
const tableName = schemaName
|
||||||
? `\`${schemaName}\`.\`${json.table.name}\``
|
? `\`${schemaName}\`.\`${json.table.name}\``
|
||||||
: `\`${json.table.name}\``
|
: `\`${json.table.name}\``
|
||||||
const externalType = json.table.schema[updatedColumn].externalType!
|
|
||||||
return {
|
return {
|
||||||
sql: `alter table ${tableName} change column \`${json.meta.renamed.old}\` \`${updatedColumn}\` ${externalType};`,
|
sql: `alter table ${tableName} rename column \`${json.meta.renamed.old}\` to \`${updatedColumn}\`;`,
|
||||||
bindings: [],
|
bindings: [],
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
query = buildUpdateTable(
|
query = buildUpdateTable(
|
||||||
client,
|
client,
|
||||||
json.table,
|
json.table,
|
||||||
|
@ -237,6 +237,27 @@ class SqlTableQueryBuilder {
|
||||||
json.meta.table,
|
json.meta.table,
|
||||||
json.meta.renamed!
|
json.meta.renamed!
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// renameColumn for SQL Server returns a parameterised `sp_rename` query,
|
||||||
|
// which is not supported by SQL Server and gives a syntax error.
|
||||||
|
if (this.sqlClient === SqlClient.MS_SQL && json.meta.renamed) {
|
||||||
|
const oldColumn = json.meta.renamed.old
|
||||||
|
const updatedColumn = json.meta.renamed.updated
|
||||||
|
const tableName = schemaName
|
||||||
|
? `${schemaName}.${json.table.name}`
|
||||||
|
: `${json.table.name}`
|
||||||
|
const sql = query.toSQL()
|
||||||
|
if (Array.isArray(sql)) {
|
||||||
|
for (const query of sql) {
|
||||||
|
if (query.sql.startsWith("exec sp_rename")) {
|
||||||
|
query.sql = `exec sp_rename '${tableName}.${oldColumn}', '${updatedColumn}', 'COLUMN'`
|
||||||
|
query.bindings = []
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return sql
|
||||||
|
}
|
||||||
break
|
break
|
||||||
case Operation.DELETE_TABLE:
|
case Operation.DELETE_TABLE:
|
||||||
query = buildDeleteTable(client, json.table)
|
query = buildDeleteTable(client, json.table)
|
||||||
|
|
|
@ -722,7 +722,7 @@ describe("SQL query builder", () => {
|
||||||
})
|
})
|
||||||
expect(query).toEqual({
|
expect(query).toEqual({
|
||||||
bindings: [],
|
bindings: [],
|
||||||
sql: `alter table \`${TABLE_NAME}\` change column \`name\` \`first_name\` varchar(45);`,
|
sql: `alter table \`${TABLE_NAME}\` rename column \`name\` to \`first_name\`;`,
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
|
|
|
@ -1,5 +1,3 @@
|
||||||
jest.unmock("pg")
|
|
||||||
|
|
||||||
import { Datasource, SourceName } from "@budibase/types"
|
import { Datasource, SourceName } from "@budibase/types"
|
||||||
import * as postgres from "./postgres"
|
import * as postgres from "./postgres"
|
||||||
import * as mongodb from "./mongodb"
|
import * as mongodb from "./mongodb"
|
||||||
|
|
|
@ -48,6 +48,18 @@ export async function save(
|
||||||
oldTable = await getTable(tableId)
|
oldTable = await getTable(tableId)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (
|
||||||
|
!oldTable &&
|
||||||
|
(tableToSave.primary == null || tableToSave.primary.length === 0)
|
||||||
|
) {
|
||||||
|
tableToSave.primary = ["id"]
|
||||||
|
tableToSave.schema.id = {
|
||||||
|
type: FieldType.NUMBER,
|
||||||
|
autocolumn: true,
|
||||||
|
name: "id",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if (hasTypeChanged(tableToSave, oldTable)) {
|
if (hasTypeChanged(tableToSave, oldTable)) {
|
||||||
throw new Error("A column type has changed.")
|
throw new Error("A column type has changed.")
|
||||||
}
|
}
|
||||||
|
@ -183,6 +195,10 @@ export async function save(
|
||||||
// that the datasource definition changed
|
// that the datasource definition changed
|
||||||
const updatedDatasource = await datasourceSdk.get(datasource._id!)
|
const updatedDatasource = await datasourceSdk.get(datasource._id!)
|
||||||
|
|
||||||
|
if (updatedDatasource.isSQL) {
|
||||||
|
tableToSave.sql = true
|
||||||
|
}
|
||||||
|
|
||||||
return { datasource: updatedDatasource, table: tableToSave }
|
return { datasource: updatedDatasource, table: tableToSave }
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -142,7 +142,9 @@ export function enrichViewSchemas(table: Table): TableResponse {
|
||||||
return {
|
return {
|
||||||
...table,
|
...table,
|
||||||
views: Object.values(table.views ?? [])
|
views: Object.values(table.views ?? [])
|
||||||
.map(v => sdk.views.enrichSchema(v, table.schema))
|
.map(v =>
|
||||||
|
sdk.views.isV2(v) ? sdk.views.enrichSchema(v, table.schema) : v
|
||||||
|
)
|
||||||
.reduce((p, v) => {
|
.reduce((p, v) => {
|
||||||
p[v.name!] = v
|
p[v.name!] = v
|
||||||
return p
|
return p
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
import { ViewV2 } from "@budibase/types"
|
import { ViewV2, ViewV2Enriched } from "@budibase/types"
|
||||||
import { context, HTTPError } from "@budibase/backend-core"
|
import { context, HTTPError } from "@budibase/backend-core"
|
||||||
|
|
||||||
import sdk from "../../../sdk"
|
import sdk from "../../../sdk"
|
||||||
|
@ -6,26 +6,34 @@ import * as utils from "../../../db/utils"
|
||||||
import { enrichSchema, isV2 } from "."
|
import { enrichSchema, isV2 } from "."
|
||||||
import { breakExternalTableId } from "../../../integrations/utils"
|
import { breakExternalTableId } from "../../../integrations/utils"
|
||||||
|
|
||||||
export async function get(
|
export async function get(viewId: string): Promise<ViewV2> {
|
||||||
viewId: string,
|
|
||||||
opts?: { enriched: boolean }
|
|
||||||
): Promise<ViewV2> {
|
|
||||||
const { tableId } = utils.extractViewInfoFromID(viewId)
|
const { tableId } = utils.extractViewInfoFromID(viewId)
|
||||||
|
|
||||||
const { datasourceId, tableName } = breakExternalTableId(tableId)
|
const { datasourceId, tableName } = breakExternalTableId(tableId)
|
||||||
const ds = await sdk.datasources.get(datasourceId!)
|
const ds = await sdk.datasources.get(datasourceId!)
|
||||||
|
|
||||||
const table = ds.entities![tableName!]
|
const table = ds.entities![tableName!]
|
||||||
const views = Object.values(table.views!)
|
const views = Object.values(table.views!).filter(isV2)
|
||||||
const found = views.find(v => isV2(v) && v.id === viewId)
|
const found = views.find(v => v.id === viewId)
|
||||||
if (!found) {
|
if (!found) {
|
||||||
throw new Error("No view found")
|
throw new Error("No view found")
|
||||||
}
|
}
|
||||||
if (opts?.enriched) {
|
return found
|
||||||
return enrichSchema(found, table.schema) as ViewV2
|
}
|
||||||
} else {
|
|
||||||
return found as ViewV2
|
export async function getEnriched(viewId: string): Promise<ViewV2Enriched> {
|
||||||
|
const { tableId } = utils.extractViewInfoFromID(viewId)
|
||||||
|
|
||||||
|
const { datasourceId, tableName } = breakExternalTableId(tableId)
|
||||||
|
const ds = await sdk.datasources.get(datasourceId!)
|
||||||
|
|
||||||
|
const table = ds.entities![tableName!]
|
||||||
|
const views = Object.values(table.views!).filter(isV2)
|
||||||
|
const found = views.find(v => v.id === viewId)
|
||||||
|
if (!found) {
|
||||||
|
throw new Error("No view found")
|
||||||
}
|
}
|
||||||
|
return enrichSchema(found, table.schema)
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function create(
|
export async function create(
|
||||||
|
|
|
@ -1,8 +1,13 @@
|
||||||
import { RenameColumn, TableSchema, View, ViewV2 } from "@budibase/types"
|
import {
|
||||||
|
RenameColumn,
|
||||||
|
TableSchema,
|
||||||
|
View,
|
||||||
|
ViewV2,
|
||||||
|
ViewV2Enriched,
|
||||||
|
} from "@budibase/types"
|
||||||
import { db as dbCore } from "@budibase/backend-core"
|
import { db as dbCore } from "@budibase/backend-core"
|
||||||
import { cloneDeep } from "lodash"
|
import { cloneDeep } from "lodash"
|
||||||
|
|
||||||
import sdk from "../../../sdk"
|
|
||||||
import * as utils from "../../../db/utils"
|
import * as utils from "../../../db/utils"
|
||||||
import { isExternalTableID } from "../../../integrations/utils"
|
import { isExternalTableID } from "../../../integrations/utils"
|
||||||
|
|
||||||
|
@ -16,12 +21,14 @@ function pickApi(tableId: any) {
|
||||||
return internal
|
return internal
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function get(
|
export async function get(viewId: string): Promise<ViewV2> {
|
||||||
viewId: string,
|
|
||||||
opts?: { enriched: boolean }
|
|
||||||
): Promise<ViewV2> {
|
|
||||||
const { tableId } = utils.extractViewInfoFromID(viewId)
|
const { tableId } = utils.extractViewInfoFromID(viewId)
|
||||||
return pickApi(tableId).get(viewId, opts)
|
return pickApi(tableId).get(viewId)
|
||||||
|
}
|
||||||
|
|
||||||
|
export async function getEnriched(viewId: string): Promise<ViewV2Enriched> {
|
||||||
|
const { tableId } = utils.extractViewInfoFromID(viewId)
|
||||||
|
return pickApi(tableId).getEnriched(viewId)
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function create(
|
export async function create(
|
||||||
|
@ -52,11 +59,10 @@ export function allowedFields(view: View | ViewV2) {
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
|
||||||
export function enrichSchema(view: View | ViewV2, tableSchema: TableSchema) {
|
export function enrichSchema(
|
||||||
if (!sdk.views.isV2(view)) {
|
view: ViewV2,
|
||||||
return view
|
tableSchema: TableSchema
|
||||||
}
|
): ViewV2Enriched {
|
||||||
|
|
||||||
let schema = cloneDeep(tableSchema)
|
let schema = cloneDeep(tableSchema)
|
||||||
const anyViewOrder = Object.values(view.schema || {}).some(
|
const anyViewOrder = Object.values(view.schema || {}).some(
|
||||||
ui => ui.order != null
|
ui => ui.order != null
|
||||||
|
|
|
@ -1,26 +1,30 @@
|
||||||
import { ViewV2 } from "@budibase/types"
|
import { ViewV2, ViewV2Enriched } from "@budibase/types"
|
||||||
import { context, HTTPError } from "@budibase/backend-core"
|
import { context, HTTPError } from "@budibase/backend-core"
|
||||||
|
|
||||||
import sdk from "../../../sdk"
|
import sdk from "../../../sdk"
|
||||||
import * as utils from "../../../db/utils"
|
import * as utils from "../../../db/utils"
|
||||||
import { enrichSchema, isV2 } from "."
|
import { enrichSchema, isV2 } from "."
|
||||||
|
|
||||||
export async function get(
|
export async function get(viewId: string): Promise<ViewV2> {
|
||||||
viewId: string,
|
|
||||||
opts?: { enriched: boolean }
|
|
||||||
): Promise<ViewV2> {
|
|
||||||
const { tableId } = utils.extractViewInfoFromID(viewId)
|
const { tableId } = utils.extractViewInfoFromID(viewId)
|
||||||
const table = await sdk.tables.getTable(tableId)
|
const table = await sdk.tables.getTable(tableId)
|
||||||
const views = Object.values(table.views!)
|
const views = Object.values(table.views!).filter(isV2)
|
||||||
const found = views.find(v => isV2(v) && v.id === viewId)
|
const found = views.find(v => v.id === viewId)
|
||||||
if (!found) {
|
if (!found) {
|
||||||
throw new Error("No view found")
|
throw new Error("No view found")
|
||||||
}
|
}
|
||||||
if (opts?.enriched) {
|
return found
|
||||||
return enrichSchema(found, table.schema) as ViewV2
|
}
|
||||||
} else {
|
|
||||||
return found as ViewV2
|
export async function getEnriched(viewId: string): Promise<ViewV2Enriched> {
|
||||||
|
const { tableId } = utils.extractViewInfoFromID(viewId)
|
||||||
|
const table = await sdk.tables.getTable(tableId)
|
||||||
|
const views = Object.values(table.views!).filter(isV2)
|
||||||
|
const found = views.find(v => v.id === viewId)
|
||||||
|
if (!found) {
|
||||||
|
throw new Error("No view found")
|
||||||
}
|
}
|
||||||
|
return enrichSchema(found, table.schema)
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function create(
|
export async function create(
|
||||||
|
|
|
@ -1,4 +1,12 @@
|
||||||
import newid from "../../db/newid"
|
import TestConfig from "../../tests/utilities/TestConfiguration"
|
||||||
|
import { db as dbCore } from "@budibase/backend-core"
|
||||||
|
import sdk from "../index"
|
||||||
|
import {
|
||||||
|
FieldType,
|
||||||
|
INTERNAL_TABLE_SOURCE_ID,
|
||||||
|
TableSourceType,
|
||||||
|
} from "@budibase/types"
|
||||||
|
import { FIND_LIMIT } from "../app/rows/attachments"
|
||||||
|
|
||||||
const attachment = {
|
const attachment = {
|
||||||
size: 73479,
|
size: 73479,
|
||||||
|
@ -8,69 +16,48 @@ const attachment = {
|
||||||
key: "app_bbb/attachments/a.png",
|
key: "app_bbb/attachments/a.png",
|
||||||
}
|
}
|
||||||
|
|
||||||
const row = {
|
|
||||||
_id: "ro_ta_aaa",
|
|
||||||
photo: [attachment],
|
|
||||||
otherCol: "string",
|
|
||||||
}
|
|
||||||
|
|
||||||
const table = {
|
|
||||||
_id: "ta_aaa",
|
|
||||||
name: "photos",
|
|
||||||
schema: {
|
|
||||||
photo: {
|
|
||||||
type: "attachment",
|
|
||||||
name: "photo",
|
|
||||||
},
|
|
||||||
otherCol: {
|
|
||||||
type: "string",
|
|
||||||
name: "otherCol",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
jest.mock("@budibase/backend-core", () => {
|
|
||||||
const core = jest.requireActual("@budibase/backend-core")
|
|
||||||
return {
|
|
||||||
...core,
|
|
||||||
db: {
|
|
||||||
...core.db,
|
|
||||||
directCouchFind: jest.fn(),
|
|
||||||
},
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
import { db as dbCore } from "@budibase/backend-core"
|
|
||||||
import sdk from "../index"
|
|
||||||
|
|
||||||
describe("should be able to re-write attachment URLs", () => {
|
describe("should be able to re-write attachment URLs", () => {
|
||||||
|
const config = new TestConfig()
|
||||||
|
|
||||||
|
beforeAll(async () => {
|
||||||
|
await config.init()
|
||||||
|
})
|
||||||
|
|
||||||
it("should update URLs on a number of rows over the limit", async () => {
|
it("should update URLs on a number of rows over the limit", async () => {
|
||||||
const db = dbCore.getDB("app_aaa")
|
const table = await config.api.table.save({
|
||||||
await db.put(table)
|
name: "photos",
|
||||||
const limit = 30
|
type: "table",
|
||||||
let rows = []
|
sourceId: INTERNAL_TABLE_SOURCE_ID,
|
||||||
for (let i = 0; i < limit; i++) {
|
sourceType: TableSourceType.INTERNAL,
|
||||||
const rowToWrite = {
|
schema: {
|
||||||
...row,
|
photo: {
|
||||||
_id: `${row._id}_${newid()}`,
|
type: FieldType.ATTACHMENT,
|
||||||
}
|
name: "photo",
|
||||||
const { rev } = await db.put(rowToWrite)
|
},
|
||||||
rows.push({
|
otherCol: {
|
||||||
...rowToWrite,
|
type: FieldType.STRING,
|
||||||
_rev: rev,
|
name: "otherCol",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
for (let i = 0; i < FIND_LIMIT * 4; i++) {
|
||||||
|
await config.api.row.save(table._id!, {
|
||||||
|
photo: [attachment],
|
||||||
|
otherCol: "string",
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
dbCore.directCouchFind
|
const db = dbCore.getDB(config.getAppId())
|
||||||
// @ts-ignore
|
|
||||||
.mockReturnValueOnce({ rows: rows.slice(0, 25), bookmark: "aaa" })
|
|
||||||
.mockReturnValueOnce({ rows: rows.slice(25, limit), bookmark: "bbb" })
|
|
||||||
await sdk.backups.updateAttachmentColumns(db.name, db)
|
await sdk.backups.updateAttachmentColumns(db.name, db)
|
||||||
const finalRows = await sdk.rows.getAllInternalRows(db.name)
|
|
||||||
for (let rowToCheck of finalRows) {
|
const rows = (await sdk.rows.getAllInternalRows(db.name)).filter(
|
||||||
expect(rowToCheck.otherCol).toBe(row.otherCol)
|
row => row.tableId === table._id
|
||||||
expect(rowToCheck.photo[0].url).toBe("")
|
)
|
||||||
expect(rowToCheck.photo[0].key).toBe(`${db.name}/attachments/a.png`)
|
for (const row of rows) {
|
||||||
|
expect(row.otherCol).toBe("string")
|
||||||
|
expect(row.photo[0].url).toBe("")
|
||||||
|
expect(row.photo[0].key).toBe(`${db.name}/attachments/a.png`)
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
|
@ -35,11 +35,20 @@ describe("syncGlobalUsers", () => {
|
||||||
builder: { global: true },
|
builder: { global: true },
|
||||||
})
|
})
|
||||||
await config.doInContext(config.appId, async () => {
|
await config.doInContext(config.appId, async () => {
|
||||||
expect(await rawUserMetadata()).toHaveLength(1)
|
let metadata = await rawUserMetadata()
|
||||||
|
expect(metadata).not.toContainEqual(
|
||||||
|
expect.objectContaining({
|
||||||
|
_id: db.generateUserMetadataID(user1._id!),
|
||||||
|
})
|
||||||
|
)
|
||||||
|
expect(metadata).not.toContainEqual(
|
||||||
|
expect.objectContaining({
|
||||||
|
_id: db.generateUserMetadataID(user2._id!),
|
||||||
|
})
|
||||||
|
)
|
||||||
await syncGlobalUsers()
|
await syncGlobalUsers()
|
||||||
|
|
||||||
const metadata = await rawUserMetadata()
|
metadata = await rawUserMetadata()
|
||||||
expect(metadata).toHaveLength(3)
|
|
||||||
expect(metadata).toContainEqual(
|
expect(metadata).toContainEqual(
|
||||||
expect.objectContaining({
|
expect.objectContaining({
|
||||||
_id: db.generateUserMetadataID(user1._id!),
|
_id: db.generateUserMetadataID(user1._id!),
|
||||||
|
@ -62,7 +71,6 @@ describe("syncGlobalUsers", () => {
|
||||||
await syncGlobalUsers()
|
await syncGlobalUsers()
|
||||||
|
|
||||||
const metadata = await rawUserMetadata()
|
const metadata = await rawUserMetadata()
|
||||||
expect(metadata).toHaveLength(1)
|
|
||||||
expect(metadata).not.toContainEqual(
|
expect(metadata).not.toContainEqual(
|
||||||
expect.objectContaining({
|
expect.objectContaining({
|
||||||
_id: db.generateUserMetadataID(user._id!),
|
_id: db.generateUserMetadataID(user._id!),
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
import env from "./environment"
|
import env from "../environment"
|
||||||
import * as redis from "./utilities/redis"
|
import * as redis from "../utilities/redis"
|
||||||
import { generateApiKey, getChecklist } from "./utilities/workerRequests"
|
import { generateApiKey, getChecklist } from "../utilities/workerRequests"
|
||||||
import {
|
import {
|
||||||
events,
|
events,
|
||||||
installation,
|
installation,
|
||||||
|
@ -9,22 +9,22 @@ import {
|
||||||
users,
|
users,
|
||||||
cache,
|
cache,
|
||||||
} from "@budibase/backend-core"
|
} from "@budibase/backend-core"
|
||||||
import fs from "fs"
|
import { watch } from "../watch"
|
||||||
import { watch } from "./watch"
|
import * as automations from "../automations"
|
||||||
import * as automations from "./automations"
|
import * as fileSystem from "../utilities/fileSystem"
|
||||||
import * as fileSystem from "./utilities/fileSystem"
|
import { default as eventEmitter, init as eventInit } from "../events"
|
||||||
import { default as eventEmitter, init as eventInit } from "./events"
|
import * as migrations from "../migrations"
|
||||||
import * as migrations from "./migrations"
|
import * as bullboard from "../automations/bullboard"
|
||||||
import * as bullboard from "./automations/bullboard"
|
|
||||||
import * as pro from "@budibase/pro"
|
import * as pro from "@budibase/pro"
|
||||||
import * as api from "./api"
|
import * as api from "../api"
|
||||||
import sdk from "./sdk"
|
import sdk from "../sdk"
|
||||||
import { initialise as initialiseWebsockets } from "./websockets"
|
import { initialise as initialiseWebsockets } from "../websockets"
|
||||||
import { automationsEnabled, printFeatures } from "./features"
|
import { automationsEnabled, printFeatures } from "../features"
|
||||||
|
import * as jsRunner from "../jsRunner"
|
||||||
import Koa from "koa"
|
import Koa from "koa"
|
||||||
import { Server } from "http"
|
import { Server } from "http"
|
||||||
import { AddressInfo } from "net"
|
import { AddressInfo } from "net"
|
||||||
import * as jsRunner from "./jsRunner"
|
import fs from "fs"
|
||||||
|
|
||||||
let STARTUP_RAN = false
|
let STARTUP_RAN = false
|
||||||
|
|
||||||
|
@ -61,8 +61,11 @@ function shutdown(server?: Server) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function startup(app?: Koa, server?: Server) {
|
export async function startup(
|
||||||
if (STARTUP_RAN) {
|
opts: { app?: Koa; server?: Server; rerun?: boolean } = {}
|
||||||
|
) {
|
||||||
|
const { app, server, rerun } = opts
|
||||||
|
if (STARTUP_RAN && !rerun) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
printFeatures()
|
printFeatures()
|
||||||
|
@ -139,9 +142,9 @@ export async function startup(app?: Koa, server?: Server) {
|
||||||
try {
|
try {
|
||||||
const user = await users.UserDB.createAdminUser(
|
const user = await users.UserDB.createAdminUser(
|
||||||
bbAdminEmail,
|
bbAdminEmail,
|
||||||
bbAdminPassword,
|
|
||||||
tenantId,
|
tenantId,
|
||||||
{
|
{
|
||||||
|
password: bbAdminPassword,
|
||||||
hashPassword: true,
|
hashPassword: true,
|
||||||
requirePassword: true,
|
requirePassword: true,
|
||||||
skipPasswordValidation: true,
|
skipPasswordValidation: true,
|
|
@ -0,0 +1,34 @@
|
||||||
|
import TestConfiguration from "../../tests/utilities/TestConfiguration"
|
||||||
|
import { startup } from "../index"
|
||||||
|
import { users, utils, tenancy } from "@budibase/backend-core"
|
||||||
|
|
||||||
|
describe("check BB_ADMIN environment variables", () => {
|
||||||
|
const config = new TestConfiguration()
|
||||||
|
beforeAll(async () => {
|
||||||
|
await config.init()
|
||||||
|
})
|
||||||
|
|
||||||
|
it("should be able to create a user with the BB_ADMIN environment variables", async () => {
|
||||||
|
const EMAIL = "budibase@budibase.com",
|
||||||
|
PASSWORD = "budibase"
|
||||||
|
await tenancy.doInTenant(tenancy.DEFAULT_TENANT_ID, async () => {
|
||||||
|
await config.withEnv(
|
||||||
|
{
|
||||||
|
BB_ADMIN_USER_EMAIL: EMAIL,
|
||||||
|
BB_ADMIN_USER_PASSWORD: PASSWORD,
|
||||||
|
MULTI_TENANCY: "0",
|
||||||
|
SELF_HOSTED: "1",
|
||||||
|
},
|
||||||
|
async () => {
|
||||||
|
await startup({ rerun: true })
|
||||||
|
const user = await users.getGlobalUserByEmail(EMAIL, {
|
||||||
|
cleanup: false,
|
||||||
|
})
|
||||||
|
expect(user).toBeDefined()
|
||||||
|
expect(user?.password).toBeDefined()
|
||||||
|
expect(await utils.compare(PASSWORD, user?.password!)).toEqual(true)
|
||||||
|
}
|
||||||
|
)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
})
|
|
@ -4,6 +4,7 @@ import {
|
||||||
CreateDatasourceResponse,
|
CreateDatasourceResponse,
|
||||||
UpdateDatasourceResponse,
|
UpdateDatasourceResponse,
|
||||||
UpdateDatasourceRequest,
|
UpdateDatasourceRequest,
|
||||||
|
QueryJson,
|
||||||
} from "@budibase/types"
|
} from "@budibase/types"
|
||||||
import { Expectations, TestAPI } from "./base"
|
import { Expectations, TestAPI } from "./base"
|
||||||
|
|
||||||
|
@ -45,4 +46,24 @@ export class DatasourceAPI extends TestAPI {
|
||||||
expectations,
|
expectations,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
delete = async (datasource: Datasource, expectations?: Expectations) => {
|
||||||
|
return await this._delete(
|
||||||
|
`/api/datasources/${datasource._id!}/${datasource._rev!}`,
|
||||||
|
{ expectations }
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
get = async (id: string, expectations?: Expectations) => {
|
||||||
|
return await this._get<Datasource>(`/api/datasources/${id}`, {
|
||||||
|
expectations,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
query = async (query: QueryJson, expectations?: Expectations) => {
|
||||||
|
return await this._post<any>(`/api/datasources/query`, {
|
||||||
|
body: query,
|
||||||
|
expectations,
|
||||||
|
})
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -6,10 +6,11 @@ import {
|
||||||
PreviewQueryResponse,
|
PreviewQueryResponse,
|
||||||
} from "@budibase/types"
|
} from "@budibase/types"
|
||||||
import { Expectations, TestAPI } from "./base"
|
import { Expectations, TestAPI } from "./base"
|
||||||
|
import { constants } from "@budibase/backend-core"
|
||||||
|
|
||||||
export class QueryAPI extends TestAPI {
|
export class QueryAPI extends TestAPI {
|
||||||
save = async (body: Query): Promise<Query> => {
|
save = async (body: Query, expectations?: Expectations): Promise<Query> => {
|
||||||
return await this._post<Query>(`/api/queries`, { body })
|
return await this._post<Query>(`/api/queries`, { body, expectations })
|
||||||
}
|
}
|
||||||
|
|
||||||
execute = async (
|
execute = async (
|
||||||
|
@ -26,9 +27,36 @@ export class QueryAPI extends TestAPI {
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
previewQuery = async (queryPreview: PreviewQueryRequest) => {
|
preview = async (
|
||||||
|
queryPreview: PreviewQueryRequest,
|
||||||
|
expectations?: Expectations
|
||||||
|
) => {
|
||||||
return await this._post<PreviewQueryResponse>(`/api/queries/preview`, {
|
return await this._post<PreviewQueryResponse>(`/api/queries/preview`, {
|
||||||
body: queryPreview,
|
body: queryPreview,
|
||||||
|
expectations,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
delete = async (query: Query, expectations?: Expectations) => {
|
||||||
|
return await this._delete(`/api/queries/${query._id!}/${query._rev!}`, {
|
||||||
|
expectations,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
get = async (queryId: string, expectations?: Expectations) => {
|
||||||
|
return await this._get<Query>(`/api/queries/${queryId}`, { expectations })
|
||||||
|
}
|
||||||
|
|
||||||
|
getProd = async (queryId: string, expectations?: Expectations) => {
|
||||||
|
return await this._get<Query>(`/api/queries/${queryId}`, {
|
||||||
|
expectations,
|
||||||
|
headers: {
|
||||||
|
[constants.Header.APP_ID]: this.config.getProdAppId(),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
fetch = async (expectations?: Expectations) => {
|
||||||
|
return await this._get<Query[]>(`/api/queries`, { expectations })
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,4 +1,6 @@
|
||||||
import {
|
import {
|
||||||
|
BulkImportRequest,
|
||||||
|
BulkImportResponse,
|
||||||
MigrateRequest,
|
MigrateRequest,
|
||||||
MigrateResponse,
|
MigrateResponse,
|
||||||
SaveTableRequest,
|
SaveTableRequest,
|
||||||
|
@ -39,4 +41,28 @@ export class TableAPI extends TestAPI {
|
||||||
expectations,
|
expectations,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
import = async (
|
||||||
|
tableId: string,
|
||||||
|
data: BulkImportRequest,
|
||||||
|
expectations?: Expectations
|
||||||
|
): Promise<BulkImportResponse> => {
|
||||||
|
return await this._post<BulkImportResponse>(
|
||||||
|
`/api/tables/${tableId}/import`,
|
||||||
|
{
|
||||||
|
body: data,
|
||||||
|
expectations,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
destroy = async (
|
||||||
|
tableId: string,
|
||||||
|
revId: string,
|
||||||
|
expectations?: Expectations
|
||||||
|
): Promise<void> => {
|
||||||
|
return await this._delete<void>(`/api/tables/${tableId}/${revId}`, {
|
||||||
|
expectations,
|
||||||
|
})
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -4,9 +4,9 @@ import {
|
||||||
ViewV2,
|
ViewV2,
|
||||||
SearchViewRowRequest,
|
SearchViewRowRequest,
|
||||||
PaginatedSearchRowResponse,
|
PaginatedSearchRowResponse,
|
||||||
|
ViewResponseEnriched,
|
||||||
} from "@budibase/types"
|
} from "@budibase/types"
|
||||||
import { Expectations, TestAPI } from "./base"
|
import { Expectations, TestAPI } from "./base"
|
||||||
import sdk from "../../../sdk"
|
|
||||||
|
|
||||||
export class ViewV2API extends TestAPI {
|
export class ViewV2API extends TestAPI {
|
||||||
create = async (
|
create = async (
|
||||||
|
@ -45,9 +45,8 @@ export class ViewV2API extends TestAPI {
|
||||||
}
|
}
|
||||||
|
|
||||||
get = async (viewId: string) => {
|
get = async (viewId: string) => {
|
||||||
return await this.config.doInContext(this.config.appId, () =>
|
return (await this._get<ViewResponseEnriched>(`/api/v2/views/${viewId}`))
|
||||||
sdk.views.get(viewId)
|
.data
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
search = async (
|
search = async (
|
||||||
|
|
|
@ -26,32 +26,56 @@ import {
|
||||||
WebhookActionType,
|
WebhookActionType,
|
||||||
} from "@budibase/types"
|
} from "@budibase/types"
|
||||||
import { LoopInput, LoopStepType } from "../../definitions/automations"
|
import { LoopInput, LoopStepType } from "../../definitions/automations"
|
||||||
|
import { merge } from "lodash"
|
||||||
|
import { generator } from "@budibase/backend-core/tests"
|
||||||
|
|
||||||
const { BUILTIN_ROLE_IDS } = roles
|
const { BUILTIN_ROLE_IDS } = roles
|
||||||
|
|
||||||
export function basicTable(): Table {
|
export function tableForDatasource(
|
||||||
return {
|
datasource?: Datasource,
|
||||||
name: "TestTable",
|
...extra: Partial<Table>[]
|
||||||
type: "table",
|
): Table {
|
||||||
sourceId: INTERNAL_TABLE_SOURCE_ID,
|
return merge(
|
||||||
sourceType: TableSourceType.INTERNAL,
|
{
|
||||||
schema: {
|
name: generator.guid(),
|
||||||
name: {
|
type: "table",
|
||||||
type: FieldType.STRING,
|
sourceType: datasource
|
||||||
name: "name",
|
? TableSourceType.EXTERNAL
|
||||||
constraints: {
|
: TableSourceType.INTERNAL,
|
||||||
type: "string",
|
sourceId: datasource ? datasource._id! : INTERNAL_TABLE_SOURCE_ID,
|
||||||
|
schema: {},
|
||||||
|
},
|
||||||
|
...extra
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
export function basicTable(
|
||||||
|
datasource?: Datasource,
|
||||||
|
...extra: Partial<Table>[]
|
||||||
|
): Table {
|
||||||
|
return tableForDatasource(
|
||||||
|
datasource,
|
||||||
|
{
|
||||||
|
name: "TestTable",
|
||||||
|
schema: {
|
||||||
|
name: {
|
||||||
|
type: FieldType.STRING,
|
||||||
|
name: "name",
|
||||||
|
constraints: {
|
||||||
|
type: "string",
|
||||||
|
},
|
||||||
},
|
},
|
||||||
},
|
description: {
|
||||||
description: {
|
type: FieldType.STRING,
|
||||||
type: FieldType.STRING,
|
name: "description",
|
||||||
name: "description",
|
constraints: {
|
||||||
constraints: {
|
type: "string",
|
||||||
type: "string",
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
...extra
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
export function basicView(tableId: string) {
|
export function basicView(tableId: string) {
|
||||||
|
|
|
@ -167,7 +167,7 @@ class QueryRunner {
|
||||||
this.hasRerun = true
|
this.hasRerun = true
|
||||||
}
|
}
|
||||||
|
|
||||||
await threadUtils.invalidateDynamicVariables(this.cachedVariables)
|
await threadUtils.invalidateCachedVariable(this.cachedVariables)
|
||||||
return this.execute()
|
return this.execute()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -254,7 +254,7 @@ class QueryRunner {
|
||||||
let { parameters } = this
|
let { parameters } = this
|
||||||
const queryId = variable.queryId,
|
const queryId = variable.queryId,
|
||||||
name = variable.name
|
name = variable.name
|
||||||
let value = await threadUtils.checkCacheForDynamicVariable(queryId, name)
|
let value = await threadUtils.getCachedVariable(queryId, name)
|
||||||
if (!value) {
|
if (!value) {
|
||||||
value = this.queryResponse[queryId]
|
value = this.queryResponse[queryId]
|
||||||
? this.queryResponse[queryId]
|
? this.queryResponse[queryId]
|
||||||
|
|
|
@ -5,7 +5,7 @@ import { redis, db as dbCore } from "@budibase/backend-core"
|
||||||
import * as jsRunner from "../jsRunner"
|
import * as jsRunner from "../jsRunner"
|
||||||
|
|
||||||
const VARIABLE_TTL_SECONDS = 3600
|
const VARIABLE_TTL_SECONDS = 3600
|
||||||
let client: any
|
let client: redis.Client | null = null
|
||||||
|
|
||||||
async function getClient() {
|
async function getClient() {
|
||||||
if (!client) {
|
if (!client) {
|
||||||
|
@ -36,23 +36,15 @@ export function threadSetup() {
|
||||||
db.init()
|
db.init()
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function checkCacheForDynamicVariable(
|
export async function getCachedVariable(queryId: string, variable: string) {
|
||||||
queryId: string,
|
return (await getClient()).get(makeVariableKey(queryId, variable))
|
||||||
variable: string
|
|
||||||
) {
|
|
||||||
const cache = await getClient()
|
|
||||||
return cache.get(makeVariableKey(queryId, variable))
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function invalidateDynamicVariables(cachedVars: QueryVariable[]) {
|
export async function invalidateCachedVariable(vars: QueryVariable[]) {
|
||||||
const cache = await getClient()
|
const cache = await getClient()
|
||||||
let promises = []
|
await Promise.all(
|
||||||
for (let variable of cachedVars) {
|
vars.map(v => cache.delete(makeVariableKey(v.queryId, v.name)))
|
||||||
promises.push(
|
)
|
||||||
cache.delete(makeVariableKey(variable.queryId, variable.name))
|
|
||||||
)
|
|
||||||
}
|
|
||||||
await Promise.all(promises)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function storeDynamicVariable(
|
export async function storeDynamicVariable(
|
||||||
|
@ -93,7 +85,7 @@ export default {
|
||||||
hasExtraData,
|
hasExtraData,
|
||||||
formatResponse,
|
formatResponse,
|
||||||
storeDynamicVariable,
|
storeDynamicVariable,
|
||||||
invalidateDynamicVariables,
|
invalidateCachedVariable,
|
||||||
checkCacheForDynamicVariable,
|
getCachedVariable,
|
||||||
threadSetup,
|
threadSetup,
|
||||||
}
|
}
|
||||||
|
|
|
@ -54,7 +54,7 @@ export function validate(rows: Rows, schema: TableSchema): ValidationResults {
|
||||||
type: columnType,
|
type: columnType,
|
||||||
subtype: columnSubtype,
|
subtype: columnSubtype,
|
||||||
autocolumn: isAutoColumn,
|
autocolumn: isAutoColumn,
|
||||||
} = schema[columnName]
|
} = schema[columnName] || {}
|
||||||
|
|
||||||
// If the column had an invalid value we don't want to override it
|
// If the column had an invalid value we don't want to override it
|
||||||
if (results.schemaValidation[columnName] === false) {
|
if (results.schemaValidation[columnName] === false) {
|
||||||
|
|
|
@ -3,16 +3,11 @@ import {
|
||||||
Row,
|
Row,
|
||||||
Table,
|
Table,
|
||||||
TableRequest,
|
TableRequest,
|
||||||
TableSchema,
|
|
||||||
View,
|
View,
|
||||||
ViewV2,
|
ViewV2Enriched,
|
||||||
} from "../../../documents"
|
} from "../../../documents"
|
||||||
|
|
||||||
interface ViewV2Response extends ViewV2 {
|
export type TableViewsResponse = { [key: string]: View | ViewV2Enriched }
|
||||||
schema: TableSchema
|
|
||||||
}
|
|
||||||
|
|
||||||
export type TableViewsResponse = { [key: string]: View | ViewV2Response }
|
|
||||||
|
|
||||||
export interface TableResponse extends Table {
|
export interface TableResponse extends Table {
|
||||||
views?: TableViewsResponse
|
views?: TableViewsResponse
|
||||||
|
|
|
@ -1,14 +1,13 @@
|
||||||
import { ViewV2, UIFieldMetadata } from "../../../documents"
|
import { ViewV2, ViewV2Enriched } from "../../../documents"
|
||||||
|
|
||||||
export interface ViewResponse {
|
export interface ViewResponse {
|
||||||
data: ViewV2
|
data: ViewV2
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface CreateViewRequest
|
export interface ViewResponseEnriched {
|
||||||
extends Omit<ViewV2, "version" | "id" | "schema"> {
|
data: ViewV2Enriched
|
||||||
schema?: Record<string, UIFieldMetadata>
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface UpdateViewRequest extends Omit<ViewV2, "schema"> {
|
export interface CreateViewRequest extends Omit<ViewV2, "version" | "id"> {}
|
||||||
schema?: Record<string, UIFieldMetadata>
|
|
||||||
}
|
export interface UpdateViewRequest extends ViewV2 {}
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
import { SearchFilter, SortOrder, SortType } from "../../api"
|
import { SearchFilter, SortOrder, SortType } from "../../api"
|
||||||
import { UIFieldMetadata } from "./table"
|
import { TableSchema, UIFieldMetadata } from "./table"
|
||||||
import { Document } from "../document"
|
import { Document } from "../document"
|
||||||
import { DBView } from "../../sdk"
|
import { DBView } from "../../sdk"
|
||||||
|
|
||||||
|
@ -48,6 +48,10 @@ export interface ViewV2 {
|
||||||
schema?: Record<string, UIFieldMetadata>
|
schema?: Record<string, UIFieldMetadata>
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export interface ViewV2Enriched extends ViewV2 {
|
||||||
|
schema?: TableSchema
|
||||||
|
}
|
||||||
|
|
||||||
export type ViewSchema = ViewCountOrSumSchema | ViewStatisticsSchema
|
export type ViewSchema = ViewCountOrSumSchema | ViewStatisticsSchema
|
||||||
|
|
||||||
export interface ViewCountOrSumSchema {
|
export interface ViewCountOrSumSchema {
|
||||||
|
|
|
@ -146,16 +146,12 @@ export const adminUser = async (
|
||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const finalUser = await userSdk.db.createAdminUser(
|
const finalUser = await userSdk.db.createAdminUser(email, tenantId, {
|
||||||
email,
|
|
||||||
tenantId,
|
|
||||||
password,
|
password,
|
||||||
{
|
ssoId,
|
||||||
ssoId,
|
hashPassword,
|
||||||
hashPassword,
|
requirePassword,
|
||||||
requirePassword,
|
})
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
// events
|
// events
|
||||||
let account: CloudAccount | undefined
|
let account: CloudAccount | undefined
|
||||||
|
|
173
yarn.lock
173
yarn.lock
|
@ -7261,37 +7261,7 @@ axios-retry@^3.1.9:
|
||||||
"@babel/runtime" "^7.15.4"
|
"@babel/runtime" "^7.15.4"
|
||||||
is-retry-allowed "^2.2.0"
|
is-retry-allowed "^2.2.0"
|
||||||
|
|
||||||
axios@0.24.0:
|
axios@0.24.0, axios@1.1.3, axios@1.6.3, axios@^0.21.1, axios@^0.21.4, axios@^0.26.0, axios@^1.0.0, axios@^1.1.3, axios@^1.5.0:
|
||||||
version "0.24.0"
|
|
||||||
resolved "https://registry.yarnpkg.com/axios/-/axios-0.24.0.tgz#804e6fa1e4b9c5288501dd9dff56a7a0940d20d6"
|
|
||||||
integrity sha512-Q6cWsys88HoPgAaFAVUb0WpPk0O8iTeisR9IMqy9G8AbO4NlpVknrnQS03zzF9PGAWgO3cgletO3VjV/P7VztA==
|
|
||||||
dependencies:
|
|
||||||
follow-redirects "^1.14.4"
|
|
||||||
|
|
||||||
axios@1.1.3:
|
|
||||||
version "1.1.3"
|
|
||||||
resolved "https://registry.yarnpkg.com/axios/-/axios-1.1.3.tgz#8274250dada2edf53814ed7db644b9c2866c1e35"
|
|
||||||
integrity sha512-00tXVRwKx/FZr/IDVFt4C+f9FYairX517WoGCL6dpOntqLkZofjhu43F/Xl44UOpqa+9sLFDrG/XAnFsUYgkDA==
|
|
||||||
dependencies:
|
|
||||||
follow-redirects "^1.15.0"
|
|
||||||
form-data "^4.0.0"
|
|
||||||
proxy-from-env "^1.1.0"
|
|
||||||
|
|
||||||
axios@^0.21.1, axios@^0.21.4:
|
|
||||||
version "0.21.4"
|
|
||||||
resolved "https://registry.yarnpkg.com/axios/-/axios-0.21.4.tgz#c67b90dc0568e5c1cf2b0b858c43ba28e2eda575"
|
|
||||||
integrity sha512-ut5vewkiu8jjGBdqpM44XxjuCjq9LAKeHVmoVfHVzy8eHgxxq8SbAVQNovDA8mVi05kP0Ea/n/UzcSHcTJQfNg==
|
|
||||||
dependencies:
|
|
||||||
follow-redirects "^1.14.0"
|
|
||||||
|
|
||||||
axios@^0.26.0:
|
|
||||||
version "0.26.1"
|
|
||||||
resolved "https://registry.yarnpkg.com/axios/-/axios-0.26.1.tgz#1ede41c51fcf51bbbd6fd43669caaa4f0495aaa9"
|
|
||||||
integrity sha512-fPwcX4EvnSHuInCMItEhAGnaSEXRBjtzh9fOtsE6E1G6p7vl7edEeZe11QHf18+6+9gR5PbKV/sGKNaD8YaMeA==
|
|
||||||
dependencies:
|
|
||||||
follow-redirects "^1.14.8"
|
|
||||||
|
|
||||||
axios@^1.0.0, axios@^1.1.3, axios@^1.5.0:
|
|
||||||
version "1.6.3"
|
version "1.6.3"
|
||||||
resolved "https://registry.yarnpkg.com/axios/-/axios-1.6.3.tgz#7f50f23b3aa246eff43c54834272346c396613f4"
|
resolved "https://registry.yarnpkg.com/axios/-/axios-1.6.3.tgz#7f50f23b3aa246eff43c54834272346c396613f4"
|
||||||
integrity sha512-fWyNdeawGam70jXSVlKl+SUNVcL6j6W79CuSIPfi6HnDUmSCH6gyUys/HrqHeA/wU0Az41rRgean494d0Jb+ww==
|
integrity sha512-fWyNdeawGam70jXSVlKl+SUNVcL6j6W79CuSIPfi6HnDUmSCH6gyUys/HrqHeA/wU0Az41rRgean494d0Jb+ww==
|
||||||
|
@ -11242,11 +11212,6 @@ fn.name@1.x.x:
|
||||||
resolved "https://registry.yarnpkg.com/fn.name/-/fn.name-1.1.0.tgz#26cad8017967aea8731bc42961d04a3d5988accc"
|
resolved "https://registry.yarnpkg.com/fn.name/-/fn.name-1.1.0.tgz#26cad8017967aea8731bc42961d04a3d5988accc"
|
||||||
integrity sha512-GRnmB5gPyJpAhTQdSZTSp9uaPSvl09KoYcMQtsB9rQoOmzs9dH6ffeccH+Z+cv6P68Hu5bC6JjRh4Ah/mHSNRw==
|
integrity sha512-GRnmB5gPyJpAhTQdSZTSp9uaPSvl09KoYcMQtsB9rQoOmzs9dH6ffeccH+Z+cv6P68Hu5bC6JjRh4Ah/mHSNRw==
|
||||||
|
|
||||||
follow-redirects@^1.14.0, follow-redirects@^1.14.4, follow-redirects@^1.14.8:
|
|
||||||
version "1.15.6"
|
|
||||||
resolved "https://registry.yarnpkg.com/follow-redirects/-/follow-redirects-1.15.6.tgz#7f815c0cda4249c74ff09e95ef97c23b5fd0399b"
|
|
||||||
integrity sha512-wWN62YITEaOpSK584EZXJafH1AGpO8RVgElfkuXbTOrPX4fIfOyEpW/CsiNd8JdYrAoOvafRTOEnvsO++qCqFA==
|
|
||||||
|
|
||||||
follow-redirects@^1.15.0:
|
follow-redirects@^1.15.0:
|
||||||
version "1.15.2"
|
version "1.15.2"
|
||||||
resolved "https://registry.yarnpkg.com/follow-redirects/-/follow-redirects-1.15.2.tgz#b460864144ba63f2681096f274c4e57026da2c13"
|
resolved "https://registry.yarnpkg.com/follow-redirects/-/follow-redirects-1.15.2.tgz#b460864144ba63f2681096f274c4e57026da2c13"
|
||||||
|
@ -12365,12 +12330,7 @@ http-assert@^1.3.0:
|
||||||
deep-equal "~1.0.1"
|
deep-equal "~1.0.1"
|
||||||
http-errors "~1.8.0"
|
http-errors "~1.8.0"
|
||||||
|
|
||||||
http-cache-semantics@3.8.1:
|
http-cache-semantics@3.8.1, http-cache-semantics@4.1.1, http-cache-semantics@^4.0.0, http-cache-semantics@^4.1.0, http-cache-semantics@^4.1.1:
|
||||||
version "3.8.1"
|
|
||||||
resolved "https://registry.yarnpkg.com/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz#39b0e16add9b605bf0a9ef3d9daaf4843b4cacd2"
|
|
||||||
integrity sha512-5ai2iksyV8ZXmnZhHH4rWPoxxistEexSi5936zIQ1bnNTW5VnA85B6P/VpXiRM017IgRvb2kKo1a//y+0wSp3w==
|
|
||||||
|
|
||||||
http-cache-semantics@^4.0.0, http-cache-semantics@^4.1.0, http-cache-semantics@^4.1.1:
|
|
||||||
version "4.1.1"
|
version "4.1.1"
|
||||||
resolved "https://registry.yarnpkg.com/http-cache-semantics/-/http-cache-semantics-4.1.1.tgz#abe02fcb2985460bf0323be664436ec3476a6d5a"
|
resolved "https://registry.yarnpkg.com/http-cache-semantics/-/http-cache-semantics-4.1.1.tgz#abe02fcb2985460bf0323be664436ec3476a6d5a"
|
||||||
integrity sha512-er295DKPVsV82j5kw1Gjt+ADA/XYHsajl82cGNQG2eyoPkvgUhX+nDIyelzhIWbbsXP39EHcI6l5tYs2FYqYXQ==
|
integrity sha512-er295DKPVsV82j5kw1Gjt+ADA/XYHsajl82cGNQG2eyoPkvgUhX+nDIyelzhIWbbsXP39EHcI6l5tYs2FYqYXQ==
|
||||||
|
@ -13315,6 +13275,11 @@ isobject@^3.0.1:
|
||||||
resolved "https://registry.yarnpkg.com/isobject/-/isobject-3.0.1.tgz#4e431e92b11a9731636aa1f9c8d1ccbcfdab78df"
|
resolved "https://registry.yarnpkg.com/isobject/-/isobject-3.0.1.tgz#4e431e92b11a9731636aa1f9c8d1ccbcfdab78df"
|
||||||
integrity sha512-WhB9zCku7EGTj/HQQRz5aUQEUeoQZH2bWcltRErOpymJ4boYE6wL9Tbr23krRPSZ+C5zqNSrSw+Cc7sZZ4b7vg==
|
integrity sha512-WhB9zCku7EGTj/HQQRz5aUQEUeoQZH2bWcltRErOpymJ4boYE6wL9Tbr23krRPSZ+C5zqNSrSw+Cc7sZZ4b7vg==
|
||||||
|
|
||||||
|
isobject@^4.0.0:
|
||||||
|
version "4.0.0"
|
||||||
|
resolved "https://registry.yarnpkg.com/isobject/-/isobject-4.0.0.tgz#3f1c9155e73b192022a80819bacd0343711697b0"
|
||||||
|
integrity sha512-S/2fF5wH8SJA/kmwr6HYhK/RI/OkhD84k8ntalo0iJjZikgq1XFvR5M8NPT1x5F7fBwCG3qHfnzeP/Vh/ZxCUA==
|
||||||
|
|
||||||
isolated-vm@^4.7.2:
|
isolated-vm@^4.7.2:
|
||||||
version "4.7.2"
|
version "4.7.2"
|
||||||
resolved "https://registry.yarnpkg.com/isolated-vm/-/isolated-vm-4.7.2.tgz#5670d5cce1d92004f9b825bec5b0b11fc7501b65"
|
resolved "https://registry.yarnpkg.com/isolated-vm/-/isolated-vm-4.7.2.tgz#5670d5cce1d92004f9b825bec5b0b11fc7501b65"
|
||||||
|
@ -15909,7 +15874,7 @@ msgpackr-extract@^3.0.2:
|
||||||
"@msgpackr-extract/msgpackr-extract-linux-x64" "3.0.2"
|
"@msgpackr-extract/msgpackr-extract-linux-x64" "3.0.2"
|
||||||
"@msgpackr-extract/msgpackr-extract-win32-x64" "3.0.2"
|
"@msgpackr-extract/msgpackr-extract-win32-x64" "3.0.2"
|
||||||
|
|
||||||
msgpackr@^1.5.2:
|
msgpackr@1.10.1, msgpackr@^1.5.2:
|
||||||
version "1.10.1"
|
version "1.10.1"
|
||||||
resolved "https://registry.yarnpkg.com/msgpackr/-/msgpackr-1.10.1.tgz#51953bb4ce4f3494f0c4af3f484f01cfbb306555"
|
resolved "https://registry.yarnpkg.com/msgpackr/-/msgpackr-1.10.1.tgz#51953bb4ce4f3494f0c4af3f484f01cfbb306555"
|
||||||
integrity sha512-r5VRLv9qouXuLiIBrLpl2d5ZvPt8svdQTl5/vMvE4nzDMyEX4sgW5yWhuBBj5UmgwOTWj8CIdSXn5sAfsHAWIQ==
|
integrity sha512-r5VRLv9qouXuLiIBrLpl2d5ZvPt8svdQTl5/vMvE4nzDMyEX4sgW5yWhuBBj5UmgwOTWj8CIdSXn5sAfsHAWIQ==
|
||||||
|
@ -16077,7 +16042,7 @@ nice-try@^1.0.4:
|
||||||
resolved "https://registry.yarnpkg.com/nice-try/-/nice-try-1.0.5.tgz#a3378a7696ce7d223e88fc9b764bd7ef1089e366"
|
resolved "https://registry.yarnpkg.com/nice-try/-/nice-try-1.0.5.tgz#a3378a7696ce7d223e88fc9b764bd7ef1089e366"
|
||||||
integrity sha512-1nh45deeb5olNY7eX82BkPO7SSxR5SSYJiPTrTdFUVYwAl8CKMA5N9PjTYkHiRjisVcxcQ1HXdLhx2qxxJzLNQ==
|
integrity sha512-1nh45deeb5olNY7eX82BkPO7SSxR5SSYJiPTrTdFUVYwAl8CKMA5N9PjTYkHiRjisVcxcQ1HXdLhx2qxxJzLNQ==
|
||||||
|
|
||||||
nock@^13.5.4:
|
nock@13.5.4, nock@^13.5.4:
|
||||||
version "13.5.4"
|
version "13.5.4"
|
||||||
resolved "https://registry.yarnpkg.com/nock/-/nock-13.5.4.tgz#8918f0addc70a63736170fef7106a9721e0dc479"
|
resolved "https://registry.yarnpkg.com/nock/-/nock-13.5.4.tgz#8918f0addc70a63736170fef7106a9721e0dc479"
|
||||||
integrity sha512-yAyTfdeNJGGBFxWdzSKCBYxs5FxLbCg5X5Q4ets974hcQzG1+qCxvIyOo4j2Ry6MUlhWVMX4OoYDefAIIwupjw==
|
integrity sha512-yAyTfdeNJGGBFxWdzSKCBYxs5FxLbCg5X5Q4ets974hcQzG1+qCxvIyOo4j2Ry6MUlhWVMX4OoYDefAIIwupjw==
|
||||||
|
@ -16113,25 +16078,13 @@ node-addon-api@^6.1.0:
|
||||||
resolved "https://registry.yarnpkg.com/node-addon-api/-/node-addon-api-6.1.0.tgz#ac8470034e58e67d0c6f1204a18ae6995d9c0d76"
|
resolved "https://registry.yarnpkg.com/node-addon-api/-/node-addon-api-6.1.0.tgz#ac8470034e58e67d0c6f1204a18ae6995d9c0d76"
|
||||||
integrity sha512-+eawOlIgy680F0kBzPUNFhMZGtJ1YmqM6l4+Crf4IkImjYrO/mqPwRMh352g23uIaQKFItcQ64I7KMaJxHgAVA==
|
integrity sha512-+eawOlIgy680F0kBzPUNFhMZGtJ1YmqM6l4+Crf4IkImjYrO/mqPwRMh352g23uIaQKFItcQ64I7KMaJxHgAVA==
|
||||||
|
|
||||||
node-fetch@2.6.0:
|
node-fetch@2.6.0, node-fetch@2.6.7, node-fetch@^2.6.0, node-fetch@^2.6.1, node-fetch@^2.6.7, node-fetch@^2.6.9, node-fetch@^2.7.0:
|
||||||
version "2.6.0"
|
|
||||||
resolved "https://registry.yarnpkg.com/node-fetch/-/node-fetch-2.6.0.tgz#e633456386d4aa55863f676a7ab0daa8fdecb0fd"
|
|
||||||
integrity sha512-8dG4H5ujfvFiqDmVu9fQ5bOHUC15JMjMY/Zumv26oOvvVJjM67KF8koCWIabKQ1GJIa9r2mMZscBq/TbdOcmNA==
|
|
||||||
|
|
||||||
node-fetch@2.6.7, node-fetch@^2.6.0, node-fetch@^2.6.1, node-fetch@^2.6.7:
|
|
||||||
version "2.6.7"
|
version "2.6.7"
|
||||||
resolved "https://registry.yarnpkg.com/node-fetch/-/node-fetch-2.6.7.tgz#24de9fba827e3b4ae44dc8b20256a379160052ad"
|
resolved "https://registry.yarnpkg.com/node-fetch/-/node-fetch-2.6.7.tgz#24de9fba827e3b4ae44dc8b20256a379160052ad"
|
||||||
integrity sha512-ZjMPFEfVx5j+y2yF35Kzx5sF7kDzxuDj6ziH4FFbOp87zKDZNx8yExJIb05OGF4Nlt9IHFIMBkRl41VdvcNdbQ==
|
integrity sha512-ZjMPFEfVx5j+y2yF35Kzx5sF7kDzxuDj6ziH4FFbOp87zKDZNx8yExJIb05OGF4Nlt9IHFIMBkRl41VdvcNdbQ==
|
||||||
dependencies:
|
dependencies:
|
||||||
whatwg-url "^5.0.0"
|
whatwg-url "^5.0.0"
|
||||||
|
|
||||||
node-fetch@^2.6.9, node-fetch@^2.7.0:
|
|
||||||
version "2.7.0"
|
|
||||||
resolved "https://registry.yarnpkg.com/node-fetch/-/node-fetch-2.7.0.tgz#d0f0fa6e3e2dc1d27efcd8ad99d550bda94d187d"
|
|
||||||
integrity sha512-c4FRfUm/dbcWZ7U+1Wq0AwCyFL+3nt2bEw05wfxSz+DWpWsitgmSgYmy2dQdWyKC1694ELPqMs/YzUSNozLt8A==
|
|
||||||
dependencies:
|
|
||||||
whatwg-url "^5.0.0"
|
|
||||||
|
|
||||||
node-forge@^1.3.1:
|
node-forge@^1.3.1:
|
||||||
version "1.3.1"
|
version "1.3.1"
|
||||||
resolved "https://registry.yarnpkg.com/node-forge/-/node-forge-1.3.1.tgz#be8da2af243b2417d5f646a770663a92b7e9ded3"
|
resolved "https://registry.yarnpkg.com/node-forge/-/node-forge-1.3.1.tgz#be8da2af243b2417d5f646a770663a92b7e9ded3"
|
||||||
|
@ -17276,15 +17229,7 @@ passport-strategy@1.x.x, passport-strategy@^1.0.0:
|
||||||
resolved "https://registry.yarnpkg.com/passport-strategy/-/passport-strategy-1.0.0.tgz#b5539aa8fc225a3d1ad179476ddf236b440f52e4"
|
resolved "https://registry.yarnpkg.com/passport-strategy/-/passport-strategy-1.0.0.tgz#b5539aa8fc225a3d1ad179476ddf236b440f52e4"
|
||||||
integrity sha512-CB97UUvDKJde2V0KDWWB3lyf6PC3FaZP7YxZ2G8OAtn9p4HI9j9JLP9qjOGZFvyl8uwNT8qM+hGnz/n16NI7oA==
|
integrity sha512-CB97UUvDKJde2V0KDWWB3lyf6PC3FaZP7YxZ2G8OAtn9p4HI9j9JLP9qjOGZFvyl8uwNT8qM+hGnz/n16NI7oA==
|
||||||
|
|
||||||
passport@^0.4.0:
|
passport@0.6.0, passport@^0.4.0, passport@^0.6.0:
|
||||||
version "0.4.1"
|
|
||||||
resolved "https://registry.yarnpkg.com/passport/-/passport-0.4.1.tgz#941446a21cb92fc688d97a0861c38ce9f738f270"
|
|
||||||
integrity sha512-IxXgZZs8d7uFSt3eqNjM9NQ3g3uQCW5avD8mRNoXV99Yig50vjuaez6dQK2qC0kVWPRTujxY0dWgGfT09adjYg==
|
|
||||||
dependencies:
|
|
||||||
passport-strategy "1.x.x"
|
|
||||||
pause "0.0.1"
|
|
||||||
|
|
||||||
passport@^0.6.0:
|
|
||||||
version "0.6.0"
|
version "0.6.0"
|
||||||
resolved "https://registry.yarnpkg.com/passport/-/passport-0.6.0.tgz#e869579fab465b5c0b291e841e6cc95c005fac9d"
|
resolved "https://registry.yarnpkg.com/passport/-/passport-0.6.0.tgz#e869579fab465b5c0b291e841e6cc95c005fac9d"
|
||||||
integrity sha512-0fe+p3ZnrWRW74fe8+SvCyf4a3Pb2/h7gFkQ8yTJpAO50gDzlfjZUZTO1k5Eg9kUct22OxHLqDZoKUWRHOh9ug==
|
integrity sha512-0fe+p3ZnrWRW74fe8+SvCyf4a3Pb2/h7gFkQ8yTJpAO50gDzlfjZUZTO1k5Eg9kUct22OxHLqDZoKUWRHOh9ug==
|
||||||
|
@ -18003,9 +17948,9 @@ postgres-interval@^1.1.0:
|
||||||
xtend "^4.0.0"
|
xtend "^4.0.0"
|
||||||
|
|
||||||
posthog-js@^1.116.6:
|
posthog-js@^1.116.6:
|
||||||
version "1.116.6"
|
version "1.117.0"
|
||||||
resolved "https://registry.yarnpkg.com/posthog-js/-/posthog-js-1.116.6.tgz#9a5c9f49230a76642f4c44d93b96710f886c2880"
|
resolved "https://registry.yarnpkg.com/posthog-js/-/posthog-js-1.117.0.tgz#59c3e520f6269f76ea82dce8760fbc33cdd7f48f"
|
||||||
integrity sha512-rvt8HxzJD4c2B/xsUa4jle8ApdqljeBI2Qqjp4XJMohQf18DXRyM6b96H5/UMs8jxYuZG14Er0h/kEIWeU6Fmw==
|
integrity sha512-+I8q5G9YG6r6wOLKPT+C+AV7MRhyVFJMTJS7dfwLmmT+mkVxQ5bfC59hBkJUObOR+YRn5jn2JT/sgIslU94EZg==
|
||||||
dependencies:
|
dependencies:
|
||||||
fflate "^0.4.8"
|
fflate "^0.4.8"
|
||||||
preact "^10.19.3"
|
preact "^10.19.3"
|
||||||
|
@ -18585,7 +18530,7 @@ pseudomap@^1.0.2:
|
||||||
resolved "https://registry.yarnpkg.com/pseudomap/-/pseudomap-1.0.2.tgz#f052a28da70e618917ef0a8ac34c1ae5a68286b3"
|
resolved "https://registry.yarnpkg.com/pseudomap/-/pseudomap-1.0.2.tgz#f052a28da70e618917ef0a8ac34c1ae5a68286b3"
|
||||||
integrity sha512-b/YwNhb8lk1Zz2+bXXpS/LK9OisiZZ1SNsSLxN1x2OXVEhW2Ckr/7mWE5vrC1ZTiJlD9g19jWszTmJsB+oEpFQ==
|
integrity sha512-b/YwNhb8lk1Zz2+bXXpS/LK9OisiZZ1SNsSLxN1x2OXVEhW2Ckr/7mWE5vrC1ZTiJlD9g19jWszTmJsB+oEpFQ==
|
||||||
|
|
||||||
psl@^1.1.28, psl@^1.1.33:
|
psl@^1.1.33:
|
||||||
version "1.9.0"
|
version "1.9.0"
|
||||||
resolved "https://registry.yarnpkg.com/psl/-/psl-1.9.0.tgz#d0df2a137f00794565fcaf3b2c00cd09f8d5a5a7"
|
resolved "https://registry.yarnpkg.com/psl/-/psl-1.9.0.tgz#d0df2a137f00794565fcaf3b2c00cd09f8d5a5a7"
|
||||||
integrity sha512-E/ZsdU4HLs/68gYzgGTkMicWTLPdAftJLfJFlLUAAKZGkStNU72sZjT66SnMDVOfOWY/YAoiD7Jxa9iHvngcag==
|
integrity sha512-E/ZsdU4HLs/68gYzgGTkMicWTLPdAftJLfJFlLUAAKZGkStNU72sZjT66SnMDVOfOWY/YAoiD7Jxa9iHvngcag==
|
||||||
|
@ -19602,11 +19547,6 @@ sax@1.2.1:
|
||||||
resolved "https://registry.yarnpkg.com/sax/-/sax-1.2.1.tgz#7b8e656190b228e81a66aea748480d828cd2d37a"
|
resolved "https://registry.yarnpkg.com/sax/-/sax-1.2.1.tgz#7b8e656190b228e81a66aea748480d828cd2d37a"
|
||||||
integrity sha512-8I2a3LovHTOpm7NV5yOyO8IHqgVsfK4+UuySrXU8YXkSRX7k6hCV9b3HrkKCr3nMpgj+0bmocaJJWpvp1oc7ZA==
|
integrity sha512-8I2a3LovHTOpm7NV5yOyO8IHqgVsfK4+UuySrXU8YXkSRX7k6hCV9b3HrkKCr3nMpgj+0bmocaJJWpvp1oc7ZA==
|
||||||
|
|
||||||
sax@>=0.1.1:
|
|
||||||
version "1.3.0"
|
|
||||||
resolved "https://registry.yarnpkg.com/sax/-/sax-1.3.0.tgz#a5dbe77db3be05c9d1ee7785dbd3ea9de51593d0"
|
|
||||||
integrity sha512-0s+oAmw9zLl1V1cS9BtZN7JAd0cW5e0QH4W3LWEK6a4LaLEA2OTpGYWDY+6XasBLtz6wkm3u1xRw95mRuJ59WA==
|
|
||||||
|
|
||||||
sax@>=0.6.0:
|
sax@>=0.6.0:
|
||||||
version "1.2.4"
|
version "1.2.4"
|
||||||
resolved "https://registry.yarnpkg.com/sax/-/sax-1.2.4.tgz#2816234e2378bddc4e5354fab5caa895df7100d9"
|
resolved "https://registry.yarnpkg.com/sax/-/sax-1.2.4.tgz#2816234e2378bddc4e5354fab5caa895df7100d9"
|
||||||
|
@ -19688,40 +19628,13 @@ semver-diff@^3.1.1:
|
||||||
dependencies:
|
dependencies:
|
||||||
semver "^6.3.0"
|
semver "^6.3.0"
|
||||||
|
|
||||||
"semver@2 || 3 || 4 || 5", semver@^5.5.0, semver@^5.6.0, semver@^5.7.1:
|
"semver@2 || 3 || 4 || 5", semver@7.5.3, semver@^5.5.0, semver@^5.6.0, semver@^5.7.1, semver@^6.0.0, semver@^6.1.1, semver@^6.1.2, semver@^6.2.0, semver@^6.3.0, semver@^6.3.1, semver@^7.0.0, semver@^7.1.1, semver@^7.1.2, semver@^7.3.2, semver@^7.3.4, semver@^7.3.5, semver@^7.3.7, semver@^7.3.8, semver@^7.5.3, semver@^7.5.4, semver@~2.3.1, semver@~7.0.0:
|
||||||
version "5.7.2"
|
|
||||||
resolved "https://registry.yarnpkg.com/semver/-/semver-5.7.2.tgz#48d55db737c3287cd4835e17fa13feace1c41ef8"
|
|
||||||
integrity sha512-cBznnQ9KjJqU67B52RMC65CMarK2600WFnbkcaiwWq3xy/5haFJlshgnpjovMVJ+Hff49d8GEn0b87C5pDQ10g==
|
|
||||||
|
|
||||||
semver@7.5.3, semver@^7.0.0, semver@^7.1.1, semver@^7.1.2, semver@^7.3.2, semver@^7.3.4, semver@^7.3.5, semver@^7.3.7, semver@^7.3.8, semver@^7.5.3:
|
|
||||||
version "7.5.3"
|
version "7.5.3"
|
||||||
resolved "https://registry.yarnpkg.com/semver/-/semver-7.5.3.tgz#161ce8c2c6b4b3bdca6caadc9fa3317a4c4fe88e"
|
resolved "https://registry.yarnpkg.com/semver/-/semver-7.5.3.tgz#161ce8c2c6b4b3bdca6caadc9fa3317a4c4fe88e"
|
||||||
integrity sha512-QBlUtyVk/5EeHbi7X0fw6liDZc7BBmEaSYn01fMU1OUYbf6GPsbTtd8WmnqbI20SeycoHSeiybkE/q1Q+qlThQ==
|
integrity sha512-QBlUtyVk/5EeHbi7X0fw6liDZc7BBmEaSYn01fMU1OUYbf6GPsbTtd8WmnqbI20SeycoHSeiybkE/q1Q+qlThQ==
|
||||||
dependencies:
|
dependencies:
|
||||||
lru-cache "^6.0.0"
|
lru-cache "^6.0.0"
|
||||||
|
|
||||||
semver@^6.0.0, semver@^6.1.1, semver@^6.1.2, semver@^6.2.0, semver@^6.3.0, semver@^6.3.1:
|
|
||||||
version "6.3.1"
|
|
||||||
resolved "https://registry.yarnpkg.com/semver/-/semver-6.3.1.tgz#556d2ef8689146e46dcea4bfdd095f3434dffcb4"
|
|
||||||
integrity sha512-BR7VvDCVHO+q2xBEWskxS6DJE1qRnb7DxzUrogb71CWoSficBxYsiAGd+Kl0mmq/MprG9yArRkyrQxTO6XjMzA==
|
|
||||||
|
|
||||||
semver@^7.5.4:
|
|
||||||
version "7.6.0"
|
|
||||||
resolved "https://registry.yarnpkg.com/semver/-/semver-7.6.0.tgz#1a46a4db4bffcccd97b743b5005c8325f23d4e2d"
|
|
||||||
integrity sha512-EnwXhrlwXMk9gKu5/flx5sv/an57AkRplG3hTK68W7FRDN+k+OWBj65M7719OkA82XLBxrcX0KSHj+X5COhOVg==
|
|
||||||
dependencies:
|
|
||||||
lru-cache "^6.0.0"
|
|
||||||
|
|
||||||
semver@~2.3.1:
|
|
||||||
version "2.3.2"
|
|
||||||
resolved "https://registry.yarnpkg.com/semver/-/semver-2.3.2.tgz#b9848f25d6cf36333073ec9ef8856d42f1233e52"
|
|
||||||
integrity sha512-abLdIKCosKfpnmhS52NCTjO4RiLspDfsn37prjzGrp9im5DPJOgh82Os92vtwGh6XdQryKI/7SREZnV+aqiXrA==
|
|
||||||
|
|
||||||
semver@~7.0.0:
|
|
||||||
version "7.0.0"
|
|
||||||
resolved "https://registry.yarnpkg.com/semver/-/semver-7.0.0.tgz#5f3ca35761e47e05b206c6daff2cf814f0316b8e"
|
|
||||||
integrity sha512-+GB6zVA9LWh6zovYQLALHwv5rb2PHGlJi3lfiqIHxR0uuwCgefcOJc59v9fv1w8GbStwxuuqqAjI9NMAOOgq1A==
|
|
||||||
|
|
||||||
seq-queue@^0.0.5:
|
seq-queue@^0.0.5:
|
||||||
version "0.0.5"
|
version "0.0.5"
|
||||||
resolved "https://registry.yarnpkg.com/seq-queue/-/seq-queue-0.0.5.tgz#d56812e1c017a6e4e7c3e3a37a1da6d78dd3c93e"
|
resolved "https://registry.yarnpkg.com/seq-queue/-/seq-queue-0.0.5.tgz#d56812e1c017a6e4e7c3e3a37a1da6d78dd3c93e"
|
||||||
|
@ -21307,7 +21220,7 @@ touch@^3.1.0:
|
||||||
dependencies:
|
dependencies:
|
||||||
nopt "~1.0.10"
|
nopt "~1.0.10"
|
||||||
|
|
||||||
"tough-cookie@^2.3.3 || ^3.0.1 || ^4.0.0", tough-cookie@^4.0.0, tough-cookie@^4.1.2:
|
tough-cookie@4.1.3, "tough-cookie@^2.3.3 || ^3.0.1 || ^4.0.0", tough-cookie@^4.0.0, tough-cookie@^4.1.2, tough-cookie@~2.5.0:
|
||||||
version "4.1.3"
|
version "4.1.3"
|
||||||
resolved "https://registry.yarnpkg.com/tough-cookie/-/tough-cookie-4.1.3.tgz#97b9adb0728b42280aa3d814b6b999b2ff0318bf"
|
resolved "https://registry.yarnpkg.com/tough-cookie/-/tough-cookie-4.1.3.tgz#97b9adb0728b42280aa3d814b6b999b2ff0318bf"
|
||||||
integrity sha512-aX/y5pVRkfRnfmuX+OdbSdXvPe6ieKX/G2s7e98f4poJHnqH3281gDPm/metm6E/WRamfx7WC4HUqkWHfQHprw==
|
integrity sha512-aX/y5pVRkfRnfmuX+OdbSdXvPe6ieKX/G2s7e98f4poJHnqH3281gDPm/metm6E/WRamfx7WC4HUqkWHfQHprw==
|
||||||
|
@ -21317,14 +21230,6 @@ touch@^3.1.0:
|
||||||
universalify "^0.2.0"
|
universalify "^0.2.0"
|
||||||
url-parse "^1.5.3"
|
url-parse "^1.5.3"
|
||||||
|
|
||||||
tough-cookie@~2.5.0:
|
|
||||||
version "2.5.0"
|
|
||||||
resolved "https://registry.yarnpkg.com/tough-cookie/-/tough-cookie-2.5.0.tgz#cd9fb2a0aa1d5a12b473bd9fb96fa3dcff65ade2"
|
|
||||||
integrity sha512-nlLsUzgm1kfLXSXfRZMc1KLAugd4hqJHDTvc2hDIwS3mZAfMEuMbc03SujMF+GEcpaX/qboeycw6iO8JwVv2+g==
|
|
||||||
dependencies:
|
|
||||||
psl "^1.1.28"
|
|
||||||
punycode "^2.1.1"
|
|
||||||
|
|
||||||
tr46@^2.1.0:
|
tr46@^2.1.0:
|
||||||
version "2.1.0"
|
version "2.1.0"
|
||||||
resolved "https://registry.yarnpkg.com/tr46/-/tr46-2.1.0.tgz#fa87aa81ca5d5941da8cbf1f9b749dc969a4e240"
|
resolved "https://registry.yarnpkg.com/tr46/-/tr46-2.1.0.tgz#fa87aa81ca5d5941da8cbf1f9b749dc969a4e240"
|
||||||
|
@ -21801,6 +21706,14 @@ unpipe@1.0.0:
|
||||||
resolved "https://registry.yarnpkg.com/unpipe/-/unpipe-1.0.0.tgz#b2bf4ee8514aae6165b4817829d21b2ef49904ec"
|
resolved "https://registry.yarnpkg.com/unpipe/-/unpipe-1.0.0.tgz#b2bf4ee8514aae6165b4817829d21b2ef49904ec"
|
||||||
integrity sha512-pjy2bYhSsufwWlKwPc+l3cN7+wuJlK6uz0YdJEOlQDbl6jo/YlPi4mb8agUkVC8BF7V8NuzeyPNqRksA3hztKQ==
|
integrity sha512-pjy2bYhSsufwWlKwPc+l3cN7+wuJlK6uz0YdJEOlQDbl6jo/YlPi4mb8agUkVC8BF7V8NuzeyPNqRksA3hztKQ==
|
||||||
|
|
||||||
|
unset-value@2.0.1:
|
||||||
|
version "2.0.1"
|
||||||
|
resolved "https://registry.yarnpkg.com/unset-value/-/unset-value-2.0.1.tgz#57bed0c22d26f28d69acde5df9a11b77c74d2df3"
|
||||||
|
integrity sha512-2hvrBfjUE00PkqN+q0XP6yRAOGrR06uSiUoIQGZkc7GxvQ9H7v8quUPNtZjMg4uux69i8HWpIjLPUKwCuRGyNg==
|
||||||
|
dependencies:
|
||||||
|
has-value "^2.0.2"
|
||||||
|
isobject "^4.0.0"
|
||||||
|
|
||||||
untildify@^4.0.0:
|
untildify@^4.0.0:
|
||||||
version "4.0.0"
|
version "4.0.0"
|
||||||
resolved "https://registry.yarnpkg.com/untildify/-/untildify-4.0.0.tgz#2bc947b953652487e4600949fb091e3ae8cd919b"
|
resolved "https://registry.yarnpkg.com/untildify/-/untildify-4.0.0.tgz#2bc947b953652487e4600949fb091e3ae8cd919b"
|
||||||
|
@ -22571,33 +22484,10 @@ xml-parse-from-string@^1.0.0:
|
||||||
resolved "https://registry.yarnpkg.com/xml-parse-from-string/-/xml-parse-from-string-1.0.1.tgz#a9029e929d3dbcded169f3c6e28238d95a5d5a28"
|
resolved "https://registry.yarnpkg.com/xml-parse-from-string/-/xml-parse-from-string-1.0.1.tgz#a9029e929d3dbcded169f3c6e28238d95a5d5a28"
|
||||||
integrity sha512-ErcKwJTF54uRzzNMXq2X5sMIy88zJvfN2DmdoQvy7PAFJ+tPRU6ydWuOKNMyfmOjdyBQTFREi60s0Y0SyI0G0g==
|
integrity sha512-ErcKwJTF54uRzzNMXq2X5sMIy88zJvfN2DmdoQvy7PAFJ+tPRU6ydWuOKNMyfmOjdyBQTFREi60s0Y0SyI0G0g==
|
||||||
|
|
||||||
xml2js@0.1.x:
|
xml2js@0.1.x, xml2js@0.4.19, xml2js@0.5.0, xml2js@0.6.2, xml2js@^0.4.19, xml2js@^0.4.5:
|
||||||
version "0.1.14"
|
version "0.6.2"
|
||||||
resolved "https://registry.yarnpkg.com/xml2js/-/xml2js-0.1.14.tgz#5274e67f5a64c5f92974cd85139e0332adc6b90c"
|
resolved "https://registry.yarnpkg.com/xml2js/-/xml2js-0.6.2.tgz#dd0b630083aa09c161e25a4d0901e2b2a929b499"
|
||||||
integrity sha512-pbdws4PPPNc1HPluSUKamY4GWMk592K7qwcj6BExbVOhhubub8+pMda/ql68b6L3luZs/OGjGSB5goV7SnmgnA==
|
integrity sha512-T4rieHaC1EXcES0Kxxj4JWgaUQHDk+qwHcYOCFHfiwKz7tOVPLq7Hjq9dM1WCMhylqMEfP7hMcOIChvotiZegA==
|
||||||
dependencies:
|
|
||||||
sax ">=0.1.1"
|
|
||||||
|
|
||||||
xml2js@0.4.19:
|
|
||||||
version "0.4.19"
|
|
||||||
resolved "https://registry.yarnpkg.com/xml2js/-/xml2js-0.4.19.tgz#686c20f213209e94abf0d1bcf1efaa291c7827a7"
|
|
||||||
integrity sha512-esZnJZJOiJR9wWKMyuvSE1y6Dq5LCuJanqhxslH2bxM6duahNZ+HMpCLhBQGZkbX6xRf8x1Y2eJlgt2q3qo49Q==
|
|
||||||
dependencies:
|
|
||||||
sax ">=0.6.0"
|
|
||||||
xmlbuilder "~9.0.1"
|
|
||||||
|
|
||||||
xml2js@0.5.0:
|
|
||||||
version "0.5.0"
|
|
||||||
resolved "https://registry.yarnpkg.com/xml2js/-/xml2js-0.5.0.tgz#d9440631fbb2ed800203fad106f2724f62c493b7"
|
|
||||||
integrity sha512-drPFnkQJik/O+uPKpqSgr22mpuFHqKdbS835iAQrUC73L2F5WkboIRd63ai/2Yg6I1jzifPFKH2NTK+cfglkIA==
|
|
||||||
dependencies:
|
|
||||||
sax ">=0.6.0"
|
|
||||||
xmlbuilder "~11.0.0"
|
|
||||||
|
|
||||||
xml2js@^0.4.19, xml2js@^0.4.5:
|
|
||||||
version "0.4.23"
|
|
||||||
resolved "https://registry.yarnpkg.com/xml2js/-/xml2js-0.4.23.tgz#a0c69516752421eb2ac758ee4d4ccf58843eac66"
|
|
||||||
integrity sha512-ySPiMjM0+pLDftHgXY4By0uswI3SPKLDw/i3UXbnO8M/p28zqexCUoPmQFrYD+/1BzhGJSs2i1ERWKJAtiLrug==
|
|
||||||
dependencies:
|
dependencies:
|
||||||
sax ">=0.6.0"
|
sax ">=0.6.0"
|
||||||
xmlbuilder "~11.0.0"
|
xmlbuilder "~11.0.0"
|
||||||
|
@ -22607,11 +22497,6 @@ xmlbuilder@~11.0.0:
|
||||||
resolved "https://registry.yarnpkg.com/xmlbuilder/-/xmlbuilder-11.0.1.tgz#be9bae1c8a046e76b31127726347d0ad7002beb3"
|
resolved "https://registry.yarnpkg.com/xmlbuilder/-/xmlbuilder-11.0.1.tgz#be9bae1c8a046e76b31127726347d0ad7002beb3"
|
||||||
integrity sha512-fDlsI/kFEx7gLvbecc0/ohLG50fugQp8ryHzMTuW9vSa1GJ0XYWKnhsUx7oie3G98+r56aTQIUB4kht42R3JvA==
|
integrity sha512-fDlsI/kFEx7gLvbecc0/ohLG50fugQp8ryHzMTuW9vSa1GJ0XYWKnhsUx7oie3G98+r56aTQIUB4kht42R3JvA==
|
||||||
|
|
||||||
xmlbuilder@~9.0.1:
|
|
||||||
version "9.0.7"
|
|
||||||
resolved "https://registry.yarnpkg.com/xmlbuilder/-/xmlbuilder-9.0.7.tgz#132ee63d2ec5565c557e20f4c22df9aca686b10d"
|
|
||||||
integrity sha512-7YXTQc3P2l9+0rjaUbLwMKRhtmwg1M1eDf6nag7urC7pIPYLD9W/jmzQ4ptRSUbodw5S0jfoGTflLemQibSpeQ==
|
|
||||||
|
|
||||||
xmlchars@^2.2.0:
|
xmlchars@^2.2.0:
|
||||||
version "2.2.0"
|
version "2.2.0"
|
||||||
resolved "https://registry.yarnpkg.com/xmlchars/-/xmlchars-2.2.0.tgz#060fe1bcb7f9c76fe2a17db86a9bc3ab894210cb"
|
resolved "https://registry.yarnpkg.com/xmlchars/-/xmlchars-2.2.0.tgz#060fe1bcb7f9c76fe2a17db86a9bc3ab894210cb"
|
||||||
|
|
Loading…
Reference in New Issue