Merge branch 'master' into cheeks-fixes

This commit is contained in:
Andrew Kingston 2024-12-18 15:41:04 +00:00 committed by GitHub
commit adc0d7869b
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
17 changed files with 297 additions and 149 deletions

View File

@ -70,6 +70,10 @@ export function encodeTableId(tableId: string) {
}
}
export function encodeViewId(viewId: string) {
return encodeURIComponent(viewId)
}
export function breakExternalTableId(tableId: string) {
const parts = tableId.split(DOUBLE_SEPARATOR)
let datasourceId = parts.shift()

View File

@ -7,7 +7,7 @@
"build": "routify -b && NODE_OPTIONS=\"--max_old_space_size=4096\" vite build --emptyOutDir",
"start": "routify -c rollup",
"dev": "routify -c dev:vite",
"dev:vite": "vite --host 0.0.0.0",
"dev:vite": "vite --host 0.0.0.0 --mode=dev",
"rollup": "rollup -c -w",
"test": "vitest run",
"test:watch": "vitest",

View File

@ -33,7 +33,7 @@
...datasource,
name,
}
await datasources.update({
await datasources.save({
datasource: updatedDatasource,
integration: integrationForDatasource(get(integrations), datasource),
})

View File

@ -41,7 +41,7 @@
get(integrations),
datasource
)
await datasources.update({ datasource, integration })
await datasources.save({ datasource, integration })
await afterSave({ datasource, action })
} catch (err) {

View File

@ -176,7 +176,7 @@
notifications.success(`Request saved successfully`)
if (dynamicVariables) {
datasource.config.dynamicVariables = rebuildVariables(saveId)
datasource = await datasources.update({
datasource = await datasources.save({
integration: integrationInfo,
datasource,
})

View File

@ -13,7 +13,7 @@
async function saveDatasource({ config, name }) {
try {
await datasources.update({
await datasources.save({
integration,
datasource: { ...datasource, config, name },
})

View File

@ -16,7 +16,7 @@
get(integrations),
updatedDatasource
)
await datasources.update({ datasource: updatedDatasource, integration })
await datasources.save({ datasource: updatedDatasource, integration })
notifications.success(
`Datasource ${updatedDatasource.name} updated successfully`
)

View File

@ -1,4 +1,4 @@
import { writable, derived, get } from "svelte/store"
import { derived, get } from "svelte/store"
import {
IntegrationTypes,
DEFAULT_BB_DATASOURCE_ID,
@ -17,6 +17,7 @@ import {
} from "@budibase/types"
// @ts-ignore
import { TableNames } from "constants"
import BudiStore from "stores/BudiStore"
// when building the internal DS - seems to represent it slightly differently to the backend typing of a DS
interface InternalDatasource extends Omit<Datasource, "entities"> {
@ -41,24 +42,39 @@ class TableImportError extends Error {
}
}
interface DatasourceStore {
interface BuilderDatasourceStore {
list: Datasource[]
selectedDatasourceId: null | string
}
export function createDatasourcesStore() {
const store = writable<DatasourceStore>({
interface DerivedDatasourceStore extends Omit<BuilderDatasourceStore, "list"> {
list: (Datasource | InternalDatasource)[]
selected?: Datasource | InternalDatasource
hasDefaultData: boolean
hasData: boolean
}
export class DatasourceStore extends BudiStore<DerivedDatasourceStore> {
constructor() {
super({
list: [],
selectedDatasourceId: null,
hasDefaultData: false,
hasData: false,
})
const derivedStore = derived([store, tables], ([$store, $tables]) => {
const derivedStore = derived<
[DatasourceStore, BudiStore<any>],
DerivedDatasourceStore
>([this, tables as any], ([$store, $tables]) => {
// Set the internal datasource entities from the table list, which we're
// able to keep updated unlike the egress generated definition of the
// internal datasource
let internalDS: Datasource | InternalDatasource | undefined =
$store.list?.find(ds => ds._id === BUDIBASE_INTERNAL_DB_ID)
let otherDS = $store.list?.filter(ds => ds._id !== BUDIBASE_INTERNAL_DB_ID)
let otherDS = $store.list?.filter(
ds => ds._id !== BUDIBASE_INTERNAL_DB_ID
)
if (internalDS) {
const tables: Table[] = $tables.list?.filter((table: Table) => {
return (
@ -89,54 +105,68 @@ export function createDatasourcesStore() {
}
})
const fetch = async () => {
this.fetch = this.fetch.bind(this)
this.init = this.fetch.bind(this)
this.select = this.select.bind(this)
this.updateSchema = this.updateSchema.bind(this)
this.create = this.create.bind(this)
this.delete = this.deleteDatasource.bind(this)
this.save = this.save.bind(this)
this.replaceDatasource = this.replaceDatasource.bind(this)
this.getTableNames = this.getTableNames.bind(this)
this.subscribe = derivedStore.subscribe
}
async fetch() {
const datasources = await API.getDatasources()
store.update(state => ({
this.store.update(state => ({
...state,
list: datasources,
}))
}
const select = (id: string) => {
store.update(state => ({
async init() {
return this.fetch()
}
select(id: string) {
this.store.update(state => ({
...state,
selectedDatasourceId: id,
}))
}
const updateDatasource = (
private updateDatasourceInStore(
response: { datasource: Datasource; errors?: Record<string, string> },
{ ignoreErrors }: { ignoreErrors?: boolean } = {}
) => {
) {
const { datasource, errors } = response
if (!ignoreErrors && errors && Object.keys(errors).length > 0) {
throw new TableImportError(errors)
}
replaceDatasource(datasource._id!, datasource)
select(datasource._id!)
this.replaceDatasource(datasource._id!, datasource)
this.select(datasource._id!)
return datasource
}
const updateSchema = async (
datasource: Datasource,
tablesFilter: string[]
) => {
async updateSchema(datasource: Datasource, tablesFilter: string[]) {
const response = await API.buildDatasourceSchema(
datasource?._id!,
tablesFilter
)
updateDatasource(response)
this.updateDatasourceInStore(response)
}
const sourceCount = (source: string) => {
return get(store).list.filter(datasource => datasource.source === source)
.length
sourceCount(source: string) {
return get(this.store).list.filter(
datasource => datasource.source === source
).length
}
const checkDatasourceValidity = async (
async checkDatasourceValidity(
integration: Integration,
datasource: Datasource
): Promise<{ valid: boolean; error?: string }> => {
): Promise<{ valid: boolean; error?: string }> {
if (integration.features?.[DatasourceFeature.CONNECTION_CHECKING]) {
const { connected, error } = await API.validateDatasource(datasource)
if (connected) {
@ -148,14 +178,14 @@ export function createDatasourcesStore() {
return { valid: true }
}
const create = async ({
async create({
integration,
config,
}: {
integration: UIIntegration
config: Record<string, any>
}) => {
const count = sourceCount(integration.name)
}) {
const count = this.sourceCount(integration.name)
const nameModifier = count === 0 ? "" : ` ${count + 1}`
const datasource: Datasource = {
@ -167,7 +197,7 @@ export function createDatasourcesStore() {
isSQL: integration.isSQL,
}
const { valid, error } = await checkDatasourceValidity(
const { valid, error } = await this.checkDatasourceValidity(
integration,
datasource
)
@ -180,41 +210,45 @@ export function createDatasourcesStore() {
fetchSchema: integration.plus,
})
return updateDatasource(response, { ignoreErrors: true })
return this.updateDatasourceInStore(response, { ignoreErrors: true })
}
const update = async ({
async save({
integration,
datasource,
}: {
integration: Integration
datasource: Datasource
}) => {
if (await checkDatasourceValidity(integration, datasource)) {
}) {
if (await this.checkDatasourceValidity(integration, datasource)) {
throw new Error("Unable to connect")
}
const response = await API.updateDatasource(datasource)
return updateDatasource(response)
return this.updateDatasourceInStore(response)
}
const deleteDatasource = async (datasource: Datasource) => {
async deleteDatasource(datasource: Datasource) {
if (!datasource?._id || !datasource?._rev) {
return
}
await API.deleteDatasource(datasource._id, datasource._rev)
replaceDatasource(datasource._id)
this.replaceDatasource(datasource._id)
}
const replaceDatasource = (datasourceId: string, datasource?: Datasource) => {
async delete(datasource: Datasource) {
return this.deleteDatasource(datasource)
}
replaceDatasource(datasourceId: string, datasource?: Datasource) {
if (!datasourceId) {
return
}
// Handle deletion
if (!datasource) {
store.update(state => ({
this.store.update(state => ({
...state,
list: state.list.filter(x => x._id !== datasourceId),
}))
@ -224,9 +258,9 @@ export function createDatasourcesStore() {
}
// Add new datasource
const index = get(store).list.findIndex(x => x._id === datasource._id)
const index = get(this.store).list.findIndex(x => x._id === datasource._id)
if (index === -1) {
store.update(state => ({
this.store.update(state => ({
...state,
list: [...state.list, datasource],
}))
@ -238,30 +272,21 @@ export function createDatasourcesStore() {
// Update existing datasource
else if (datasource) {
store.update(state => {
this.store.update(state => {
state.list[index] = datasource
return state
})
}
}
const getTableNames = async (datasource: Datasource) => {
async getTableNames(datasource: Datasource) {
const info = await API.fetchInfoForDatasource(datasource)
return info.tableNames || []
}
return {
subscribe: derivedStore.subscribe,
fetch,
init: fetch,
select,
updateSchema,
create,
update,
delete: deleteDatasource,
replaceDatasource,
getTableNames,
}
// subscribe() {
// return this.derivedStore.subscribe()
// }
}
export const datasources = createDatasourcesStore()
export const datasources = new DatasourceStore()

View File

@ -3,6 +3,7 @@ import { derived } from "svelte/store"
import { DatasourceTypes } from "constants/backend"
import { UIIntegration, Integration } from "@budibase/types"
import BudiStore from "stores/BudiStore"
const getIntegrationOrder = (type: string | undefined) => {
// if type is not known, sort to end
@ -18,8 +19,11 @@ const getIntegrationOrder = (type: string | undefined) => {
return type.charCodeAt(0) + 4
}
export const createSortedIntegrationsStore = () => {
return derived<typeof integrations, UIIntegration[]>(
export class SortedIntegrationStore extends BudiStore<UIIntegration[]> {
constructor() {
super([])
const derivedStore = derived<typeof integrations, UIIntegration[]>(
integrations,
$integrations => {
const entries: [string, Integration][] = Object.entries($integrations)
@ -41,6 +45,9 @@ export const createSortedIntegrationsStore = () => {
})
}
)
this.subscribe = derivedStore.subscribe
}
}
export const sortedIntegrations = createSortedIntegrationsStore()
export const sortedIntegrations = new SortedIntegrationStore()

View File

@ -1,12 +1,14 @@
import { it, expect, describe, beforeEach, vi } from "vitest"
import { createSortedIntegrationsStore } from "stores/builder/sortedIntegrations"
import { SortedIntegrationStore } from "stores/builder/sortedIntegrations"
import { DatasourceTypes } from "constants/backend"
import { derived } from "svelte/store"
import { integrations } from "stores/builder/integrations"
vi.mock("svelte/store", () => ({
derived: vi.fn(),
derived: vi.fn(() => ({
subscribe: vi.fn(),
})),
writable: vi.fn(() => ({
subscribe: vi.fn(),
})),
@ -14,6 +16,8 @@ vi.mock("svelte/store", () => ({
vi.mock("stores/builder/integrations", () => ({ integrations: vi.fn() }))
const mockedDerived = vi.mocked(derived)
const inputA = {
nonRelationalA: {
friendlyName: "non-relational A",
@ -104,25 +108,28 @@ const expectedOutput = [
]
describe("sorted integrations store", () => {
beforeEach(ctx => {
interface LocalContext {
returnedStore: SortedIntegrationStore
derivedCallback: any
}
beforeEach<LocalContext>(ctx => {
vi.clearAllMocks()
ctx.returnedStore = createSortedIntegrationsStore()
ctx.derivedCallback = derived.mock.calls[0][1]
ctx.returnedStore = new SortedIntegrationStore()
ctx.derivedCallback = mockedDerived.mock.calls[0]?.[1]
})
it("calls derived with the correct parameters", () => {
expect(derived).toHaveBeenCalledTimes(1)
expect(derived).toHaveBeenCalledWith(integrations, expect.toBeFunc())
expect(mockedDerived).toHaveBeenCalledTimes(1)
expect(mockedDerived).toHaveBeenCalledWith(
integrations,
expect.any(Function)
)
})
describe("derived callback", () => {
it("When no integrations are loaded", ctx => {
expect(ctx.derivedCallback({})).toEqual([])
})
it("When integrations are present", ctx => {
it<LocalContext>("When integrations are present", ctx => {
expect(ctx.derivedCallback(inputA)).toEqual(expectedOutput)
expect(ctx.derivedCallback(inputB)).toEqual(expectedOutput)
})

View File

@ -102,9 +102,8 @@
lastSearchId = Math.random()
searching = true
const thisSearchId = lastSearchId
const results = await searchFunction({
const results = await searchFunction(schema.tableId, {
paginate: false,
tableId: schema.tableId,
limit: 20,
query: {
string: {

View File

@ -52,10 +52,22 @@ export async function patch(ctx: UserCtx<PatchRowRequest, PatchRowResponse>) {
const table = await utils.getTableFromSource(source)
const { _id, ...rowData } = ctx.request.body
const dataToUpdate = await inputProcessing(
const beforeRow = await sdk.rows.external.getRow(table._id!, _id, {
relationships: true,
})
let dataToUpdate = cloneDeep(beforeRow)
const allowedField = utils.getSourceFields(source)
for (const key of Object.keys(rowData)) {
if (!allowedField.includes(key)) continue
dataToUpdate[key] = rowData[key]
}
dataToUpdate = await inputProcessing(
ctx.user?._id,
cloneDeep(source),
rowData
dataToUpdate
)
const validateResult = await sdk.rows.utils.validate({
@ -66,10 +78,6 @@ export async function patch(ctx: UserCtx<PatchRowRequest, PatchRowResponse>) {
throw { validation: validateResult.errors }
}
const beforeRow = await sdk.rows.external.getRow(table._id!, _id, {
relationships: true,
})
const response = await handleRequest(Operation.UPDATE, source, {
id: breakRowIdField(_id),
row: dataToUpdate,

View File

@ -66,7 +66,7 @@ export function getSourceId(ctx: Ctx): { tableId: string; viewId?: string } {
if (docIds.isViewId(sourceId)) {
return {
tableId: utils.extractViewInfoFromID(sourceId).tableId,
viewId: sourceId,
viewId: sql.utils.encodeViewId(sourceId),
}
}
return { tableId: sql.utils.encodeTableId(ctx.params.sourceId) }
@ -110,6 +110,21 @@ function fixBooleanFields(row: Row, table: Table) {
return row
}
export function getSourceFields(source: Table | ViewV2): string[] {
const isView = sdk.views.isView(source)
if (isView) {
const fields = Object.keys(
helpers.views.basicFields(source, { visible: true })
)
return fields
}
const fields = Object.entries(source.schema)
.filter(([_, field]) => field.visible !== false)
.map(([columnName]) => columnName)
return fields
}
export async function sqlOutputProcessing(
rows: DatasourcePlusQueryResponse,
source: Table | ViewV2,

View File

@ -1333,6 +1333,62 @@ if (descriptions.length) {
expect(resp.relationship.length).toBe(1)
})
it("should be able to keep linked data when updating from views that trims links from the main table", async () => {
let row = await config.api.row.save(table._id!, {
name: "main",
description: "main description",
})
const row2 = await config.api.row.save(otherTable._id!, {
name: "link",
description: "link description",
relationship: [row._id],
})
const view = await config.api.viewV2.create({
tableId: table._id!,
name: "view",
schema: {
name: { visible: true },
},
})
const resp = await config.api.row.patch(view.id, {
_id: row._id!,
_rev: row._rev!,
tableId: row.tableId!,
name: "test2",
relationship: [row2._id],
})
expect(resp.relationship).toBeUndefined()
const updatedRow = await config.api.row.get(table._id!, row._id!)
expect(updatedRow.relationship.length).toBe(1)
})
it("should be able to keep linked data when updating from views that trims links from the foreign table", async () => {
let row = await config.api.row.save(table._id!, {
name: "main",
description: "main description",
})
const row2 = await config.api.row.save(otherTable._id!, {
name: "link",
description: "link description",
relationship: [row._id],
})
const view = await config.api.viewV2.create({
tableId: otherTable._id!,
name: "view",
})
await config.api.row.patch(view.id, {
_id: row2._id!,
_rev: row2._rev!,
tableId: row2.tableId!,
})
const updatedRow = await config.api.row.get(table._id!, row._id!)
expect(updatedRow.relationship.length).toBe(1)
})
!isInternal &&
// MSSQL needs a setting called IDENTITY_INSERT to be set to ON to allow writing
// to identity columns. This is not something Budibase does currently.

View File

@ -55,7 +55,7 @@ if (descriptions.length) {
let datasource: Datasource | undefined
function saveTableRequest(
...overrides: Partial<Omit<SaveTableRequest, "name">>[]
...overrides: Partial<SaveTableRequest>[]
): SaveTableRequest {
const req: SaveTableRequest = {
name: generator.guid().replaceAll("-", "").substring(0, 16),
@ -1898,6 +1898,36 @@ if (descriptions.length) {
}
expect(view.queryUI).toEqual(expected)
})
it("tables and views can contain whitespaces", async () => {
const table = await config.api.table.save(
saveTableRequest({
name: `table with spaces ${generator.hash()}`,
schema: {
name: {
type: FieldType.STRING,
name: "name",
},
},
})
)
const view = await config.api.viewV2.create({
tableId: table._id!,
name: `view name with spaces`,
schema: {
name: { visible: true },
},
})
expect(await getDelegate(view)).toEqual({
...view,
schema: {
id: { ...table.schema["id"], visible: false },
name: { ...table.schema["name"], visible: true },
},
})
})
})
describe("updating table schema", () => {

View File

@ -1,10 +1,4 @@
import {
context,
db as dbCore,
docIds,
utils,
sql,
} from "@budibase/backend-core"
import { context, db as dbCore, docIds, utils } from "@budibase/backend-core"
import {
DatabaseQueryOpts,
Datasource,
@ -334,7 +328,7 @@ export function extractViewInfoFromID(viewId: string) {
const regex = new RegExp(`^(?<tableId>.+)${SEPARATOR}([^${SEPARATOR}]+)$`)
const res = regex.exec(viewId)
return {
tableId: sql.utils.encodeTableId(res!.groups!["tableId"]),
tableId: res!.groups!["tableId"],
}
}

View File

@ -46,8 +46,11 @@ export class ViewV2API extends TestAPI {
}
get = async (viewId: string) => {
return (await this._get<ViewResponseEnriched>(`/api/v2/views/${viewId}`))
.data
return (
await this._get<ViewResponseEnriched>(
`/api/v2/views/${encodeURIComponent(viewId)}`
)
).data
}
fetch = async (expectations?: Expectations) => {