799 lines
25 KiB
TypeScript
799 lines
25 KiB
TypeScript
import dayjs from "dayjs"
|
|
import {
|
|
Aggregation,
|
|
AutoFieldSubType,
|
|
AutoReason,
|
|
Datasource,
|
|
DatasourcePlusQueryResponse,
|
|
FieldSchema,
|
|
FieldType,
|
|
FilterType,
|
|
IncludeRelationship,
|
|
InternalSearchFilterOperator,
|
|
isManyToOne,
|
|
isOneToMany,
|
|
OneToManyRelationshipFieldMetadata,
|
|
Operation,
|
|
PaginationJson,
|
|
QueryJson,
|
|
RelationshipFieldMetadata,
|
|
Row,
|
|
SearchFilters,
|
|
SortJson,
|
|
SortType,
|
|
Table,
|
|
ViewV2,
|
|
} from "@budibase/types"
|
|
import {
|
|
breakExternalTableId,
|
|
breakRowIdField,
|
|
convertRowId,
|
|
generateRowIdField,
|
|
isRowId,
|
|
isSQL,
|
|
} from "../../../integrations/utils"
|
|
import {
|
|
buildExternalRelationships,
|
|
buildSqlFieldList,
|
|
generateIdForRow,
|
|
isKnexEmptyReadResponse,
|
|
isManyToMany,
|
|
sqlOutputProcessing,
|
|
} from "./utils"
|
|
import {
|
|
getDatasourceAndQuery,
|
|
processRowCountResponse,
|
|
} from "../../../sdk/app/rows/utils"
|
|
import { processObjectSync } from "@budibase/string-templates"
|
|
import { cloneDeep } from "lodash/fp"
|
|
import { db as dbCore } from "@budibase/backend-core"
|
|
import sdk from "../../../sdk"
|
|
import env from "../../../environment"
|
|
import { makeExternalQuery } from "../../../integrations/base/query"
|
|
import { dataFilters, helpers } from "@budibase/shared-core"
|
|
import { isRelationshipColumn } from "../../../db/utils"
|
|
|
|
export interface ManyRelationship {
|
|
tableId?: string
|
|
id?: string
|
|
isUpdate?: boolean
|
|
key: string
|
|
[key: string]: any
|
|
}
|
|
|
|
export interface RunConfig {
|
|
id?: any[]
|
|
filters?: SearchFilters
|
|
sort?: SortJson
|
|
paginate?: PaginationJson
|
|
datasource?: Datasource
|
|
row?: Row
|
|
rows?: Row[]
|
|
tables?: Record<string, Table>
|
|
includeSqlRelationships?: IncludeRelationship
|
|
}
|
|
|
|
export type ExternalReadRequestReturnType = {
|
|
rows: Row[]
|
|
rawResponseSize: number
|
|
}
|
|
|
|
export type ExternalRequestReturnType<T extends Operation> =
|
|
T extends Operation.READ
|
|
? ExternalReadRequestReturnType
|
|
: T extends Operation.COUNT
|
|
? number
|
|
: { row: Row; table: Table }
|
|
|
|
/**
|
|
* This function checks the incoming parameters to make sure all the inputs are
|
|
* valid based on on the table schema. The main thing this is looking for is when a
|
|
* user has made use of the _id field of a row for a foreign key or a search parameter.
|
|
* In these cases the key will be sent up as [1], rather than 1. In these cases we will
|
|
* simplify it down to the requirements. This function is quite complex as we try to be
|
|
* relatively restrictive over what types of columns we will perform this action for.
|
|
*/
|
|
function cleanupConfig(config: RunConfig, table: Table): RunConfig {
|
|
const primaryOptions = [
|
|
FieldType.STRING,
|
|
FieldType.LONGFORM,
|
|
FieldType.OPTIONS,
|
|
FieldType.NUMBER,
|
|
]
|
|
// filter out fields which cannot be keys
|
|
const fieldNames = Object.entries(table.schema)
|
|
.filter(schema => primaryOptions.find(val => val === schema[1].type))
|
|
// map to fieldName
|
|
.map(entry => entry[0])
|
|
const iterateObject = (obj: { [key: string]: any }) => {
|
|
for (let [field, value] of Object.entries(obj)) {
|
|
if (fieldNames.find(name => name === field) && isRowId(value)) {
|
|
obj[field] = convertRowId(value)
|
|
}
|
|
}
|
|
}
|
|
// check the row and filters to make sure they aren't a key of some sort
|
|
if (config.filters) {
|
|
for (let [key, filter] of Object.entries(config.filters)) {
|
|
// oneOf is an array, don't iterate it
|
|
if (
|
|
typeof filter !== "object" ||
|
|
Object.keys(filter).length === 0 ||
|
|
key === FilterType.ONE_OF
|
|
) {
|
|
continue
|
|
}
|
|
iterateObject(filter)
|
|
}
|
|
}
|
|
if (config.row) {
|
|
iterateObject(config.row)
|
|
}
|
|
|
|
return config
|
|
}
|
|
|
|
function getEndpoint(tableId: string | undefined, operation: string) {
|
|
if (!tableId) {
|
|
throw new Error("Cannot get endpoint information - no table ID specified")
|
|
}
|
|
const { datasourceId, tableName } = breakExternalTableId(tableId)
|
|
return {
|
|
datasourceId: datasourceId,
|
|
entityId: tableName,
|
|
operation: operation as Operation,
|
|
}
|
|
}
|
|
|
|
function isOneSide(
|
|
field: RelationshipFieldMetadata
|
|
): field is OneToManyRelationshipFieldMetadata {
|
|
return (
|
|
field.relationshipType && field.relationshipType.split("-")[0] === "one"
|
|
)
|
|
}
|
|
|
|
function isEditableColumn(column: FieldSchema) {
|
|
const isExternalAutoColumn =
|
|
column.autocolumn &&
|
|
column.autoReason !== AutoReason.FOREIGN_KEY &&
|
|
column.subtype !== AutoFieldSubType.AUTO_ID
|
|
const isFormula = column.type === FieldType.FORMULA
|
|
return !(isExternalAutoColumn || isFormula)
|
|
}
|
|
|
|
export class ExternalRequest<T extends Operation> {
|
|
private readonly operation: T
|
|
private readonly source: Table | ViewV2
|
|
private datasource: Datasource
|
|
|
|
public static async for<T extends Operation>(
|
|
operation: T,
|
|
source: Table | ViewV2,
|
|
opts: { datasource?: Datasource } = {}
|
|
) {
|
|
if (!opts.datasource) {
|
|
if (sdk.views.isView(source)) {
|
|
const table = await sdk.views.getTable(source.id)
|
|
opts.datasource = await sdk.datasources.get(table.sourceId)
|
|
} else {
|
|
opts.datasource = await sdk.datasources.get(source.sourceId)
|
|
}
|
|
}
|
|
|
|
return new ExternalRequest(operation, source, opts.datasource)
|
|
}
|
|
|
|
private get tables(): { [key: string]: Table } {
|
|
if (!this.datasource.entities) {
|
|
throw new Error("Datasource does not have entities")
|
|
}
|
|
return this.datasource.entities
|
|
}
|
|
|
|
private constructor(
|
|
operation: T,
|
|
source: Table | ViewV2,
|
|
datasource: Datasource
|
|
) {
|
|
this.operation = operation
|
|
this.source = source
|
|
this.datasource = datasource
|
|
}
|
|
|
|
private prepareFilters(
|
|
id: string | undefined | string[],
|
|
filters: SearchFilters,
|
|
table: Table
|
|
): SearchFilters {
|
|
const primary = table.primary
|
|
// if passed in array need to copy for shifting etc
|
|
let idCopy: undefined | string | any[] = cloneDeep(id)
|
|
if (filters) {
|
|
// need to map over the filters and make sure the _id field isn't present
|
|
let prefix = 1
|
|
const checkFilters = (innerFilters: SearchFilters): SearchFilters => {
|
|
for (const [operatorType, operator] of Object.entries(innerFilters)) {
|
|
const isArrayOp = sdk.rows.utils.isArrayFilter(operatorType)
|
|
for (const field of Object.keys(operator || {})) {
|
|
if (dbCore.removeKeyNumbering(field) === "_id") {
|
|
if (primary) {
|
|
const parts = breakRowIdField(operator[field])
|
|
if (primary.length > 1 && isArrayOp) {
|
|
operator[InternalSearchFilterOperator.COMPLEX_ID_OPERATOR] = {
|
|
id: primary,
|
|
values: parts[0],
|
|
}
|
|
} else {
|
|
for (let field of primary) {
|
|
operator[`${prefix}:${field}`] = parts.shift()
|
|
}
|
|
prefix++
|
|
}
|
|
}
|
|
// make sure this field doesn't exist on any filter
|
|
delete operator[field]
|
|
}
|
|
}
|
|
}
|
|
return dataFilters.recurseLogicalOperators(innerFilters, checkFilters)
|
|
}
|
|
checkFilters(filters)
|
|
}
|
|
// there is no id, just use the user provided filters
|
|
if (!idCopy || !table) {
|
|
return filters
|
|
}
|
|
// if used as URL parameter it will have been joined
|
|
if (!Array.isArray(idCopy)) {
|
|
idCopy = breakRowIdField(idCopy)
|
|
}
|
|
const equal: SearchFilters["equal"] = {}
|
|
if (primary && idCopy) {
|
|
for (let field of primary) {
|
|
// work through the ID and get the parts
|
|
equal[field] = idCopy.shift()
|
|
}
|
|
}
|
|
return {
|
|
equal,
|
|
}
|
|
}
|
|
|
|
private async removeManyToManyRelationships(rowId: string, table: Table) {
|
|
const tableId = table._id!
|
|
const filters = this.prepareFilters(rowId, {}, table)
|
|
// safety check, if there are no filters on deletion bad things happen
|
|
if (Object.keys(filters).length !== 0) {
|
|
return getDatasourceAndQuery({
|
|
endpoint: getEndpoint(tableId, Operation.DELETE),
|
|
filters,
|
|
meta: {
|
|
table,
|
|
},
|
|
})
|
|
} else {
|
|
return []
|
|
}
|
|
}
|
|
|
|
private async removeOneToManyRelationships(
|
|
rowId: string,
|
|
table: Table,
|
|
colName: string
|
|
) {
|
|
const tableId = table._id!
|
|
const filters = this.prepareFilters(rowId, {}, table)
|
|
// safety check, if there are no filters on deletion bad things happen
|
|
if (Object.keys(filters).length !== 0) {
|
|
return getDatasourceAndQuery({
|
|
endpoint: getEndpoint(tableId, Operation.UPDATE),
|
|
body: { [colName]: null },
|
|
filters,
|
|
meta: {
|
|
table,
|
|
},
|
|
})
|
|
} else {
|
|
return []
|
|
}
|
|
}
|
|
|
|
getTable(tableId: string | undefined): Table | undefined {
|
|
if (!tableId) {
|
|
throw new Error("Table ID is unknown, cannot find table")
|
|
}
|
|
const { tableName } = breakExternalTableId(tableId)
|
|
return this.tables[tableName]
|
|
}
|
|
|
|
async getRow(table: Table, rowId: string): Promise<Row> {
|
|
const response = await getDatasourceAndQuery({
|
|
endpoint: getEndpoint(table._id!, Operation.READ),
|
|
filters: this.prepareFilters(rowId, {}, table),
|
|
meta: {
|
|
table,
|
|
},
|
|
})
|
|
if (Array.isArray(response) && response.length > 0) {
|
|
return response[0]
|
|
} else {
|
|
throw new Error(`Cannot fetch row by ID "${rowId}"`)
|
|
}
|
|
}
|
|
|
|
inputProcessing<T extends Row | undefined>(
|
|
row: T,
|
|
table: Table
|
|
): { row: T; manyRelationships: ManyRelationship[] } {
|
|
if (!row) {
|
|
return { row, manyRelationships: [] }
|
|
}
|
|
// we don't really support composite keys for relationships, this is why [0] is used
|
|
// @ts-ignore
|
|
const tablePrimary: string = table.primary[0]
|
|
let newRow: Row = {},
|
|
manyRelationships: ManyRelationship[] = []
|
|
for (let [key, field] of Object.entries(table.schema)) {
|
|
// if set already, or not set just skip it
|
|
if (row[key] === undefined || newRow[key]) {
|
|
continue
|
|
}
|
|
if (
|
|
!(this.operation === Operation.BULK_UPSERT) &&
|
|
!isEditableColumn(field)
|
|
) {
|
|
continue
|
|
}
|
|
// parse floats/numbers
|
|
if (field.type === FieldType.NUMBER && !isNaN(parseFloat(row[key]))) {
|
|
newRow[key] = parseFloat(row[key])
|
|
} else if (field.type === FieldType.LINK) {
|
|
const { tableName: linkTableName } = breakExternalTableId(field.tableId)
|
|
// table has to exist for many to many
|
|
if (!linkTableName || !this.tables[linkTableName]) {
|
|
continue
|
|
}
|
|
const linkTable = this.tables[linkTableName]
|
|
// @ts-ignore
|
|
const linkTablePrimary = linkTable.primary[0]
|
|
// one to many
|
|
if (isOneSide(field)) {
|
|
let id = row[key][0]
|
|
if (id) {
|
|
if (typeof row[key] === "string") {
|
|
id = decodeURIComponent(row[key]).match(/\[(.*?)\]/)?.[1]
|
|
}
|
|
newRow[field.foreignKey || linkTablePrimary] =
|
|
breakRowIdField(id)[0]
|
|
} else {
|
|
// Removing from both new and row, as we don't know if it has already been processed
|
|
row[field.foreignKey || linkTablePrimary] = null
|
|
newRow[field.foreignKey || linkTablePrimary] = null
|
|
}
|
|
}
|
|
// many to many
|
|
else if (isManyToMany(field)) {
|
|
// we're not inserting a doc, will be a bunch of update calls
|
|
const otherKey: string = field.throughFrom || linkTablePrimary
|
|
const thisKey: string = field.throughTo || tablePrimary
|
|
for (const relationship of row[key]) {
|
|
manyRelationships.push({
|
|
tableId: field.through || field.tableId,
|
|
isUpdate: false,
|
|
key: otherKey,
|
|
[otherKey]: breakRowIdField(relationship)[0],
|
|
// leave the ID for enrichment later
|
|
[thisKey]: `{{ literal ${tablePrimary} }}`,
|
|
})
|
|
}
|
|
}
|
|
// many to one
|
|
else {
|
|
const thisKey: string = "id"
|
|
// @ts-ignore
|
|
const otherKey: string = field.fieldName
|
|
for (const relationship of row[key]) {
|
|
manyRelationships.push({
|
|
tableId: field.tableId,
|
|
isUpdate: true,
|
|
key: otherKey,
|
|
[thisKey]: breakRowIdField(relationship)[0],
|
|
// leave the ID for enrichment later
|
|
[otherKey]: `{{ literal ${tablePrimary} }}`,
|
|
})
|
|
}
|
|
}
|
|
} else if (
|
|
field.type === FieldType.DATETIME &&
|
|
field.timeOnly &&
|
|
row[key] &&
|
|
dayjs(row[key]).isValid()
|
|
) {
|
|
newRow[key] = dayjs(row[key]).format("HH:mm")
|
|
} else {
|
|
newRow[key] = row[key]
|
|
}
|
|
}
|
|
// we return the relationships that may need to be created in the through table
|
|
// we do this so that if the ID is generated by the DB it can be inserted
|
|
// after the fact
|
|
return { row: newRow as T, manyRelationships }
|
|
}
|
|
|
|
/**
|
|
* This is a cached lookup, of relationship records, this is mainly for creating/deleting junction
|
|
* information.
|
|
*/
|
|
async lookupRelations(tableId: string, row: Row) {
|
|
const related: {
|
|
rows: Row[]
|
|
isMany: boolean
|
|
tableId: string
|
|
field: string
|
|
}[] = []
|
|
|
|
const { tableName } = breakExternalTableId(tableId)
|
|
const table = this.tables[tableName]
|
|
// @ts-ignore
|
|
const primaryKey = table.primary[0]
|
|
// make a new request to get the row with all its relationships
|
|
// we need this to work out if any relationships need removed
|
|
for (const field of Object.values(table.schema)) {
|
|
if (
|
|
field.type !== FieldType.LINK ||
|
|
!field.fieldName ||
|
|
isOneSide(field)
|
|
) {
|
|
continue
|
|
}
|
|
let relatedTableId: string | undefined,
|
|
lookupField: string | undefined,
|
|
fieldName: string | undefined
|
|
if (isManyToMany(field)) {
|
|
relatedTableId = field.through
|
|
lookupField = primaryKey
|
|
fieldName = field.throughTo || primaryKey
|
|
} else if (isManyToOne(field)) {
|
|
relatedTableId = field.tableId
|
|
lookupField = field.foreignKey
|
|
fieldName = field.fieldName
|
|
}
|
|
if (!relatedTableId || !lookupField || !fieldName) {
|
|
throw new Error(
|
|
"Unable to lookup relationships - undefined column properties."
|
|
)
|
|
}
|
|
const { tableName: relatedTableName } =
|
|
breakExternalTableId(relatedTableId)
|
|
// @ts-ignore
|
|
const linkPrimaryKey = this.tables[relatedTableName].primary[0]
|
|
if (!lookupField || !row?.[lookupField] == null) {
|
|
continue
|
|
}
|
|
const endpoint = getEndpoint(relatedTableId, Operation.READ)
|
|
const relatedTable = this.tables[endpoint.entityId]
|
|
if (!relatedTable) {
|
|
throw new Error("unable to find related table")
|
|
}
|
|
const response = await getDatasourceAndQuery({
|
|
endpoint: endpoint,
|
|
filters: {
|
|
equal: {
|
|
[fieldName]: row[lookupField],
|
|
},
|
|
},
|
|
meta: {
|
|
table: relatedTable,
|
|
},
|
|
})
|
|
// this is the response from knex if no rows found
|
|
const rows: Row[] =
|
|
!Array.isArray(response) || isKnexEmptyReadResponse(response)
|
|
? []
|
|
: response
|
|
const storeTo = isManyToMany(field)
|
|
? field.throughFrom || linkPrimaryKey
|
|
: fieldName
|
|
|
|
related.push({
|
|
rows,
|
|
isMany: isManyToMany(field),
|
|
tableId: relatedTableId,
|
|
field: storeTo,
|
|
})
|
|
}
|
|
return related
|
|
}
|
|
|
|
/**
|
|
* Once a row has been written we may need to update a many field, e.g. updating foreign keys
|
|
* in a bunch of rows in another table, or inserting/deleting rows from a junction table (many to many).
|
|
* This is quite a complex process and is handled by this function, there are a few things going on here:
|
|
* 1. If updating foreign keys its relatively simple, just create a filter for the row that needs updated
|
|
* and write the various components.
|
|
* 2. If junction table, then we lookup what exists already, write what doesn't exist, work out what
|
|
* isn't supposed to exist anymore and delete those. This is better than the usual method of delete them
|
|
* all and then re-create, as theres no chance of losing data (e.g. delete succeed, but write fail).
|
|
*/
|
|
async handleManyRelationships(
|
|
mainTableId: string,
|
|
row: Row,
|
|
relationships: ManyRelationship[]
|
|
) {
|
|
// if we're creating (in a through table) need to wipe the existing ones first
|
|
const promises = []
|
|
const related = await this.lookupRelations(mainTableId, row)
|
|
for (let relationship of relationships) {
|
|
const { key, tableId, isUpdate, id, ...rest } = relationship
|
|
const body: { [key: string]: any } = processObjectSync(rest, row, {})
|
|
const linkTable = this.getTable(tableId)
|
|
const relationshipPrimary = linkTable?.primary || []
|
|
const linkPrimary = relationshipPrimary[0]
|
|
if (!linkTable || !linkPrimary) {
|
|
return
|
|
}
|
|
|
|
const linkSecondary = relationshipPrimary[1]
|
|
|
|
const rows =
|
|
related.find(r => r.tableId === relationship.tableId && r.field === key)
|
|
?.rows || []
|
|
|
|
const relationshipMatchPredicate = ({
|
|
row,
|
|
linkPrimary,
|
|
linkSecondary,
|
|
}: {
|
|
row: Row
|
|
linkPrimary: string
|
|
linkSecondary?: string
|
|
}) => {
|
|
const matchesPrimaryLink =
|
|
row[linkPrimary] === relationship.id ||
|
|
row[linkPrimary] === body?.[linkPrimary]
|
|
if (!matchesPrimaryLink || !linkSecondary) {
|
|
return matchesPrimaryLink
|
|
}
|
|
|
|
const matchesSecondaryLink =
|
|
row[linkSecondary] === body?.[linkSecondary]
|
|
return matchesPrimaryLink && matchesSecondaryLink
|
|
}
|
|
|
|
const existingRelationship = rows.find((row: { [key: string]: any }) =>
|
|
relationshipMatchPredicate({ row, linkPrimary, linkSecondary })
|
|
)
|
|
const operation = isUpdate ? Operation.UPDATE : Operation.CREATE
|
|
if (!existingRelationship) {
|
|
promises.push(
|
|
getDatasourceAndQuery({
|
|
endpoint: getEndpoint(tableId, operation),
|
|
// if we're doing many relationships then we're writing, only one response
|
|
body,
|
|
filters: this.prepareFilters(id, {}, linkTable),
|
|
meta: {
|
|
table: linkTable,
|
|
},
|
|
})
|
|
)
|
|
} else {
|
|
// remove the relationship from cache so it isn't adjusted again
|
|
rows.splice(rows.indexOf(existingRelationship), 1)
|
|
}
|
|
}
|
|
// finally cleanup anything that needs to be removed
|
|
for (let { isMany, rows, tableId, field } of related) {
|
|
const table: Table | undefined = this.getTable(tableId)
|
|
// if it's not the foreign key skip it, nothing to do
|
|
if (
|
|
!table ||
|
|
(!isMany && table.primary && table.primary.indexOf(field) !== -1)
|
|
) {
|
|
continue
|
|
}
|
|
for (let row of rows) {
|
|
const rowId = generateIdForRow(row, table)
|
|
const promise: Promise<any> = isMany
|
|
? this.removeManyToManyRelationships(rowId, table)
|
|
: this.removeOneToManyRelationships(rowId, table, field)
|
|
if (promise) {
|
|
promises.push(promise)
|
|
}
|
|
}
|
|
}
|
|
await Promise.all(promises)
|
|
}
|
|
|
|
async removeRelationshipsToRow(table: Table, rowId: string) {
|
|
const row = await this.getRow(table, rowId)
|
|
const related = await this.lookupRelations(table._id!, row)
|
|
for (const column of Object.values(table.schema)) {
|
|
if (!isRelationshipColumn(column) || isOneToMany(column)) {
|
|
continue
|
|
}
|
|
|
|
const relatedByTable = isManyToMany(column)
|
|
? related.find(
|
|
r => r.tableId === column.through && r.field === column.fieldName
|
|
)
|
|
: related.find(r => r.field === column.fieldName)
|
|
if (!relatedByTable) {
|
|
continue
|
|
}
|
|
|
|
const { rows, isMany, tableId } = relatedByTable
|
|
const table = this.getTable(tableId)!
|
|
await Promise.all(
|
|
rows.map(row => {
|
|
const rowId = generateIdForRow(row, table)
|
|
return isMany
|
|
? this.removeManyToManyRelationships(rowId, table)
|
|
: this.removeOneToManyRelationships(rowId, table, column.fieldName)
|
|
})
|
|
)
|
|
}
|
|
}
|
|
|
|
async run(config: RunConfig): Promise<ExternalRequestReturnType<T>> {
|
|
const { operation } = this
|
|
let table: Table
|
|
if (sdk.views.isView(this.source)) {
|
|
table = await sdk.views.getTable(this.source.id)
|
|
} else {
|
|
table = this.source
|
|
}
|
|
|
|
let isSql = isSQL(this.datasource)
|
|
|
|
// look for specific components of config which may not be considered acceptable
|
|
let { id, row, filters, sort, paginate, rows } = cleanupConfig(
|
|
config,
|
|
table
|
|
)
|
|
//if the sort column is a formula, remove it
|
|
for (let sortColumn of Object.keys(sort || {})) {
|
|
if (!sort?.[sortColumn]) {
|
|
continue
|
|
}
|
|
switch (table.schema[sortColumn]?.type) {
|
|
case FieldType.FORMULA:
|
|
delete sort?.[sortColumn]
|
|
break
|
|
case FieldType.NUMBER:
|
|
if (sort && sort[sortColumn]) {
|
|
sort[sortColumn].type = SortType.NUMBER
|
|
}
|
|
break
|
|
}
|
|
}
|
|
filters = this.prepareFilters(id, filters || {}, table)
|
|
const relationships = buildExternalRelationships(table, this.tables)
|
|
|
|
const incRelationships =
|
|
config.includeSqlRelationships === IncludeRelationship.INCLUDE
|
|
|
|
// clean up row on ingress using schema
|
|
const unprocessedRow = config.row
|
|
const processed = this.inputProcessing(row, table)
|
|
row = processed.row
|
|
let manyRelationships = processed.manyRelationships
|
|
|
|
if (!row && rows) {
|
|
manyRelationships = []
|
|
for (let i = 0; i < rows.length; i++) {
|
|
const processed = this.inputProcessing(rows[i], table)
|
|
rows[i] = processed.row
|
|
if (processed.manyRelationships.length) {
|
|
manyRelationships.push(...processed.manyRelationships)
|
|
}
|
|
}
|
|
}
|
|
|
|
if (
|
|
operation === Operation.DELETE &&
|
|
(filters == null || Object.keys(filters).length === 0)
|
|
) {
|
|
throw "Deletion must be filtered"
|
|
}
|
|
|
|
let aggregations: Aggregation[] = []
|
|
if (sdk.views.isView(this.source)) {
|
|
const calculationFields = helpers.views.calculationFields(this.source)
|
|
for (const [key, field] of Object.entries(calculationFields)) {
|
|
aggregations.push({
|
|
...field,
|
|
name: key,
|
|
})
|
|
}
|
|
}
|
|
|
|
let json: QueryJson = {
|
|
endpoint: {
|
|
datasourceId: this.datasource._id!,
|
|
entityId: table.name,
|
|
operation,
|
|
},
|
|
resource: {
|
|
// have to specify the fields to avoid column overlap (for SQL)
|
|
fields: isSql
|
|
? await buildSqlFieldList(this.source, this.tables, {
|
|
relationships: incRelationships,
|
|
})
|
|
: [],
|
|
aggregations,
|
|
},
|
|
filters,
|
|
sort,
|
|
paginate,
|
|
relationships,
|
|
body: row || rows,
|
|
// pass an id filter into extra, purely for mysql/returning
|
|
extra: {
|
|
idFilter: this.prepareFilters(
|
|
id || generateIdForRow(row, table),
|
|
{},
|
|
table
|
|
),
|
|
},
|
|
meta: {
|
|
table,
|
|
tables: this.tables,
|
|
},
|
|
}
|
|
|
|
// remove any relationships that could block deletion
|
|
if (operation === Operation.DELETE && id) {
|
|
await this.removeRelationshipsToRow(table, generateRowIdField(id))
|
|
}
|
|
|
|
// aliasing can be disabled fully if desired
|
|
const aliasing = new sdk.rows.AliasTables(Object.keys(this.tables))
|
|
let response: DatasourcePlusQueryResponse
|
|
// there's a chance after input processing nothing needs updated, so pass over the call
|
|
// we might still need to perform other operations like updating the foreign keys on other rows
|
|
if (
|
|
this.operation === Operation.UPDATE &&
|
|
Object.keys(row || {}).length === 0 &&
|
|
unprocessedRow
|
|
) {
|
|
response = [unprocessedRow]
|
|
} else {
|
|
response = env.SQL_ALIASING_DISABLE
|
|
? await getDatasourceAndQuery(json)
|
|
: await aliasing.queryWithAliasing(json, makeExternalQuery)
|
|
}
|
|
|
|
// if it's a counting operation there will be no more processing, just return the number
|
|
if (this.operation === Operation.COUNT) {
|
|
return processRowCountResponse(response) as ExternalRequestReturnType<T>
|
|
}
|
|
|
|
const responseRows = Array.isArray(response) ? response : []
|
|
// handle many-to-many relationships now if we know the ID (could be auto increment)
|
|
if (operation !== Operation.READ) {
|
|
await this.handleManyRelationships(
|
|
table._id || "",
|
|
responseRows[0],
|
|
processed.manyRelationships
|
|
)
|
|
}
|
|
const output = await sqlOutputProcessing(
|
|
response,
|
|
this.source,
|
|
this.tables,
|
|
relationships
|
|
)
|
|
// if reading it'll just be an array of rows, return whole thing
|
|
if (operation === Operation.READ) {
|
|
const rows = Array.isArray(output) ? output : [output]
|
|
return {
|
|
rows,
|
|
rawResponseSize: responseRows.length,
|
|
} as ExternalRequestReturnType<T>
|
|
} else {
|
|
return { row: output[0], table } as ExternalRequestReturnType<T>
|
|
}
|
|
}
|
|
}
|