Merge branch 'new-datepicker' of github.com:Budibase/budibase into new-datepicker
This commit is contained in:
commit
bb575f1ccd
|
@ -152,6 +152,8 @@ $ helm install --create-namespace --namespace budibase budibase . -f values.yaml
|
||||||
| services.apps.autoscaling.targetCPUUtilizationPercentage | int | `80` | Target CPU utilization percentage for the apps service. Note that for autoscaling to work, you will need to have metrics-server configured, and resources set for the apps pods. |
|
| services.apps.autoscaling.targetCPUUtilizationPercentage | int | `80` | Target CPU utilization percentage for the apps service. Note that for autoscaling to work, you will need to have metrics-server configured, and resources set for the apps pods. |
|
||||||
| services.apps.extraContainers | list | `[]` | Additional containers to be added to the apps pod. |
|
| services.apps.extraContainers | list | `[]` | Additional containers to be added to the apps pod. |
|
||||||
| services.apps.extraEnv | list | `[]` | Extra environment variables to set for apps pods. Takes a list of name=value pairs. |
|
| services.apps.extraEnv | list | `[]` | Extra environment variables to set for apps pods. Takes a list of name=value pairs. |
|
||||||
|
| services.apps.extraVolumeMounts | list | `[]` | Additional volumeMounts to the main apps container. |
|
||||||
|
| services.apps.extraVolumes | list | `[]` | Additional volumes to the apps pod. |
|
||||||
| services.apps.httpLogging | int | `1` | Whether or not to log HTTP requests to the apps service. |
|
| services.apps.httpLogging | int | `1` | Whether or not to log HTTP requests to the apps service. |
|
||||||
| services.apps.livenessProbe | object | HTTP health checks. | Liveness probe configuration for apps pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
| services.apps.livenessProbe | object | HTTP health checks. | Liveness probe configuration for apps pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
||||||
| services.apps.logLevel | string | `"info"` | The log level for the apps service. |
|
| services.apps.logLevel | string | `"info"` | The log level for the apps service. |
|
||||||
|
@ -166,6 +168,8 @@ $ helm install --create-namespace --namespace budibase budibase . -f values.yaml
|
||||||
| services.automationWorkers.enabled | bool | `true` | Whether or not to enable the automation worker service. If you disable this, automations will be processed by the apps service. |
|
| services.automationWorkers.enabled | bool | `true` | Whether or not to enable the automation worker service. If you disable this, automations will be processed by the apps service. |
|
||||||
| services.automationWorkers.extraContainers | list | `[]` | Additional containers to be added to the automationWorkers pod. |
|
| services.automationWorkers.extraContainers | list | `[]` | Additional containers to be added to the automationWorkers pod. |
|
||||||
| services.automationWorkers.extraEnv | list | `[]` | Extra environment variables to set for automation worker pods. Takes a list of name=value pairs. |
|
| services.automationWorkers.extraEnv | list | `[]` | Extra environment variables to set for automation worker pods. Takes a list of name=value pairs. |
|
||||||
|
| services.automationWorkers.extraVolumeMounts | list | `[]` | Additional volumeMounts to the main automationWorkers container. |
|
||||||
|
| services.automationWorkers.extraVolumes | list | `[]` | Additional volumes to the automationWorkers pod. |
|
||||||
| services.automationWorkers.livenessProbe | object | HTTP health checks. | Liveness probe configuration for automation worker pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
| services.automationWorkers.livenessProbe | object | HTTP health checks. | Liveness probe configuration for automation worker pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
||||||
| services.automationWorkers.logLevel | string | `"info"` | The log level for the automation worker service. |
|
| services.automationWorkers.logLevel | string | `"info"` | The log level for the automation worker service. |
|
||||||
| services.automationWorkers.readinessProbe | object | HTTP health checks. | Readiness probe configuration for automation worker pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
| services.automationWorkers.readinessProbe | object | HTTP health checks. | Readiness probe configuration for automation worker pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
||||||
|
@ -185,6 +189,8 @@ $ helm install --create-namespace --namespace budibase budibase . -f values.yaml
|
||||||
| services.objectStore.cloudfront.privateKey64 | string | `""` | Base64 encoded private key for the above public key. |
|
| services.objectStore.cloudfront.privateKey64 | string | `""` | Base64 encoded private key for the above public key. |
|
||||||
| services.objectStore.cloudfront.publicKeyId | string | `""` | ID of public key stored in cloudfront. |
|
| services.objectStore.cloudfront.publicKeyId | string | `""` | ID of public key stored in cloudfront. |
|
||||||
| services.objectStore.extraContainers | list | `[]` | Additional containers to be added to the objectStore pod. |
|
| services.objectStore.extraContainers | list | `[]` | Additional containers to be added to the objectStore pod. |
|
||||||
|
| services.objectStore.extraVolumeMounts | list | `[]` | Additional volumeMounts to the main objectStore container. |
|
||||||
|
| services.objectStore.extraVolumes | list | `[]` | Additional volumes to the objectStore pod. |
|
||||||
| services.objectStore.minio | bool | `true` | Set to false if using another object store, such as S3. You will need to set `services.objectStore.url` to point to your bucket if you do this. |
|
| services.objectStore.minio | bool | `true` | Set to false if using another object store, such as S3. You will need to set `services.objectStore.url` to point to your bucket if you do this. |
|
||||||
| services.objectStore.region | string | `""` | AWS_REGION if using S3 |
|
| services.objectStore.region | string | `""` | AWS_REGION if using S3 |
|
||||||
| services.objectStore.resources | object | `{}` | The resources to use for Minio pods. See <https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/> for more information on how to set these. |
|
| services.objectStore.resources | object | `{}` | The resources to use for Minio pods. See <https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/> for more information on how to set these. |
|
||||||
|
@ -197,6 +203,8 @@ $ helm install --create-namespace --namespace budibase budibase . -f values.yaml
|
||||||
| services.proxy.autoscaling.minReplicas | int | `1` | |
|
| services.proxy.autoscaling.minReplicas | int | `1` | |
|
||||||
| services.proxy.autoscaling.targetCPUUtilizationPercentage | int | `80` | Target CPU utilization percentage for the proxy service. Note that for autoscaling to work, you will need to have metrics-server configured, and resources set for the proxy pods. |
|
| services.proxy.autoscaling.targetCPUUtilizationPercentage | int | `80` | Target CPU utilization percentage for the proxy service. Note that for autoscaling to work, you will need to have metrics-server configured, and resources set for the proxy pods. |
|
||||||
| services.proxy.extraContainers | list | `[]` | |
|
| services.proxy.extraContainers | list | `[]` | |
|
||||||
|
| services.proxy.extraVolumeMounts | list | `[]` | Additional volumeMounts to the main proxy container. |
|
||||||
|
| services.proxy.extraVolumes | list | `[]` | Additional volumes to the proxy pod. |
|
||||||
| services.proxy.livenessProbe | object | HTTP health checks. | Liveness probe configuration for proxy pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
| services.proxy.livenessProbe | object | HTTP health checks. | Liveness probe configuration for proxy pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
||||||
| services.proxy.readinessProbe | object | HTTP health checks. | Readiness probe configuration for proxy pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
| services.proxy.readinessProbe | object | HTTP health checks. | Readiness probe configuration for proxy pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
||||||
| services.proxy.replicaCount | int | `1` | The number of proxy replicas to run. |
|
| services.proxy.replicaCount | int | `1` | The number of proxy replicas to run. |
|
||||||
|
@ -204,6 +212,9 @@ $ helm install --create-namespace --namespace budibase budibase . -f values.yaml
|
||||||
| services.proxy.startupProbe | object | HTTP health checks. | Startup probe configuration for proxy pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
| services.proxy.startupProbe | object | HTTP health checks. | Startup probe configuration for proxy pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
||||||
| services.redis.enabled | bool | `true` | Whether or not to deploy a Redis pod into your cluster. |
|
| services.redis.enabled | bool | `true` | Whether or not to deploy a Redis pod into your cluster. |
|
||||||
| services.redis.extraContainers | list | `[]` | Additional containers to be added to the redis pod. |
|
| services.redis.extraContainers | list | `[]` | Additional containers to be added to the redis pod. |
|
||||||
|
| services.redis.extraVolumeMounts | list | `[]` | Additional volumeMounts to the main redis container. |
|
||||||
|
| services.redis.extraVolumes | list | `[]` | Additional volumes to the redis pod. |
|
||||||
|
| services.redis.image | string | `"redis"` | The Redis image to use. |
|
||||||
| services.redis.password | string | `"budibase"` | The password to use when connecting to Redis. It's recommended that you change this from the default if you're running Redis in-cluster. |
|
| services.redis.password | string | `"budibase"` | The password to use when connecting to Redis. It's recommended that you change this from the default if you're running Redis in-cluster. |
|
||||||
| services.redis.port | int | `6379` | Port to expose Redis on. |
|
| services.redis.port | int | `6379` | Port to expose Redis on. |
|
||||||
| services.redis.resources | object | `{}` | The resources to use for Redis pods. See <https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/> for more information on how to set these. |
|
| services.redis.resources | object | `{}` | The resources to use for Redis pods. See <https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/> for more information on how to set these. |
|
||||||
|
@ -216,6 +227,8 @@ $ helm install --create-namespace --namespace budibase budibase . -f values.yaml
|
||||||
| services.worker.autoscaling.targetCPUUtilizationPercentage | int | `80` | Target CPU utilization percentage for the worker service. Note that for autoscaling to work, you will need to have metrics-server configured, and resources set for the worker pods. |
|
| services.worker.autoscaling.targetCPUUtilizationPercentage | int | `80` | Target CPU utilization percentage for the worker service. Note that for autoscaling to work, you will need to have metrics-server configured, and resources set for the worker pods. |
|
||||||
| services.worker.extraContainers | list | `[]` | Additional containers to be added to the worker pod. |
|
| services.worker.extraContainers | list | `[]` | Additional containers to be added to the worker pod. |
|
||||||
| services.worker.extraEnv | list | `[]` | Extra environment variables to set for worker pods. Takes a list of name=value pairs. |
|
| services.worker.extraEnv | list | `[]` | Extra environment variables to set for worker pods. Takes a list of name=value pairs. |
|
||||||
|
| services.worker.extraVolumeMounts | list | `[]` | Additional volumeMounts to the main worker container. |
|
||||||
|
| services.worker.extraVolumes | list | `[]` | Additional volumes to the worker pod. |
|
||||||
| services.worker.httpLogging | int | `1` | Whether or not to log HTTP requests to the worker service. |
|
| services.worker.httpLogging | int | `1` | Whether or not to log HTTP requests to the worker service. |
|
||||||
| services.worker.livenessProbe | object | HTTP health checks. | Liveness probe configuration for worker pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
| services.worker.livenessProbe | object | HTTP health checks. | Liveness probe configuration for worker pods. You shouldn't need to change this, but if you want to you can find more information here: <https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/> |
|
||||||
| services.worker.logLevel | string | `"info"` | The log level for the worker service. |
|
| services.worker.logLevel | string | `"info"` | The log level for the worker service. |
|
||||||
|
|
|
@ -235,6 +235,10 @@ spec:
|
||||||
args:
|
args:
|
||||||
{{- toYaml .Values.services.apps.args | nindent 10 }}
|
{{- toYaml .Values.services.apps.args | nindent 10 }}
|
||||||
{{ end }}
|
{{ end }}
|
||||||
|
{{ if .Values.services.apps.extraVolumeMounts }}
|
||||||
|
volumeMounts:
|
||||||
|
{{- toYaml .Values.services.apps.extraVolumeMounts | nindent 10 }}
|
||||||
|
{{- end }}
|
||||||
{{- if .Values.services.apps.extraContainers }}
|
{{- if .Values.services.apps.extraContainers }}
|
||||||
{{- toYaml .Values.services.apps.extraContainers | nindent 6 }}
|
{{- toYaml .Values.services.apps.extraContainers | nindent 6 }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
|
@ -261,4 +265,8 @@ spec:
|
||||||
- name: ndots
|
- name: ndots
|
||||||
value: {{ .Values.services.apps.ndots | quote }}
|
value: {{ .Values.services.apps.ndots | quote }}
|
||||||
{{ end }}
|
{{ end }}
|
||||||
|
{{ if .Values.services.apps.extraVolumes }}
|
||||||
|
volumes:
|
||||||
|
{{- toYaml .Values.services.apps.extraVolumes | nindent 6 }}
|
||||||
|
{{- end }}
|
||||||
status: {}
|
status: {}
|
||||||
|
|
|
@ -235,6 +235,10 @@ spec:
|
||||||
args:
|
args:
|
||||||
{{- toYaml .Values.services.automationWorkers.args | nindent 10 }}
|
{{- toYaml .Values.services.automationWorkers.args | nindent 10 }}
|
||||||
{{ end }}
|
{{ end }}
|
||||||
|
{{ if .Values.services.automationWorkers.extraVolumeMounts }}
|
||||||
|
volumeMounts:
|
||||||
|
{{- toYaml .Values.services.automationWorkers.extraVolumeMounts | nindent 10 }}
|
||||||
|
{{ end }}
|
||||||
{{- if .Values.services.automationWorkers.extraContainers }}
|
{{- if .Values.services.automationWorkers.extraContainers }}
|
||||||
{{- toYaml .Values.services.automationWorkers.extraContainers | nindent 6 }}
|
{{- toYaml .Values.services.automationWorkers.extraContainers | nindent 6 }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
|
@ -261,5 +265,9 @@ spec:
|
||||||
- name: ndots
|
- name: ndots
|
||||||
value: {{ .Values.services.automationWorkers.ndots | quote }}
|
value: {{ .Values.services.automationWorkers.ndots | quote }}
|
||||||
{{ end }}
|
{{ end }}
|
||||||
|
{{ if .Values.services.automationWorkers.extraVolumes }}
|
||||||
|
volumes:
|
||||||
|
{{- toYaml .Values.services.automationWorkers.extraVolumes | nindent 8 }}
|
||||||
|
{{ end }}
|
||||||
status: {}
|
status: {}
|
||||||
{{- end }}
|
{{- end }}
|
|
@ -54,6 +54,9 @@ spec:
|
||||||
volumeMounts:
|
volumeMounts:
|
||||||
- mountPath: /data
|
- mountPath: /data
|
||||||
name: minio-data
|
name: minio-data
|
||||||
|
{{ if .Values.services.objectStore.extraVolumeMounts }}
|
||||||
|
{{- toYaml .Values.services.objectStore.extraVolumeMounts | nindent 8 }}
|
||||||
|
{{- end }}
|
||||||
{{- if .Values.services.objectStore.extraContainers }}
|
{{- if .Values.services.objectStore.extraContainers }}
|
||||||
{{- toYaml .Values.services.objectStore.extraContainers | nindent 6 }}
|
{{- toYaml .Values.services.objectStore.extraContainers | nindent 6 }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
|
@ -78,5 +81,8 @@ spec:
|
||||||
- name: minio-data
|
- name: minio-data
|
||||||
persistentVolumeClaim:
|
persistentVolumeClaim:
|
||||||
claimName: minio-data
|
claimName: minio-data
|
||||||
|
{{ if .Values.services.objectStore.extraVolumes }}
|
||||||
|
{{- toYaml .Values.services.objectStore.extraVolumes | nindent 6 }}
|
||||||
|
{{- end }}
|
||||||
status: {}
|
status: {}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
|
|
|
@ -82,6 +82,10 @@ spec:
|
||||||
resources:
|
resources:
|
||||||
{{- toYaml . | nindent 10 }}
|
{{- toYaml . | nindent 10 }}
|
||||||
{{ end }}
|
{{ end }}
|
||||||
|
{{ if .Values.services.proxy.extraVolumeMounts }}
|
||||||
|
volumeMounts:
|
||||||
|
{{- toYaml .Values.services.proxy.extraVolumeMounts | nindent 8 }}
|
||||||
|
{{- end }}
|
||||||
{{- if .Values.services.proxy.extraContainers }}
|
{{- if .Values.services.proxy.extraContainers }}
|
||||||
{{- toYaml .Values.services.proxy.extraContainers | nindent 6 }}
|
{{- toYaml .Values.services.proxy.extraContainers | nindent 6 }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
|
@ -110,7 +114,10 @@ spec:
|
||||||
args:
|
args:
|
||||||
{{- toYaml .Values.services.proxy.args | nindent 8 }}
|
{{- toYaml .Values.services.proxy.args | nindent 8 }}
|
||||||
{{ end }}
|
{{ end }}
|
||||||
|
{{ if .Values.services.proxy.extraVolumes }}
|
||||||
volumes:
|
volumes:
|
||||||
|
{{- toYaml .Values.services.proxy.extraVolumes | nindent 6 }}
|
||||||
|
{{ end }}
|
||||||
{{ if .Values.services.proxy.ndots }}
|
{{ if .Values.services.proxy.ndots }}
|
||||||
dnsConfig:
|
dnsConfig:
|
||||||
options:
|
options:
|
||||||
|
|
|
@ -22,7 +22,7 @@ spec:
|
||||||
- redis-server
|
- redis-server
|
||||||
- --requirepass
|
- --requirepass
|
||||||
- {{ .Values.services.redis.password }}
|
- {{ .Values.services.redis.password }}
|
||||||
image: redis
|
image: {{ .Values.services.redis.image }}
|
||||||
imagePullPolicy: ""
|
imagePullPolicy: ""
|
||||||
name: redis-service
|
name: redis-service
|
||||||
ports:
|
ports:
|
||||||
|
@ -34,6 +34,9 @@ spec:
|
||||||
volumeMounts:
|
volumeMounts:
|
||||||
- mountPath: /data
|
- mountPath: /data
|
||||||
name: redis-data
|
name: redis-data
|
||||||
|
{{ if .Values.services.redis.extraVolumeMounts }}
|
||||||
|
{{- toYaml .Values.services.redis.extraVolumeMounts | nindent 8 }}
|
||||||
|
{{- end }}
|
||||||
{{- if .Values.services.redis.extraContainers }}
|
{{- if .Values.services.redis.extraContainers }}
|
||||||
{{- toYaml .Values.services.redis.extraContainers | nindent 6 }}
|
{{- toYaml .Values.services.redis.extraContainers | nindent 6 }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
|
@ -58,6 +61,9 @@ spec:
|
||||||
- name: redis-data
|
- name: redis-data
|
||||||
persistentVolumeClaim:
|
persistentVolumeClaim:
|
||||||
claimName: redis-data
|
claimName: redis-data
|
||||||
|
{{ if .Values.services.redis.extraVolumes }}
|
||||||
|
{{- toYaml .Values.services.redis.extraVolumes | nindent 6 }}
|
||||||
|
{{- end }}
|
||||||
|
|
||||||
status: {}
|
status: {}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
|
|
|
@ -221,6 +221,10 @@ spec:
|
||||||
args:
|
args:
|
||||||
{{- toYaml .Values.services.worker.args | nindent 10 }}
|
{{- toYaml .Values.services.worker.args | nindent 10 }}
|
||||||
{{ end }}
|
{{ end }}
|
||||||
|
{{ if .Values.services.worker.extraVolumeMounts }}
|
||||||
|
volumeMounts:
|
||||||
|
{{- toYaml .Values.services.worker.extraVolumeMounts | nindent 10 }}
|
||||||
|
{{- end }}
|
||||||
{{- if .Values.services.worker.extraContainers }}
|
{{- if .Values.services.worker.extraContainers }}
|
||||||
{{- toYaml .Values.services.worker.extraContainers | nindent 6 }}
|
{{- toYaml .Values.services.worker.extraContainers | nindent 6 }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
|
@ -247,4 +251,8 @@ spec:
|
||||||
- name: ndots
|
- name: ndots
|
||||||
value: {{ .Values.services.worker.ndots | quote }}
|
value: {{ .Values.services.worker.ndots | quote }}
|
||||||
{{ end }}
|
{{ end }}
|
||||||
|
{{ if .Values.services.worker.extraVolumes }}
|
||||||
|
volumes:
|
||||||
|
{{- toYaml .Values.services.worker.extraVolumes | nindent 6 }}
|
||||||
|
{{- end }}
|
||||||
status: {}
|
status: {}
|
||||||
|
|
|
@ -211,6 +211,16 @@ services:
|
||||||
# - name: my-sidecar
|
# - name: my-sidecar
|
||||||
# image: myimage:latest
|
# image: myimage:latest
|
||||||
|
|
||||||
|
# -- Additional volumeMounts to the main proxy container.
|
||||||
|
extraVolumeMounts: []
|
||||||
|
# - name: my-volume
|
||||||
|
# mountPath: /path/to/mount
|
||||||
|
|
||||||
|
# -- Additional volumes to the proxy pod.
|
||||||
|
extraVolumes: []
|
||||||
|
# - name: my-volume
|
||||||
|
# emptyDir: {}
|
||||||
|
|
||||||
apps:
|
apps:
|
||||||
# @ignore (you shouldn't need to change this)
|
# @ignore (you shouldn't need to change this)
|
||||||
port: 4002
|
port: 4002
|
||||||
|
@ -283,6 +293,16 @@ services:
|
||||||
# - name: my-sidecar
|
# - name: my-sidecar
|
||||||
# image: myimage:latest
|
# image: myimage:latest
|
||||||
|
|
||||||
|
# -- Additional volumeMounts to the main apps container.
|
||||||
|
extraVolumeMounts: []
|
||||||
|
# - name: my-volume
|
||||||
|
# mountPath: /path/to/mount
|
||||||
|
|
||||||
|
# -- Additional volumes to the apps pod.
|
||||||
|
extraVolumes: []
|
||||||
|
# - name: my-volume
|
||||||
|
# emptyDir: {}
|
||||||
|
|
||||||
automationWorkers:
|
automationWorkers:
|
||||||
# -- Whether or not to enable the automation worker service. If you disable this,
|
# -- Whether or not to enable the automation worker service. If you disable this,
|
||||||
# automations will be processed by the apps service.
|
# automations will be processed by the apps service.
|
||||||
|
@ -359,6 +379,16 @@ services:
|
||||||
# - name: my-sidecar
|
# - name: my-sidecar
|
||||||
# image: myimage:latest
|
# image: myimage:latest
|
||||||
|
|
||||||
|
# -- Additional volumeMounts to the main automationWorkers container.
|
||||||
|
extraVolumeMounts: []
|
||||||
|
# - name: my-volume
|
||||||
|
# mountPath: /path/to/mount
|
||||||
|
|
||||||
|
# -- Additional volumes to the automationWorkers pod.
|
||||||
|
extraVolumes: []
|
||||||
|
# - name: my-volume
|
||||||
|
# emptyDir: {}
|
||||||
|
|
||||||
worker:
|
worker:
|
||||||
# @ignore (you shouldn't need to change this)
|
# @ignore (you shouldn't need to change this)
|
||||||
port: 4003
|
port: 4003
|
||||||
|
@ -431,6 +461,16 @@ services:
|
||||||
# - name: my-sidecar
|
# - name: my-sidecar
|
||||||
# image: myimage:latest
|
# image: myimage:latest
|
||||||
|
|
||||||
|
# -- Additional volumeMounts to the main worker container.
|
||||||
|
extraVolumeMounts: []
|
||||||
|
# - name: my-volume
|
||||||
|
# mountPath: /path/to/mount
|
||||||
|
|
||||||
|
# -- Additional volumes to the worker pod.
|
||||||
|
extraVolumes: []
|
||||||
|
# - name: my-volume
|
||||||
|
# emptyDir: {}
|
||||||
|
|
||||||
couchdb:
|
couchdb:
|
||||||
# -- Whether or not to spin up a CouchDB instance in your cluster. True by
|
# -- Whether or not to spin up a CouchDB instance in your cluster. True by
|
||||||
# default, and the configuration for the CouchDB instance is under the
|
# default, and the configuration for the CouchDB instance is under the
|
||||||
|
@ -456,6 +496,8 @@ services:
|
||||||
resources: {}
|
resources: {}
|
||||||
|
|
||||||
redis:
|
redis:
|
||||||
|
# -- The Redis image to use.
|
||||||
|
image: redis
|
||||||
# -- Whether or not to deploy a Redis pod into your cluster.
|
# -- Whether or not to deploy a Redis pod into your cluster.
|
||||||
enabled: true
|
enabled: true
|
||||||
# -- Port to expose Redis on.
|
# -- Port to expose Redis on.
|
||||||
|
@ -484,6 +526,16 @@ services:
|
||||||
# - name: my-sidecar
|
# - name: my-sidecar
|
||||||
# image: myimage:latest
|
# image: myimage:latest
|
||||||
|
|
||||||
|
# -- Additional volumeMounts to the main redis container.
|
||||||
|
extraVolumeMounts: []
|
||||||
|
# - name: my-volume
|
||||||
|
# mountPath: /path/to/mount
|
||||||
|
|
||||||
|
# -- Additional volumes to the redis pod.
|
||||||
|
extraVolumes: []
|
||||||
|
# - name: my-volume
|
||||||
|
# emptyDir: {}
|
||||||
|
|
||||||
objectStore:
|
objectStore:
|
||||||
# -- Set to false if using another object store, such as S3. You will need
|
# -- Set to false if using another object store, such as S3. You will need
|
||||||
# to set `services.objectStore.url` to point to your bucket if you do this.
|
# to set `services.objectStore.url` to point to your bucket if you do this.
|
||||||
|
@ -530,6 +582,16 @@ services:
|
||||||
# - name: my-sidecar
|
# - name: my-sidecar
|
||||||
# image: myimage:latest
|
# image: myimage:latest
|
||||||
|
|
||||||
|
# -- Additional volumeMounts to the main objectStore container.
|
||||||
|
extraVolumeMounts: []
|
||||||
|
# - name: my-volume
|
||||||
|
# mountPath: /path/to/mount
|
||||||
|
|
||||||
|
# -- Additional volumes to the objectStore pod.
|
||||||
|
extraVolumes: []
|
||||||
|
# - name: my-volume
|
||||||
|
# emptyDir: {}
|
||||||
|
|
||||||
# Override values in couchDB subchart. We're only specifying the values we're changing.
|
# Override values in couchDB subchart. We're only specifying the values we're changing.
|
||||||
# If you want to see all of the available values, see:
|
# If you want to see all of the available values, see:
|
||||||
# https://github.com/apache/couchdb-helm/tree/couchdb-4.3.0/couchdb
|
# https://github.com/apache/couchdb-helm/tree/couchdb-4.3.0/couchdb
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
{
|
{
|
||||||
"version": "2.23.4",
|
"version": "2.23.5",
|
||||||
"npmClient": "yarn",
|
"npmClient": "yarn",
|
||||||
"packages": [
|
"packages": [
|
||||||
"packages/*",
|
"packages/*",
|
||||||
|
|
|
@ -56,6 +56,7 @@
|
||||||
"dev:noserver": "yarn run kill-builder && lerna run --stream dev:stack:up --ignore @budibase/account-portal-server && lerna run --stream dev --ignore @budibase/backend-core --ignore @budibase/server --ignore @budibase/worker --ignore=@budibase/account-portal-ui --ignore @budibase/account-portal-server",
|
"dev:noserver": "yarn run kill-builder && lerna run --stream dev:stack:up --ignore @budibase/account-portal-server && lerna run --stream dev --ignore @budibase/backend-core --ignore @budibase/server --ignore @budibase/worker --ignore=@budibase/account-portal-ui --ignore @budibase/account-portal-server",
|
||||||
"dev:server": "yarn run kill-server && lerna run --stream dev --scope @budibase/worker --scope @budibase/server",
|
"dev:server": "yarn run kill-server && lerna run --stream dev --scope @budibase/worker --scope @budibase/server",
|
||||||
"dev:accountportal": "yarn kill-accountportal && lerna run dev --stream --scope @budibase/account-portal-ui --scope @budibase/account-portal-server",
|
"dev:accountportal": "yarn kill-accountportal && lerna run dev --stream --scope @budibase/account-portal-ui --scope @budibase/account-portal-server",
|
||||||
|
"dev:camunda": "./scripts/deploy-camunda.sh",
|
||||||
"dev:all": "yarn run kill-all && lerna run --stream dev",
|
"dev:all": "yarn run kill-all && lerna run --stream dev",
|
||||||
"dev:built": "yarn run kill-all && cd packages/server && yarn dev:stack:up && cd ../../ && lerna run --stream dev:built",
|
"dev:built": "yarn run kill-all && cd packages/server && yarn dev:stack:up && cd ../../ && lerna run --stream dev:built",
|
||||||
"dev:docker": "yarn build --scope @budibase/server --scope @budibase/worker && docker-compose -f hosting/docker-compose.build.yaml -f hosting/docker-compose.dev.yaml --env-file hosting/.env up --build --scale proxy-service=0",
|
"dev:docker": "yarn build --scope @budibase/server --scope @budibase/worker && docker-compose -f hosting/docker-compose.build.yaml -f hosting/docker-compose.dev.yaml --env-file hosting/.env up --build --scale proxy-service=0",
|
||||||
|
|
|
@ -1 +1 @@
|
||||||
Subproject commit a0ee9cad8cefb8f9f40228705711be174f018fa9
|
Subproject commit bd0e01d639ec3b2547e7c859a1c43b622dce8344
|
|
@ -320,6 +320,7 @@ async function performAppCreate(ctx: UserCtx<CreateAppRequest, App>) {
|
||||||
"theme",
|
"theme",
|
||||||
"customTheme",
|
"customTheme",
|
||||||
"icon",
|
"icon",
|
||||||
|
"snippets",
|
||||||
]
|
]
|
||||||
keys.forEach(key => {
|
keys.forEach(key => {
|
||||||
if (existing[key]) {
|
if (existing[key]) {
|
||||||
|
|
|
@ -8,6 +8,8 @@ import {
|
||||||
FieldType,
|
FieldType,
|
||||||
RowSearchParams,
|
RowSearchParams,
|
||||||
SearchFilters,
|
SearchFilters,
|
||||||
|
SortOrder,
|
||||||
|
SortType,
|
||||||
Table,
|
Table,
|
||||||
TableSchema,
|
TableSchema,
|
||||||
} from "@budibase/types"
|
} from "@budibase/types"
|
||||||
|
@ -62,7 +64,32 @@ describe.each([
|
||||||
class SearchAssertion {
|
class SearchAssertion {
|
||||||
constructor(private readonly query: RowSearchParams) {}
|
constructor(private readonly query: RowSearchParams) {}
|
||||||
|
|
||||||
async toFind(expectedRows: any[]) {
|
// Asserts that the query returns rows matching exactly the set of rows
|
||||||
|
// passed in. The order of the rows matters. Rows returned in an order
|
||||||
|
// different to the one passed in will cause the assertion to fail. Extra
|
||||||
|
// rows returned by the query will also cause the assertion to fail.
|
||||||
|
async toMatchExactly(expectedRows: any[]) {
|
||||||
|
const { rows: foundRows } = await config.api.row.search(table._id!, {
|
||||||
|
...this.query,
|
||||||
|
tableId: table._id!,
|
||||||
|
})
|
||||||
|
|
||||||
|
// eslint-disable-next-line jest/no-standalone-expect
|
||||||
|
expect(foundRows).toHaveLength(expectedRows.length)
|
||||||
|
// eslint-disable-next-line jest/no-standalone-expect
|
||||||
|
expect(foundRows).toEqual(
|
||||||
|
expectedRows.map((expectedRow: any) =>
|
||||||
|
expect.objectContaining(
|
||||||
|
foundRows.find(foundRow => _.isMatch(foundRow, expectedRow))
|
||||||
|
)
|
||||||
|
)
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Asserts that the query returns rows matching exactly the set of rows
|
||||||
|
// passed in. The order of the rows is not important, but extra rows will
|
||||||
|
// cause the assertion to fail.
|
||||||
|
async toContainExactly(expectedRows: any[]) {
|
||||||
const { rows: foundRows } = await config.api.row.search(table._id!, {
|
const { rows: foundRows } = await config.api.row.search(table._id!, {
|
||||||
...this.query,
|
...this.query,
|
||||||
tableId: table._id!,
|
tableId: table._id!,
|
||||||
|
@ -82,8 +109,39 @@ describe.each([
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Asserts that the query returns rows matching the set of rows passed in.
|
||||||
|
// The order of the rows is not important. Extra rows will not cause the
|
||||||
|
// assertion to fail.
|
||||||
|
async toContain(expectedRows: any[]) {
|
||||||
|
const { rows: foundRows } = await config.api.row.search(table._id!, {
|
||||||
|
...this.query,
|
||||||
|
tableId: table._id!,
|
||||||
|
})
|
||||||
|
|
||||||
|
// eslint-disable-next-line jest/no-standalone-expect
|
||||||
|
expect(foundRows).toEqual(
|
||||||
|
expect.arrayContaining(
|
||||||
|
expectedRows.map((expectedRow: any) =>
|
||||||
|
expect.objectContaining(
|
||||||
|
foundRows.find(foundRow => _.isMatch(foundRow, expectedRow))
|
||||||
|
)
|
||||||
|
)
|
||||||
|
)
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
async toFindNothing() {
|
async toFindNothing() {
|
||||||
await this.toFind([])
|
await this.toContainExactly([])
|
||||||
|
}
|
||||||
|
|
||||||
|
async toHaveLength(length: number) {
|
||||||
|
const { rows: foundRows } = await config.api.row.search(table._id!, {
|
||||||
|
...this.query,
|
||||||
|
tableId: table._id!,
|
||||||
|
})
|
||||||
|
|
||||||
|
// eslint-disable-next-line jest/no-standalone-expect
|
||||||
|
expect(foundRows).toHaveLength(length)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -105,28 +163,33 @@ describe.each([
|
||||||
|
|
||||||
describe("misc", () => {
|
describe("misc", () => {
|
||||||
it("should return all if no query is passed", () =>
|
it("should return all if no query is passed", () =>
|
||||||
expectSearch({} as RowSearchParams).toFind([
|
expectSearch({} as RowSearchParams).toContainExactly([
|
||||||
{ name: "foo" },
|
{ name: "foo" },
|
||||||
{ name: "bar" },
|
{ name: "bar" },
|
||||||
]))
|
]))
|
||||||
|
|
||||||
it("should return all if empty query is passed", () =>
|
it("should return all if empty query is passed", () =>
|
||||||
expectQuery({}).toFind([{ name: "foo" }, { name: "bar" }]))
|
expectQuery({}).toContainExactly([{ name: "foo" }, { name: "bar" }]))
|
||||||
|
|
||||||
it("should return all if onEmptyFilter is RETURN_ALL", () =>
|
it("should return all if onEmptyFilter is RETURN_ALL", () =>
|
||||||
expectQuery({
|
expectQuery({
|
||||||
onEmptyFilter: EmptyFilterOption.RETURN_ALL,
|
onEmptyFilter: EmptyFilterOption.RETURN_ALL,
|
||||||
}).toFind([{ name: "foo" }, { name: "bar" }]))
|
}).toContainExactly([{ name: "foo" }, { name: "bar" }]))
|
||||||
|
|
||||||
it("should return nothing if onEmptyFilter is RETURN_NONE", () =>
|
it("should return nothing if onEmptyFilter is RETURN_NONE", () =>
|
||||||
expectQuery({
|
expectQuery({
|
||||||
onEmptyFilter: EmptyFilterOption.RETURN_NONE,
|
onEmptyFilter: EmptyFilterOption.RETURN_NONE,
|
||||||
}).toFindNothing())
|
}).toFindNothing())
|
||||||
|
|
||||||
|
it("should respect limit", () =>
|
||||||
|
expectSearch({ limit: 1, paginate: true, query: {} }).toHaveLength(1))
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("equal", () => {
|
describe("equal", () => {
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({ equal: { name: "foo" } }).toFind([{ name: "foo" }]))
|
expectQuery({ equal: { name: "foo" } }).toContainExactly([
|
||||||
|
{ name: "foo" },
|
||||||
|
]))
|
||||||
|
|
||||||
it("fails to find nonexistent row", () =>
|
it("fails to find nonexistent row", () =>
|
||||||
expectQuery({ equal: { name: "none" } }).toFindNothing())
|
expectQuery({ equal: { name: "none" } }).toFindNothing())
|
||||||
|
@ -134,15 +197,21 @@ describe.each([
|
||||||
|
|
||||||
describe("notEqual", () => {
|
describe("notEqual", () => {
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({ notEqual: { name: "foo" } }).toFind([{ name: "bar" }]))
|
expectQuery({ notEqual: { name: "foo" } }).toContainExactly([
|
||||||
|
{ name: "bar" },
|
||||||
|
]))
|
||||||
|
|
||||||
it("fails to find nonexistent row", () =>
|
it("fails to find nonexistent row", () =>
|
||||||
expectQuery({ notEqual: { name: "bar" } }).toFind([{ name: "foo" }]))
|
expectQuery({ notEqual: { name: "bar" } }).toContainExactly([
|
||||||
|
{ name: "foo" },
|
||||||
|
]))
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("oneOf", () => {
|
describe("oneOf", () => {
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({ oneOf: { name: ["foo"] } }).toFind([{ name: "foo" }]))
|
expectQuery({ oneOf: { name: ["foo"] } }).toContainExactly([
|
||||||
|
{ name: "foo" },
|
||||||
|
]))
|
||||||
|
|
||||||
it("fails to find nonexistent row", () =>
|
it("fails to find nonexistent row", () =>
|
||||||
expectQuery({ oneOf: { name: ["none"] } }).toFindNothing())
|
expectQuery({ oneOf: { name: ["none"] } }).toFindNothing())
|
||||||
|
@ -150,11 +219,69 @@ describe.each([
|
||||||
|
|
||||||
describe("fuzzy", () => {
|
describe("fuzzy", () => {
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({ fuzzy: { name: "oo" } }).toFind([{ name: "foo" }]))
|
expectQuery({ fuzzy: { name: "oo" } }).toContainExactly([
|
||||||
|
{ name: "foo" },
|
||||||
|
]))
|
||||||
|
|
||||||
it("fails to find nonexistent row", () =>
|
it("fails to find nonexistent row", () =>
|
||||||
expectQuery({ fuzzy: { name: "none" } }).toFindNothing())
|
expectQuery({ fuzzy: { name: "none" } }).toFindNothing())
|
||||||
})
|
})
|
||||||
|
|
||||||
|
describe("range", () => {
|
||||||
|
it("successfully finds multiple rows", () =>
|
||||||
|
expectQuery({
|
||||||
|
range: { name: { low: "a", high: "z" } },
|
||||||
|
}).toContainExactly([{ name: "bar" }, { name: "foo" }]))
|
||||||
|
|
||||||
|
it("successfully finds a row with a high bound", () =>
|
||||||
|
expectQuery({
|
||||||
|
range: { name: { low: "a", high: "c" } },
|
||||||
|
}).toContainExactly([{ name: "bar" }]))
|
||||||
|
|
||||||
|
it("successfully finds a row with a low bound", () =>
|
||||||
|
expectQuery({
|
||||||
|
range: { name: { low: "f", high: "z" } },
|
||||||
|
}).toContainExactly([{ name: "foo" }]))
|
||||||
|
|
||||||
|
it("successfully finds no rows", () =>
|
||||||
|
expectQuery({
|
||||||
|
range: { name: { low: "g", high: "h" } },
|
||||||
|
}).toFindNothing())
|
||||||
|
})
|
||||||
|
|
||||||
|
describe("sort", () => {
|
||||||
|
it("sorts ascending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "name",
|
||||||
|
sortOrder: SortOrder.ASCENDING,
|
||||||
|
}).toMatchExactly([{ name: "bar" }, { name: "foo" }]))
|
||||||
|
|
||||||
|
it("sorts descending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "name",
|
||||||
|
sortOrder: SortOrder.DESCENDING,
|
||||||
|
}).toMatchExactly([{ name: "foo" }, { name: "bar" }]))
|
||||||
|
|
||||||
|
describe("sortType STRING", () => {
|
||||||
|
it("sorts ascending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "name",
|
||||||
|
sortType: SortType.STRING,
|
||||||
|
sortOrder: SortOrder.ASCENDING,
|
||||||
|
}).toMatchExactly([{ name: "bar" }, { name: "foo" }]))
|
||||||
|
|
||||||
|
it("sorts descending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "name",
|
||||||
|
sortType: SortType.STRING,
|
||||||
|
sortOrder: SortOrder.DESCENDING,
|
||||||
|
}).toMatchExactly([{ name: "foo" }, { name: "bar" }]))
|
||||||
|
})
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("numbers", () => {
|
describe("numbers", () => {
|
||||||
|
@ -167,7 +294,7 @@ describe.each([
|
||||||
|
|
||||||
describe("equal", () => {
|
describe("equal", () => {
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({ equal: { age: 1 } }).toFind([{ age: 1 }]))
|
expectQuery({ equal: { age: 1 } }).toContainExactly([{ age: 1 }]))
|
||||||
|
|
||||||
it("fails to find nonexistent row", () =>
|
it("fails to find nonexistent row", () =>
|
||||||
expectQuery({ equal: { age: 2 } }).toFindNothing())
|
expectQuery({ equal: { age: 2 } }).toFindNothing())
|
||||||
|
@ -175,15 +302,15 @@ describe.each([
|
||||||
|
|
||||||
describe("notEqual", () => {
|
describe("notEqual", () => {
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({ notEqual: { age: 1 } }).toFind([{ age: 10 }]))
|
expectQuery({ notEqual: { age: 1 } }).toContainExactly([{ age: 10 }]))
|
||||||
|
|
||||||
it("fails to find nonexistent row", () =>
|
it("fails to find nonexistent row", () =>
|
||||||
expectQuery({ notEqual: { age: 10 } }).toFind([{ age: 1 }]))
|
expectQuery({ notEqual: { age: 10 } }).toContainExactly([{ age: 1 }]))
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("oneOf", () => {
|
describe("oneOf", () => {
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({ oneOf: { age: [1] } }).toFind([{ age: 1 }]))
|
expectQuery({ oneOf: { age: [1] } }).toContainExactly([{ age: 1 }]))
|
||||||
|
|
||||||
it("fails to find nonexistent row", () =>
|
it("fails to find nonexistent row", () =>
|
||||||
expectQuery({ oneOf: { age: [2] } }).toFindNothing())
|
expectQuery({ oneOf: { age: [2] } }).toFindNothing())
|
||||||
|
@ -193,17 +320,56 @@ describe.each([
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({
|
expectQuery({
|
||||||
range: { age: { low: 1, high: 5 } },
|
range: { age: { low: 1, high: 5 } },
|
||||||
}).toFind([{ age: 1 }]))
|
}).toContainExactly([{ age: 1 }]))
|
||||||
|
|
||||||
it("successfully finds multiple rows", () =>
|
it("successfully finds multiple rows", () =>
|
||||||
expectQuery({
|
expectQuery({
|
||||||
range: { age: { low: 1, high: 10 } },
|
range: { age: { low: 1, high: 10 } },
|
||||||
}).toFind([{ age: 1 }, { age: 10 }]))
|
}).toContainExactly([{ age: 1 }, { age: 10 }]))
|
||||||
|
|
||||||
it("successfully finds a row with a high bound", () =>
|
it("successfully finds a row with a high bound", () =>
|
||||||
expectQuery({
|
expectQuery({
|
||||||
range: { age: { low: 5, high: 10 } },
|
range: { age: { low: 5, high: 10 } },
|
||||||
}).toFind([{ age: 10 }]))
|
}).toContainExactly([{ age: 10 }]))
|
||||||
|
|
||||||
|
it("successfully finds no rows", () =>
|
||||||
|
expectQuery({
|
||||||
|
range: { age: { low: 5, high: 9 } },
|
||||||
|
}).toFindNothing())
|
||||||
|
})
|
||||||
|
|
||||||
|
describe("sort", () => {
|
||||||
|
it("sorts ascending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "age",
|
||||||
|
sortOrder: SortOrder.ASCENDING,
|
||||||
|
}).toMatchExactly([{ age: 1 }, { age: 10 }]))
|
||||||
|
|
||||||
|
it("sorts descending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "age",
|
||||||
|
sortOrder: SortOrder.DESCENDING,
|
||||||
|
}).toMatchExactly([{ age: 10 }, { age: 1 }]))
|
||||||
|
})
|
||||||
|
|
||||||
|
describe("sortType NUMBER", () => {
|
||||||
|
it("sorts ascending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "age",
|
||||||
|
sortType: SortType.NUMBER,
|
||||||
|
sortOrder: SortOrder.ASCENDING,
|
||||||
|
}).toMatchExactly([{ age: 1 }, { age: 10 }]))
|
||||||
|
|
||||||
|
it("sorts descending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "age",
|
||||||
|
sortType: SortType.NUMBER,
|
||||||
|
sortOrder: SortOrder.DESCENDING,
|
||||||
|
}).toMatchExactly([{ age: 10 }, { age: 1 }]))
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
|
@ -211,6 +377,7 @@ describe.each([
|
||||||
const JAN_1ST = "2020-01-01T00:00:00.000Z"
|
const JAN_1ST = "2020-01-01T00:00:00.000Z"
|
||||||
const JAN_2ND = "2020-01-02T00:00:00.000Z"
|
const JAN_2ND = "2020-01-02T00:00:00.000Z"
|
||||||
const JAN_5TH = "2020-01-05T00:00:00.000Z"
|
const JAN_5TH = "2020-01-05T00:00:00.000Z"
|
||||||
|
const JAN_9TH = "2020-01-09T00:00:00.000Z"
|
||||||
const JAN_10TH = "2020-01-10T00:00:00.000Z"
|
const JAN_10TH = "2020-01-10T00:00:00.000Z"
|
||||||
|
|
||||||
beforeAll(async () => {
|
beforeAll(async () => {
|
||||||
|
@ -223,7 +390,9 @@ describe.each([
|
||||||
|
|
||||||
describe("equal", () => {
|
describe("equal", () => {
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({ equal: { dob: JAN_1ST } }).toFind([{ dob: JAN_1ST }]))
|
expectQuery({ equal: { dob: JAN_1ST } }).toContainExactly([
|
||||||
|
{ dob: JAN_1ST },
|
||||||
|
]))
|
||||||
|
|
||||||
it("fails to find nonexistent row", () =>
|
it("fails to find nonexistent row", () =>
|
||||||
expectQuery({ equal: { dob: JAN_2ND } }).toFindNothing())
|
expectQuery({ equal: { dob: JAN_2ND } }).toFindNothing())
|
||||||
|
@ -231,15 +400,21 @@ describe.each([
|
||||||
|
|
||||||
describe("notEqual", () => {
|
describe("notEqual", () => {
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({ notEqual: { dob: JAN_1ST } }).toFind([{ dob: JAN_10TH }]))
|
expectQuery({ notEqual: { dob: JAN_1ST } }).toContainExactly([
|
||||||
|
{ dob: JAN_10TH },
|
||||||
|
]))
|
||||||
|
|
||||||
it("fails to find nonexistent row", () =>
|
it("fails to find nonexistent row", () =>
|
||||||
expectQuery({ notEqual: { dob: JAN_10TH } }).toFind([{ dob: JAN_1ST }]))
|
expectQuery({ notEqual: { dob: JAN_10TH } }).toContainExactly([
|
||||||
|
{ dob: JAN_1ST },
|
||||||
|
]))
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("oneOf", () => {
|
describe("oneOf", () => {
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({ oneOf: { dob: [JAN_1ST] } }).toFind([{ dob: JAN_1ST }]))
|
expectQuery({ oneOf: { dob: [JAN_1ST] } }).toContainExactly([
|
||||||
|
{ dob: JAN_1ST },
|
||||||
|
]))
|
||||||
|
|
||||||
it("fails to find nonexistent row", () =>
|
it("fails to find nonexistent row", () =>
|
||||||
expectQuery({ oneOf: { dob: [JAN_2ND] } }).toFindNothing())
|
expectQuery({ oneOf: { dob: [JAN_2ND] } }).toFindNothing())
|
||||||
|
@ -249,17 +424,130 @@ describe.each([
|
||||||
it("successfully finds a row", () =>
|
it("successfully finds a row", () =>
|
||||||
expectQuery({
|
expectQuery({
|
||||||
range: { dob: { low: JAN_1ST, high: JAN_5TH } },
|
range: { dob: { low: JAN_1ST, high: JAN_5TH } },
|
||||||
}).toFind([{ dob: JAN_1ST }]))
|
}).toContainExactly([{ dob: JAN_1ST }]))
|
||||||
|
|
||||||
it("successfully finds multiple rows", () =>
|
it("successfully finds multiple rows", () =>
|
||||||
expectQuery({
|
expectQuery({
|
||||||
range: { dob: { low: JAN_1ST, high: JAN_10TH } },
|
range: { dob: { low: JAN_1ST, high: JAN_10TH } },
|
||||||
}).toFind([{ dob: JAN_1ST }, { dob: JAN_10TH }]))
|
}).toContainExactly([{ dob: JAN_1ST }, { dob: JAN_10TH }]))
|
||||||
|
|
||||||
it("successfully finds a row with a high bound", () =>
|
it("successfully finds a row with a high bound", () =>
|
||||||
expectQuery({
|
expectQuery({
|
||||||
range: { dob: { low: JAN_5TH, high: JAN_10TH } },
|
range: { dob: { low: JAN_5TH, high: JAN_10TH } },
|
||||||
}).toFind([{ dob: JAN_10TH }]))
|
}).toContainExactly([{ dob: JAN_10TH }]))
|
||||||
|
|
||||||
|
it("successfully finds no rows", () =>
|
||||||
|
expectQuery({
|
||||||
|
range: { dob: { low: JAN_5TH, high: JAN_9TH } },
|
||||||
|
}).toFindNothing())
|
||||||
|
})
|
||||||
|
|
||||||
|
describe("sort", () => {
|
||||||
|
it("sorts ascending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "dob",
|
||||||
|
sortOrder: SortOrder.ASCENDING,
|
||||||
|
}).toMatchExactly([{ dob: JAN_1ST }, { dob: JAN_10TH }]))
|
||||||
|
|
||||||
|
it("sorts descending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "dob",
|
||||||
|
sortOrder: SortOrder.DESCENDING,
|
||||||
|
}).toMatchExactly([{ dob: JAN_10TH }, { dob: JAN_1ST }]))
|
||||||
|
|
||||||
|
describe("sortType STRING", () => {
|
||||||
|
it("sorts ascending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "dob",
|
||||||
|
sortType: SortType.STRING,
|
||||||
|
sortOrder: SortOrder.ASCENDING,
|
||||||
|
}).toMatchExactly([{ dob: JAN_1ST }, { dob: JAN_10TH }]))
|
||||||
|
|
||||||
|
it("sorts descending", () =>
|
||||||
|
expectSearch({
|
||||||
|
query: {},
|
||||||
|
sort: "dob",
|
||||||
|
sortType: SortType.STRING,
|
||||||
|
sortOrder: SortOrder.DESCENDING,
|
||||||
|
}).toMatchExactly([{ dob: JAN_10TH }, { dob: JAN_1ST }]))
|
||||||
|
})
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe("array of strings", () => {
|
||||||
|
beforeAll(async () => {
|
||||||
|
await createTable({
|
||||||
|
numbers: {
|
||||||
|
name: "numbers",
|
||||||
|
type: FieldType.ARRAY,
|
||||||
|
constraints: { inclusion: ["one", "two", "three"] },
|
||||||
|
},
|
||||||
|
})
|
||||||
|
await createRows([{ numbers: ["one", "two"] }, { numbers: ["three"] }])
|
||||||
|
})
|
||||||
|
|
||||||
|
describe("contains", () => {
|
||||||
|
it("successfully finds a row", () =>
|
||||||
|
expectQuery({ contains: { numbers: ["one"] } }).toContainExactly([
|
||||||
|
{ numbers: ["one", "two"] },
|
||||||
|
]))
|
||||||
|
|
||||||
|
it("fails to find nonexistent row", () =>
|
||||||
|
expectQuery({ contains: { numbers: ["none"] } }).toFindNothing())
|
||||||
|
|
||||||
|
it("fails to find row containing all", () =>
|
||||||
|
expectQuery({
|
||||||
|
contains: { numbers: ["one", "two", "three"] },
|
||||||
|
}).toFindNothing())
|
||||||
|
|
||||||
|
it("finds all with empty list", () =>
|
||||||
|
expectQuery({ contains: { numbers: [] } }).toContainExactly([
|
||||||
|
{ numbers: ["one", "two"] },
|
||||||
|
{ numbers: ["three"] },
|
||||||
|
]))
|
||||||
|
})
|
||||||
|
|
||||||
|
describe("notContains", () => {
|
||||||
|
it("successfully finds a row", () =>
|
||||||
|
expectQuery({ notContains: { numbers: ["one"] } }).toContainExactly([
|
||||||
|
{ numbers: ["three"] },
|
||||||
|
]))
|
||||||
|
|
||||||
|
it("fails to find nonexistent row", () =>
|
||||||
|
expectQuery({
|
||||||
|
notContains: { numbers: ["one", "two", "three"] },
|
||||||
|
}).toContainExactly([
|
||||||
|
{ numbers: ["one", "two"] },
|
||||||
|
{ numbers: ["three"] },
|
||||||
|
]))
|
||||||
|
|
||||||
|
it("finds all with empty list", () =>
|
||||||
|
expectQuery({ notContains: { numbers: [] } }).toContainExactly([
|
||||||
|
{ numbers: ["one", "two"] },
|
||||||
|
{ numbers: ["three"] },
|
||||||
|
]))
|
||||||
|
})
|
||||||
|
|
||||||
|
describe("containsAny", () => {
|
||||||
|
it("successfully finds rows", () =>
|
||||||
|
expectQuery({
|
||||||
|
containsAny: { numbers: ["one", "two", "three"] },
|
||||||
|
}).toContainExactly([
|
||||||
|
{ numbers: ["one", "two"] },
|
||||||
|
{ numbers: ["three"] },
|
||||||
|
]))
|
||||||
|
|
||||||
|
it("fails to find nonexistent row", () =>
|
||||||
|
expectQuery({ containsAny: { numbers: ["none"] } }).toFindNothing())
|
||||||
|
|
||||||
|
it("finds all with empty list", () =>
|
||||||
|
expectQuery({ containsAny: { numbers: [] } }).toContainExactly([
|
||||||
|
{ numbers: ["one", "two"] },
|
||||||
|
{ numbers: ["three"] },
|
||||||
|
]))
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
|
@ -20,6 +20,7 @@ export enum FilterTypes {
|
||||||
NOT_EMPTY = "notEmpty",
|
NOT_EMPTY = "notEmpty",
|
||||||
CONTAINS = "contains",
|
CONTAINS = "contains",
|
||||||
NOT_CONTAINS = "notContains",
|
NOT_CONTAINS = "notContains",
|
||||||
|
CONTAINS_ANY = "containsAny",
|
||||||
ONE_OF = "oneOf",
|
ONE_OF = "oneOf",
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -30,6 +31,7 @@ export const NoEmptyFilterStrings = [
|
||||||
FilterTypes.NOT_EQUAL,
|
FilterTypes.NOT_EQUAL,
|
||||||
FilterTypes.CONTAINS,
|
FilterTypes.CONTAINS,
|
||||||
FilterTypes.NOT_CONTAINS,
|
FilterTypes.NOT_CONTAINS,
|
||||||
|
FilterTypes.CONTAINS_ANY,
|
||||||
]
|
]
|
||||||
|
|
||||||
export const CanSwitchTypes = [
|
export const CanSwitchTypes = [
|
||||||
|
|
|
@ -233,6 +233,11 @@ class InternalBuilder {
|
||||||
(statement ? andOr : "") +
|
(statement ? andOr : "") +
|
||||||
`LOWER(${likeKey(this.client, key)}) LIKE ?`
|
`LOWER(${likeKey(this.client, key)}) LIKE ?`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (statement === "") {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
// @ts-ignore
|
// @ts-ignore
|
||||||
query = query[rawFnc](`${not}(${statement})`, value)
|
query = query[rawFnc](`${not}(${statement})`, value)
|
||||||
})
|
})
|
||||||
|
|
|
@ -29,6 +29,10 @@ function pickApi(tableId: any) {
|
||||||
return internal
|
return internal
|
||||||
}
|
}
|
||||||
|
|
||||||
|
function isEmptyArray(value: any) {
|
||||||
|
return Array.isArray(value) && value.length === 0
|
||||||
|
}
|
||||||
|
|
||||||
// don't do a pure falsy check, as 0 is included
|
// don't do a pure falsy check, as 0 is included
|
||||||
// https://github.com/Budibase/budibase/issues/10118
|
// https://github.com/Budibase/budibase/issues/10118
|
||||||
export function removeEmptyFilters(filters: SearchFilters) {
|
export function removeEmptyFilters(filters: SearchFilters) {
|
||||||
|
@ -47,7 +51,7 @@ export function removeEmptyFilters(filters: SearchFilters) {
|
||||||
for (let [key, value] of Object.entries(
|
for (let [key, value] of Object.entries(
|
||||||
filters[filterType] as object
|
filters[filterType] as object
|
||||||
)) {
|
)) {
|
||||||
if (value == null || value === "") {
|
if (value == null || value === "" || isEmptyArray(value)) {
|
||||||
// @ts-ignore
|
// @ts-ignore
|
||||||
delete filters[filterField][key]
|
delete filters[filterField][key]
|
||||||
}
|
}
|
||||||
|
|
|
@ -132,7 +132,7 @@ export async function search(
|
||||||
type: "row",
|
type: "row",
|
||||||
}
|
}
|
||||||
|
|
||||||
if (params.sort && !params.sortType) {
|
if (params.sort) {
|
||||||
const sortField = table.schema[params.sort]
|
const sortField = table.schema[params.sort]
|
||||||
const sortType =
|
const sortType =
|
||||||
sortField.type === FieldType.NUMBER ? SortType.NUMBER : SortType.STRING
|
sortField.type === FieldType.NUMBER ? SortType.NUMBER : SortType.STRING
|
||||||
|
|
|
@ -102,6 +102,7 @@ export function isVerifiableSSOProvider(provider: AccountSSOProvider): boolean {
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface AccountSSO {
|
export interface AccountSSO {
|
||||||
|
ssoId?: string
|
||||||
provider: AccountSSOProvider
|
provider: AccountSSOProvider
|
||||||
providerType: AccountSSOProviderType
|
providerType: AccountSSOProviderType
|
||||||
oauth2?: OAuthTokens
|
oauth2?: OAuthTokens
|
||||||
|
|
|
@ -1,22 +1,111 @@
|
||||||
import { Document } from "../document"
|
import { Document } from "../document"
|
||||||
|
|
||||||
export enum FieldType {
|
export enum FieldType {
|
||||||
|
/**
|
||||||
|
* a primitive type, stores a string, called Text within Budibase. This is one of the default
|
||||||
|
* types of Budibase, if an external type is not fully understood, we will treat it as text.
|
||||||
|
*/
|
||||||
STRING = "string",
|
STRING = "string",
|
||||||
|
/**
|
||||||
|
* similar to string type, called Long Form Text within Budibase. This is mainly a frontend
|
||||||
|
* orientated type which enables a larger text input area. This can also be used
|
||||||
|
* in conjunction with the 'useRichText' option to support a markdown editor/viewer.
|
||||||
|
*/
|
||||||
LONGFORM = "longform",
|
LONGFORM = "longform",
|
||||||
|
/**
|
||||||
|
* similar to string type, called Options within Budibase. This works very similarly to
|
||||||
|
* the string type within the backend, but is validated to a list of options. This will
|
||||||
|
* display a <select> input within the builder/client.
|
||||||
|
*/
|
||||||
OPTIONS = "options",
|
OPTIONS = "options",
|
||||||
|
/**
|
||||||
|
* a primitive type, stores a number, as a floating point, called Number within Budibase.
|
||||||
|
* this type will always represent numbers as reals/floating point - there is no integer only
|
||||||
|
* type within Budibase.
|
||||||
|
*/
|
||||||
NUMBER = "number",
|
NUMBER = "number",
|
||||||
|
/**
|
||||||
|
* a primitive type, stores a boolean, called Boolean within Budibase. This is often represented
|
||||||
|
* as a toggle or checkbox within forms/grids.
|
||||||
|
*/
|
||||||
BOOLEAN = "boolean",
|
BOOLEAN = "boolean",
|
||||||
|
/**
|
||||||
|
* a JSON type, this type is always an array of strings, called Multi-select within Budibase.
|
||||||
|
* This type can be compared to the options type, as it functions similarly, but allows picking
|
||||||
|
* multiple options rather than a single option.
|
||||||
|
*/
|
||||||
ARRAY = "array",
|
ARRAY = "array",
|
||||||
|
/**
|
||||||
|
* a string type, this is always a string when input/returned from the API, called Date/Time within
|
||||||
|
* Budibase. We utilise ISO date strings for representing dates, this type has a range of subtypes
|
||||||
|
* to restrict it to date only, time only and ignore timezone capabilities.
|
||||||
|
*/
|
||||||
DATETIME = "datetime",
|
DATETIME = "datetime",
|
||||||
|
/**
|
||||||
|
* a JSON type, an array of metadata about files held in object storage, called Attachment List within
|
||||||
|
* Budibase. To utilise this type there is an API for uploading files to Budibase, which returns metadata
|
||||||
|
* that can be stored against columns of this type. Currently this is not supported on external databases.
|
||||||
|
*/
|
||||||
ATTACHMENTS = "attachment",
|
ATTACHMENTS = "attachment",
|
||||||
|
/**
|
||||||
|
* a JSON type, similar to the attachments type, called Attachment within Budibase. This type functions
|
||||||
|
* much the same as the attachment list, but only holds a single attachment metadata as an object.
|
||||||
|
* This simplifies the binding experience of using this column type.
|
||||||
|
*/
|
||||||
ATTACHMENT_SINGLE = "attachment_single",
|
ATTACHMENT_SINGLE = "attachment_single",
|
||||||
|
/**
|
||||||
|
* a complex type, called Relationships within Budibase. This is the most complex type of Budibase,
|
||||||
|
* nothing should be stored against rows under link columns; this type simply represents the
|
||||||
|
* relationship between tables as part of the table schema. When rows are input to the Budibase API
|
||||||
|
* relationships to be made are represented as a list of row IDs to link. When rows are returned
|
||||||
|
* from the Budibase API it will contain a list of row IDs and display column values of the related rows.
|
||||||
|
*/
|
||||||
LINK = "link",
|
LINK = "link",
|
||||||
|
/**
|
||||||
|
* a complex type, called Formulas within Budibase. This type has two variants, static and dynamic, with
|
||||||
|
* static only being supported against internal tables. Dynamic formulas calculate a provided HBS/JS binding
|
||||||
|
* based on the row context and enrich it when rows are being returned from the API. Static bindings calculate
|
||||||
|
* this when rows are being stored, so that the formula output can be searched upon within the DB.
|
||||||
|
*/
|
||||||
FORMULA = "formula",
|
FORMULA = "formula",
|
||||||
|
/**
|
||||||
|
* a complex type, called Auto Column within Budibase. This type has a few variants, with options such as a
|
||||||
|
* date for created at/updated at, an auto ID column with auto-increments as rows are saved and a user
|
||||||
|
* relationship type which stores the created by/updated by user details. These subtypes all depend on the
|
||||||
|
* date, number of link types respectively. There is one case where these will be executed in the browser,
|
||||||
|
* that is part of the initial formula definition, the formula will be live evaluated in the browser.
|
||||||
|
*/
|
||||||
AUTO = "auto",
|
AUTO = "auto",
|
||||||
|
/**
|
||||||
|
* a JSON type, called JSON within Budibase. This type allows any arbitrary JSON to be input to this column
|
||||||
|
* type, which will be represented as a JSON object in the row. This type depends on a schema being
|
||||||
|
* provided to make the JSON searchable/bindable, the JSON cannot be fully dynamic.
|
||||||
|
*/
|
||||||
JSON = "json",
|
JSON = "json",
|
||||||
|
/**
|
||||||
|
* @deprecated an internal type, this is an old deprecated type which is no longer used - still represented to note it
|
||||||
|
* could appear in very old tables.
|
||||||
|
*/
|
||||||
INTERNAL = "internal",
|
INTERNAL = "internal",
|
||||||
|
/**
|
||||||
|
* a string type, called Barcode/QR within Budibase. This type is used to denote to forms to that this column
|
||||||
|
* should be filled in using a camera to read a barcode, there is a form component which will be used when this
|
||||||
|
* type is found. The column will contain the contents of any barcode scanned.
|
||||||
|
*/
|
||||||
BARCODEQR = "barcodeqr",
|
BARCODEQR = "barcodeqr",
|
||||||
|
/**
|
||||||
|
* a string type, this allows representing very large integers, but they are held/managed within Budibase as
|
||||||
|
* strings. When stored in external databases Budibase will attempt to use a real big integer type and depend
|
||||||
|
* on the database parsing the string to this type as part of saving.
|
||||||
|
*/
|
||||||
BIGINT = "bigint",
|
BIGINT = "bigint",
|
||||||
|
/**
|
||||||
|
* a JSON type, called User within Budibase. This type is used to represent a link to an internal Budibase
|
||||||
|
* resource, like a user or group, today only users are supported. This type will be represented as an
|
||||||
|
* array of internal resource IDs (e.g. user IDs) within the row - this ID list will be enriched with
|
||||||
|
* the full resources when rows are returned from the API. The full resources can be input to the API, or
|
||||||
|
* an array of resource IDs, the API will squash these down and validate them before saving the row.
|
||||||
|
*/
|
||||||
BB_REFERENCE = "bb_reference",
|
BB_REFERENCE = "bb_reference",
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1,31 @@
|
||||||
|
#!/bin/bash
|
||||||
|
yarn global add zbctl
|
||||||
|
export ZEEBE_ADDRESS='localhost:26500'
|
||||||
|
|
||||||
|
cd ../budibase-bpm
|
||||||
|
|
||||||
|
is_camunda_ready() {
|
||||||
|
if (zbctl --insecure status 2>/dev/null) | grep -q 'Healthy'; then
|
||||||
|
return 1
|
||||||
|
else
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
docker-compose up -d
|
||||||
|
echo "waiting for Camunda to be ready..."
|
||||||
|
|
||||||
|
while is_camunda_ready -eq 0; do sleep 1; done
|
||||||
|
|
||||||
|
cd src/main/resources/models
|
||||||
|
|
||||||
|
echo "deploy processes..."
|
||||||
|
zbctl deploy resource offboarding.bpmn --insecure
|
||||||
|
zbctl deploy resource onboarding.bpmn --insecure
|
||||||
|
|
||||||
|
cd ../../../../../budibase/packages/account-portal/packages/server
|
||||||
|
|
||||||
|
yarn worker:run & cd ../../../.. && yarn dev:accountportal
|
||||||
|
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue