Compare commits
20 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| 35a8712eef | |||
| b200b3579a | |||
| a13ca2d954 | |||
| 36ed678c5a | |||
| c2c8b91dc5 | |||
| 9d208292a5 | |||
| 8bb3b61f2d | |||
| 5023f229ca | |||
| 55b547db75 | |||
| 377459f430 | |||
| df99bf4743 | |||
| fc4fde490f | |||
| a6d2a4d34f | |||
| 232158f85d | |||
| 3c998edb3c | |||
| 92722b0fef | |||
| 265c586198 | |||
| 440ecc8193 | |||
| 1bae960139 | |||
| 1dec96551d |
18
.idea/codeStyles/Project.xml
generated
18
.idea/codeStyles/Project.xml
generated
@@ -2,16 +2,30 @@
|
|||||||
<code_scheme name="Project" version="173">
|
<code_scheme name="Project" version="173">
|
||||||
<JetCodeStyleSettings>
|
<JetCodeStyleSettings>
|
||||||
<option name="PACKAGES_TO_USE_STAR_IMPORTS">
|
<option name="PACKAGES_TO_USE_STAR_IMPORTS">
|
||||||
|
<value />
|
||||||
|
</option>
|
||||||
|
<option name="PACKAGES_IMPORT_LAYOUT">
|
||||||
<value>
|
<value>
|
||||||
<package name="java.util" withSubpackages="true" static="false" />
|
<package name="" alias="false" withSubpackages="true" />
|
||||||
|
<package name="" alias="true" withSubpackages="true" />
|
||||||
</value>
|
</value>
|
||||||
</option>
|
</option>
|
||||||
<option name="SPACE_BEFORE_EXTEND_COLON" value="false" />
|
<option name="SPACE_BEFORE_EXTEND_COLON" value="false" />
|
||||||
<option name="NAME_COUNT_TO_USE_STAR_IMPORT" value="2147483647" />
|
<option name="NAME_COUNT_TO_USE_STAR_IMPORT" value="2147483647" />
|
||||||
|
<option name="NAME_COUNT_TO_USE_STAR_IMPORT_FOR_MEMBERS" value="2147483647" />
|
||||||
|
<option name="IMPORT_NESTED_CLASSES" value="true" />
|
||||||
<option name="CODE_STYLE_DEFAULTS" value="KOTLIN_OFFICIAL" />
|
<option name="CODE_STYLE_DEFAULTS" value="KOTLIN_OFFICIAL" />
|
||||||
</JetCodeStyleSettings>
|
</JetCodeStyleSettings>
|
||||||
<codeStyleSettings language="kotlin">
|
<codeStyleSettings language="kotlin">
|
||||||
<option name="CODE_STYLE_DEFAULTS" value="KOTLIN_OFFICIAL" />
|
<option name="CALL_PARAMETERS_WRAP" value="5" />
|
||||||
|
<option name="CALL_PARAMETERS_LPAREN_ON_NEXT_LINE" value="true" />
|
||||||
|
<option name="CALL_PARAMETERS_RPAREN_ON_NEXT_LINE" value="true" />
|
||||||
|
<option name="METHOD_PARAMETERS_WRAP" value="5" />
|
||||||
|
<option name="METHOD_PARAMETERS_LPAREN_ON_NEXT_LINE" value="true" />
|
||||||
|
<option name="METHOD_PARAMETERS_RPAREN_ON_NEXT_LINE" value="true" />
|
||||||
|
<option name="EXTENDS_LIST_WRAP" value="1" />
|
||||||
|
<option name="METHOD_CALL_CHAIN_WRAP" value="1" />
|
||||||
|
<option name="ASSIGNMENT_WRAP" value="1" />
|
||||||
<indentOptions>
|
<indentOptions>
|
||||||
<option name="CONTINUATION_INDENT_SIZE" value="4" />
|
<option name="CONTINUATION_INDENT_SIZE" value="4" />
|
||||||
</indentOptions>
|
</indentOptions>
|
||||||
|
|||||||
2
.idea/misc.xml
generated
2
.idea/misc.xml
generated
@@ -3,7 +3,7 @@
|
|||||||
<component name="FrameworkDetectionExcludesConfiguration">
|
<component name="FrameworkDetectionExcludesConfiguration">
|
||||||
<file type="web" url="file://$PROJECT_DIR$" />
|
<file type="web" url="file://$PROJECT_DIR$" />
|
||||||
</component>
|
</component>
|
||||||
<component name="ProjectRootManager" version="2" languageLevel="JDK_11" default="true" project-jdk-name="adopt-openjdk-11" project-jdk-type="JavaSDK">
|
<component name="ProjectRootManager" version="2" languageLevel="JDK_11" default="true" project-jdk-name="corretto-11" project-jdk-type="JavaSDK">
|
||||||
<output url="file://$PROJECT_DIR$/out" />
|
<output url="file://$PROJECT_DIR$/out" />
|
||||||
</component>
|
</component>
|
||||||
<component name="TaskProjectConfiguration">
|
<component name="TaskProjectConfiguration">
|
||||||
|
|||||||
3
.idea/runConfigurations/Check.xml
generated
3
.idea/runConfigurations/Check.xml
generated
@@ -1,7 +1,8 @@
|
|||||||
<component name="ProjectRunConfigurationManager">
|
<component name="ProjectRunConfigurationManager">
|
||||||
<configuration default="false" name="Check" type="CompoundRunConfigurationType">
|
<configuration default="false" name="Check" type="CompoundRunConfigurationType">
|
||||||
<toRun name="tests" type="JUnit" />
|
<toRun name="run DB" type="docker-deploy" />
|
||||||
<toRun name="Lint" type="GradleRunConfiguration" />
|
<toRun name="Lint" type="GradleRunConfiguration" />
|
||||||
|
<toRun name="tests" type="JUnit" />
|
||||||
<method v="2" />
|
<method v="2" />
|
||||||
</configuration>
|
</configuration>
|
||||||
</component>
|
</component>
|
||||||
@@ -1,34 +0,0 @@
|
|||||||
<component name="ProjectRunConfigurationManager">
|
|
||||||
<configuration default="false" name="Test and Publish To MavenLocal" type="GradleRunConfiguration" factoryName="Gradle" singleton="true">
|
|
||||||
<ExternalSystemSettings>
|
|
||||||
<option name="executionName" />
|
|
||||||
<option name="externalProjectPath" value="$PROJECT_DIR$" />
|
|
||||||
<option name="externalSystemIdString" value="GRADLE" />
|
|
||||||
<option name="scriptParameters" value="" />
|
|
||||||
<option name="taskDescriptions">
|
|
||||||
<list />
|
|
||||||
</option>
|
|
||||||
<option name="taskNames">
|
|
||||||
<list>
|
|
||||||
<option value="publishToMavenLocal" />
|
|
||||||
</list>
|
|
||||||
</option>
|
|
||||||
<option name="vmOptions" value="" />
|
|
||||||
</ExternalSystemSettings>
|
|
||||||
<extension name="net.ashald.envfile">
|
|
||||||
<option name="IS_ENABLED" value="false" />
|
|
||||||
<option name="IS_SUBST" value="false" />
|
|
||||||
<option name="IS_PATH_MACRO_SUPPORTED" value="false" />
|
|
||||||
<option name="IS_IGNORE_MISSING_FILES" value="false" />
|
|
||||||
<option name="IS_ENABLE_EXPERIMENTAL_INTEGRATIONS" value="false" />
|
|
||||||
<ENTRIES>
|
|
||||||
<ENTRY IS_ENABLED="true" PARSER="runconfig" />
|
|
||||||
</ENTRIES>
|
|
||||||
</extension>
|
|
||||||
<GradleScriptDebugEnabled>true</GradleScriptDebugEnabled>
|
|
||||||
<method v="2">
|
|
||||||
<option name="RunConfigurationTask" enabled="true" run_configuration_name="Lint" run_configuration_type="GradleRunConfiguration" />
|
|
||||||
<option name="RunConfigurationTask" enabled="true" run_configuration_name="tests" run_configuration_type="JUnit" />
|
|
||||||
</method>
|
|
||||||
</configuration>
|
|
||||||
</component>
|
|
||||||
4
.idea/runConfigurations/tests.xml
generated
4
.idea/runConfigurations/tests.xml
generated
@@ -1,6 +1,6 @@
|
|||||||
<component name="ProjectRunConfigurationManager">
|
<component name="ProjectRunConfigurationManager">
|
||||||
<configuration default="false" name="tests" type="JUnit" factoryName="JUnit" singleton="false">
|
<configuration default="false" name="tests" type="JUnit" factoryName="JUnit" singleton="false">
|
||||||
<module name="postgresjson.test" />
|
<module name="postgres-json.test" />
|
||||||
<useClassPathOnly />
|
<useClassPathOnly />
|
||||||
<extension name="net.ashald.envfile">
|
<extension name="net.ashald.envfile">
|
||||||
<option name="IS_ENABLED" value="false" />
|
<option name="IS_ENABLED" value="false" />
|
||||||
@@ -13,7 +13,7 @@
|
|||||||
</ENTRIES>
|
</ENTRIES>
|
||||||
</extension>
|
</extension>
|
||||||
<option name="ALTERNATIVE_JRE_PATH_ENABLED" value="true" />
|
<option name="ALTERNATIVE_JRE_PATH_ENABLED" value="true" />
|
||||||
<option name="ALTERNATIVE_JRE_PATH" value="$USER_HOME$/.jdks/adopt-openjdk-11.0.5" />
|
<option name="ALTERNATIVE_JRE_PATH" value="corretto-11" />
|
||||||
<option name="MAIN_CLASS_NAME" value="" />
|
<option name="MAIN_CLASS_NAME" value="" />
|
||||||
<option name="METHOD_NAME" value="" />
|
<option name="METHOD_NAME" value="" />
|
||||||
<option name="TEST_OBJECT" value="directory" />
|
<option name="TEST_OBJECT" value="directory" />
|
||||||
|
|||||||
@@ -4,14 +4,15 @@ plugins {
|
|||||||
jacoco
|
jacoco
|
||||||
|
|
||||||
id("maven-publish")
|
id("maven-publish")
|
||||||
id("org.jetbrains.kotlin.jvm") version "1.3.50"
|
kotlin("jvm") version "1.4.30"
|
||||||
|
|
||||||
id("org.jlleitschuh.gradle.ktlint") version "8.2.0"
|
id("org.jlleitschuh.gradle.ktlint") version "10.0.0"
|
||||||
id("org.owasp.dependencycheck") version "5.1.0"
|
id("org.owasp.dependencycheck") version "6.1.1"
|
||||||
id("fr.coppernic.versioning") version "3.1.2"
|
id("fr.coppernic.versioning") version "3.2.1"
|
||||||
|
id("com.avast.gradle.docker-compose") version "0.14.0"
|
||||||
}
|
}
|
||||||
|
|
||||||
group = "flecomte"
|
group = "com.github.flecomte"
|
||||||
version = versioning.info.tag
|
version = versioning.info.tag
|
||||||
|
|
||||||
repositories {
|
repositories {
|
||||||
@@ -22,22 +23,46 @@ repositories {
|
|||||||
tasks.withType<KotlinCompile> {
|
tasks.withType<KotlinCompile> {
|
||||||
kotlinOptions {
|
kotlinOptions {
|
||||||
jvmTarget = "11"
|
jvmTarget = "11"
|
||||||
|
sourceCompatibility = "11"
|
||||||
|
targetCompatibility = "11"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
val compileKotlin: KotlinCompile by tasks
|
||||||
|
compileKotlin.kotlinOptions {
|
||||||
|
jvmTarget = "1.8"
|
||||||
|
}
|
||||||
|
val compileTestKotlin: KotlinCompile by tasks
|
||||||
|
compileTestKotlin.kotlinOptions {
|
||||||
|
jvmTarget = "1.8"
|
||||||
|
}
|
||||||
|
|
||||||
|
tasks.test {
|
||||||
|
useJUnit()
|
||||||
|
useJUnitPlatform()
|
||||||
|
systemProperty("junit.jupiter.execution.parallel.enabled", true)
|
||||||
|
finalizedBy(tasks.ktlintCheck)
|
||||||
|
}
|
||||||
|
|
||||||
|
tasks.publishToMavenLocal {
|
||||||
|
dependsOn(tasks.test)
|
||||||
|
}
|
||||||
|
|
||||||
dependencies {
|
dependencies {
|
||||||
implementation("org.jetbrains.kotlin:kotlin-stdlib-jdk8")
|
implementation("org.jetbrains.kotlin:kotlin-stdlib-jdk8")
|
||||||
implementation("org.jetbrains.kotlin:kotlin-reflect:1.3.31")
|
implementation("org.jetbrains.kotlin:kotlin-reflect:1.4.31")
|
||||||
implementation("com.fasterxml.jackson.module:jackson-module-kotlin:2.9.9")
|
implementation("com.fasterxml.jackson.module:jackson-module-kotlin:2.12.1")
|
||||||
implementation("com.fasterxml.jackson.datatype:jackson-datatype-joda:2.9.9")
|
implementation("com.fasterxml.jackson.datatype:jackson-datatype-joda:2.12.1")
|
||||||
implementation("com.github.jasync-sql:jasync-postgresql:1.0.7")
|
implementation("com.github.jasync-sql:jasync-postgresql:1.1.7")
|
||||||
implementation("org.slf4j:slf4j-api:1.7.26")
|
implementation("org.slf4j:slf4j-api:1.7.30")
|
||||||
|
implementation("com.avast.gradle:gradle-docker-compose-plugin:0.14.0")
|
||||||
|
|
||||||
testImplementation("ch.qos.logback:logback-classic:1.2.3")
|
testImplementation("ch.qos.logback:logback-classic:1.2.3")
|
||||||
testImplementation("ch.qos.logback:logback-core:1.2.3")
|
testImplementation("ch.qos.logback:logback-core:1.2.3")
|
||||||
testImplementation("io.mockk:mockk:1.9")
|
testImplementation("io.mockk:mockk:1.10.6")
|
||||||
testImplementation("org.junit.jupiter:junit-jupiter:5.4.2")
|
testImplementation("org.junit.jupiter:junit-jupiter:5.7.1")
|
||||||
testImplementation("org.amshove.kluent:kluent:1.47")
|
testImplementation("org.jetbrains.kotlin:kotlin-test-junit:1.4.30")
|
||||||
|
testImplementation("org.amshove.kluent:kluent:1.65")
|
||||||
}
|
}
|
||||||
|
|
||||||
val sourcesJar by tasks.creating(Jar::class) {
|
val sourcesJar by tasks.creating(Jar::class) {
|
||||||
@@ -45,6 +70,14 @@ val sourcesJar by tasks.creating(Jar::class) {
|
|||||||
from(sourceSets.getByName("main").allSource)
|
from(sourceSets.getByName("main").allSource)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
apply(plugin = "docker-compose")
|
||||||
|
dockerCompose {
|
||||||
|
projectName = "postgres-json"
|
||||||
|
useComposeFiles = listOf("docker-compose.yml")
|
||||||
|
stopContainers = true
|
||||||
|
isRequiredBy(project.tasks.test)
|
||||||
|
}
|
||||||
|
|
||||||
publishing {
|
publishing {
|
||||||
repositories {
|
repositories {
|
||||||
maven {
|
maven {
|
||||||
|
|||||||
15
docker-compose.yml
Normal file
15
docker-compose.yml
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
# To execute this docker-compose yml file use docker-compose -f <file_name> up
|
||||||
|
# Add the "-d" flag at the end for detached execution
|
||||||
|
version: '3.7'
|
||||||
|
services:
|
||||||
|
db:
|
||||||
|
container_name: postgres_json
|
||||||
|
build:
|
||||||
|
context: docker/postgresql
|
||||||
|
restart: always
|
||||||
|
ports:
|
||||||
|
- 5555:5432
|
||||||
|
environment:
|
||||||
|
POSTGRES_DB: json_test
|
||||||
|
POSTGRES_USER: test
|
||||||
|
POSTGRES_PASSWORD: test
|
||||||
9
docker/postgresql/Dockerfile
Normal file
9
docker/postgresql/Dockerfile
Normal file
@@ -0,0 +1,9 @@
|
|||||||
|
FROM postgres:11
|
||||||
|
|
||||||
|
COPY postgresql.conf /tmp/postgresql.conf
|
||||||
|
COPY extension.sh /docker-entrypoint-initdb.d/000-extension.sh
|
||||||
|
COPY setup.sh /docker-entrypoint-initdb.d/100-setup.sh
|
||||||
|
|
||||||
|
ENTRYPOINT ["docker-entrypoint.sh"]
|
||||||
|
EXPOSE 5432
|
||||||
|
CMD ["postgres"]
|
||||||
7
docker/postgresql/extension.sh
Normal file
7
docker/postgresql/extension.sh
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
#!/usr/bin/env bash
|
||||||
|
set -e
|
||||||
|
|
||||||
|
psql -v ON_ERROR_STOP=1 --username "test" --dbname "json_test" <<-EOSQL
|
||||||
|
create extension if not exists plpgsql;
|
||||||
|
create extension if not exists "uuid-ossp";
|
||||||
|
EOSQL
|
||||||
668
docker/postgresql/postgresql.conf
Normal file
668
docker/postgresql/postgresql.conf
Normal file
@@ -0,0 +1,668 @@
|
|||||||
|
# -----------------------------
|
||||||
|
# PostgreSQL configuration file
|
||||||
|
# -----------------------------
|
||||||
|
#
|
||||||
|
# This file consists of lines of the form:
|
||||||
|
#
|
||||||
|
# name = value
|
||||||
|
#
|
||||||
|
# (The "=" is optional.) Whitespace may be used. Comments are introduced with
|
||||||
|
# "#" anywhere on a line. The complete list of parameter names and allowed
|
||||||
|
# values can be found in the PostgreSQL documentation.
|
||||||
|
#
|
||||||
|
# The commented-out settings shown in this file represent the default values.
|
||||||
|
# Re-commenting a setting is NOT sufficient to revert it to the default value;
|
||||||
|
# you need to reload the server.
|
||||||
|
#
|
||||||
|
# This file is read on server startup and when the server receives a SIGHUP
|
||||||
|
# signal. If you edit the file on a running system, you have to SIGHUP the
|
||||||
|
# server for the changes to take effect, run "pg_ctl reload", or execute
|
||||||
|
# "SELECT pg_reload_conf()". Some parameters, which are marked below,
|
||||||
|
# require a server shutdown and restart to take effect.
|
||||||
|
#
|
||||||
|
# Any parameter can also be given as a command-line option to the server, e.g.,
|
||||||
|
# "postgres -c log_connections=on". Some parameters can be changed at run time
|
||||||
|
# with the "SET" SQL command.
|
||||||
|
#
|
||||||
|
# Memory units: kB = kilobytes Time units: ms = milliseconds
|
||||||
|
# MB = megabytes s = seconds
|
||||||
|
# GB = gigabytes min = minutes
|
||||||
|
# TB = terabytes h = hours
|
||||||
|
# d = days
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# FILE LOCATIONS
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# The default values of these variables are driven from the -D command-line
|
||||||
|
# option or PGDATA environment variable, represented here as ConfigDir.
|
||||||
|
|
||||||
|
#data_directory = 'ConfigDir' # use data in another directory
|
||||||
|
# (change requires restart)
|
||||||
|
#hba_file = 'ConfigDir/pg_hba.conf' # host-based authentication file
|
||||||
|
# (change requires restart)
|
||||||
|
#ident_file = 'ConfigDir/pg_ident.conf' # ident configuration file
|
||||||
|
# (change requires restart)
|
||||||
|
|
||||||
|
# If external_pid_file is not explicitly set, no extra PID file is written.
|
||||||
|
#external_pid_file = '' # write an extra PID file
|
||||||
|
# (change requires restart)
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# CONNECTIONS AND AUTHENTICATION
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# - Connection Settings -
|
||||||
|
|
||||||
|
listen_addresses = '*'
|
||||||
|
# comma-separated list of addresses;
|
||||||
|
# defaults to 'localhost'; use '*' for all
|
||||||
|
# (change requires restart)
|
||||||
|
#port = 5432 # (change requires restart)
|
||||||
|
#max_connections = 100 # (change requires restart)
|
||||||
|
#superuser_reserved_connections = 3 # (change requires restart)
|
||||||
|
#unix_socket_directories = '/tmp' # comma-separated list of directories
|
||||||
|
# (change requires restart)
|
||||||
|
#unix_socket_group = '' # (change requires restart)
|
||||||
|
#unix_socket_permissions = 0777 # begin with 0 to use octal notation
|
||||||
|
# (change requires restart)
|
||||||
|
#bonjour = off # advertise server via Bonjour
|
||||||
|
# (change requires restart)
|
||||||
|
#bonjour_name = '' # defaults to the computer name
|
||||||
|
# (change requires restart)
|
||||||
|
|
||||||
|
# - Security and Authentication -
|
||||||
|
|
||||||
|
#authentication_timeout = 1min # 1s-600s
|
||||||
|
#ssl = off
|
||||||
|
#ssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL' # allowed SSL ciphers
|
||||||
|
#ssl_prefer_server_ciphers = on
|
||||||
|
#ssl_ecdh_curve = 'prime256v1'
|
||||||
|
#ssl_dh_params_file = ''
|
||||||
|
#ssl_cert_file = 'server.crt'
|
||||||
|
#ssl_key_file = 'server.key'
|
||||||
|
#ssl_ca_file = ''
|
||||||
|
#ssl_crl_file = ''
|
||||||
|
#password_encryption = md5 # md5 or scram-sha-256
|
||||||
|
#db_user_namespace = off
|
||||||
|
#row_security = on
|
||||||
|
|
||||||
|
# GSSAPI using Kerberos
|
||||||
|
#krb_server_keyfile = ''
|
||||||
|
#krb_caseins_users = off
|
||||||
|
|
||||||
|
# - TCP Keepalives -
|
||||||
|
# see "man 7 tcp" for details
|
||||||
|
|
||||||
|
#tcp_keepalives_idle = 0 # TCP_KEEPIDLE, in seconds;
|
||||||
|
# 0 selects the system default
|
||||||
|
#tcp_keepalives_interval = 0 # TCP_KEEPINTVL, in seconds;
|
||||||
|
# 0 selects the system default
|
||||||
|
#tcp_keepalives_count = 0 # TCP_KEEPCNT;
|
||||||
|
# 0 selects the system default
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# RESOURCE USAGE (except WAL)
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# - Memory -
|
||||||
|
|
||||||
|
shared_buffers = 1GB # min 128kB
|
||||||
|
# (change requires restart)
|
||||||
|
#huge_pages = try # on, off, or try
|
||||||
|
# (change requires restart)
|
||||||
|
#temp_buffers = 8MB # min 800kB
|
||||||
|
#max_prepared_transactions = 0 # zero disables the feature
|
||||||
|
# (change requires restart)
|
||||||
|
# Caution: it is not advisable to set max_prepared_transactions nonzero unless
|
||||||
|
# you actively intend to use prepared transactions.
|
||||||
|
work_mem = 256MB # min 64kB
|
||||||
|
#maintenance_work_mem = 64MB # min 1MB
|
||||||
|
#replacement_sort_tuples = 150000 # limits use of replacement selection sort
|
||||||
|
#autovacuum_work_mem = -1 # min 1MB, or -1 to use maintenance_work_mem
|
||||||
|
#max_stack_depth = 2MB # min 100kB
|
||||||
|
#dynamic_shared_memory_type = posix # the default is the first option
|
||||||
|
# supported by the operating system:
|
||||||
|
# posix
|
||||||
|
# sysv
|
||||||
|
# windows
|
||||||
|
# mmap
|
||||||
|
# use none to disable dynamic shared memory
|
||||||
|
# (change requires restart)
|
||||||
|
|
||||||
|
# - Disk -
|
||||||
|
|
||||||
|
#temp_file_limit = -1 # limits per-process temp file space
|
||||||
|
# in kB, or -1 for no limit
|
||||||
|
|
||||||
|
# - Kernel Resource Usage -
|
||||||
|
|
||||||
|
#max_files_per_process = 1000 # min 25
|
||||||
|
# (change requires restart)
|
||||||
|
shared_preload_libraries = 'pg_stat_statements'
|
||||||
|
pg_stat_statements.max = 10000
|
||||||
|
pg_stat_statements.track = all
|
||||||
|
|
||||||
|
# - Cost-Based Vacuum Delay -
|
||||||
|
|
||||||
|
#vacuum_cost_delay = 0 # 0-100 milliseconds
|
||||||
|
#vacuum_cost_page_hit = 1 # 0-10000 credits
|
||||||
|
#vacuum_cost_page_miss = 10 # 0-10000 credits
|
||||||
|
#vacuum_cost_page_dirty = 20 # 0-10000 credits
|
||||||
|
#vacuum_cost_limit = 200 # 1-10000 credits
|
||||||
|
|
||||||
|
# - Background Writer -
|
||||||
|
|
||||||
|
#bgwriter_delay = 200ms # 10-10000ms between rounds
|
||||||
|
#bgwriter_lru_maxpages = 100 # 0-1000 max buffers written/round
|
||||||
|
#bgwriter_lru_multiplier = 2.0 # 0-10.0 multiplier on buffers scanned/round
|
||||||
|
#bgwriter_flush_after = 0 # measured in pages, 0 disables
|
||||||
|
|
||||||
|
# - Asynchronous Behavior -
|
||||||
|
|
||||||
|
effective_io_concurrency = 100 # 1-1000; 0 disables prefetching
|
||||||
|
#max_worker_processes = 8 # (change requires restart)
|
||||||
|
max_parallel_workers_per_gather = 4 # taken from max_parallel_workers
|
||||||
|
#max_parallel_workers = 8 # maximum number of max_worker_processes that
|
||||||
|
# can be used in parallel queries
|
||||||
|
#old_snapshot_threshold = -1 # 1min-60d; -1 disables; 0 is immediate
|
||||||
|
# (change requires restart)
|
||||||
|
#backend_flush_after = 0 # measured in pages, 0 disables
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# WRITE AHEAD LOG
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# - Settings -
|
||||||
|
|
||||||
|
#wal_level = replica # minimal, replica, or logical
|
||||||
|
# (change requires restart)
|
||||||
|
#fsync = on # flush data to disk for crash safety
|
||||||
|
# (turning this off can cause
|
||||||
|
# unrecoverable data corruption)
|
||||||
|
#synchronous_commit = on # synchronization level;
|
||||||
|
# off, local, remote_write, remote_apply, or on
|
||||||
|
#wal_sync_method = fsync # the default is the first option
|
||||||
|
# supported by the operating system:
|
||||||
|
# open_datasync
|
||||||
|
# fdatasync (default on Linux)
|
||||||
|
# fsync
|
||||||
|
# fsync_writethrough
|
||||||
|
# open_sync
|
||||||
|
#full_page_writes = on # recover from partial page writes
|
||||||
|
#wal_compression = off # enable compression of full-page writes
|
||||||
|
#wal_log_hints = off # also do full page writes of non-critical updates
|
||||||
|
# (change requires restart)
|
||||||
|
#wal_buffers = -1 # min 32kB, -1 sets based on shared_buffers
|
||||||
|
# (change requires restart)
|
||||||
|
#wal_writer_delay = 200ms # 1-10000 milliseconds
|
||||||
|
#wal_writer_flush_after = 1MB # measured in pages, 0 disables
|
||||||
|
|
||||||
|
#commit_delay = 0 # range 0-100000, in microseconds
|
||||||
|
#commit_siblings = 5 # range 1-1000
|
||||||
|
|
||||||
|
# - Checkpoints -
|
||||||
|
|
||||||
|
#checkpoint_timeout = 5min # range 30s-1d
|
||||||
|
#max_wal_size = 1GB
|
||||||
|
#min_wal_size = 80MB
|
||||||
|
#checkpoint_completion_target = 0.5 # checkpoint target duration, 0.0 - 1.0
|
||||||
|
#checkpoint_flush_after = 0 # measured in pages, 0 disables
|
||||||
|
#checkpoint_warning = 30s # 0 disables
|
||||||
|
|
||||||
|
# - Archiving -
|
||||||
|
|
||||||
|
#archive_mode = off # enables archiving; off, on, or always
|
||||||
|
# (change requires restart)
|
||||||
|
#archive_command = '' # command to use to archive a logfile segment
|
||||||
|
# placeholders: %p = path of file to archive
|
||||||
|
# %f = file name only
|
||||||
|
# e.g. 'test ! -f /mnt/server/archivedir/%f && cp %p /mnt/server/archivedir/%f'
|
||||||
|
#archive_timeout = 0 # force a logfile segment switch after this
|
||||||
|
# number of seconds; 0 disables
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# REPLICATION
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# - Sending Server(s) -
|
||||||
|
|
||||||
|
# Set these on the master and on any standby that will send replication data.
|
||||||
|
|
||||||
|
#max_wal_senders = 10 # max number of walsender processes
|
||||||
|
# (change requires restart)
|
||||||
|
#wal_keep_segments = 0 # in logfile segments, 16MB each; 0 disables
|
||||||
|
#wal_sender_timeout = 60s # in milliseconds; 0 disables
|
||||||
|
|
||||||
|
#max_replication_slots = 10 # max number of replication slots
|
||||||
|
# (change requires restart)
|
||||||
|
#track_commit_timestamp = off # collect timestamp of transaction commit
|
||||||
|
# (change requires restart)
|
||||||
|
|
||||||
|
# - Master Server -
|
||||||
|
|
||||||
|
# These settings are ignored on a standby server.
|
||||||
|
|
||||||
|
#synchronous_standby_names = '' # standby servers that provide sync rep
|
||||||
|
# method to choose sync standbys, number of sync standbys,
|
||||||
|
# and comma-separated list of application_name
|
||||||
|
# from standby(s); '*' = all
|
||||||
|
#vacuum_defer_cleanup_age = 0 # number of xacts by which cleanup is delayed
|
||||||
|
|
||||||
|
# - Standby Servers -
|
||||||
|
|
||||||
|
# These settings are ignored on a master server.
|
||||||
|
|
||||||
|
#hot_standby = on # "off" disallows queries during recovery
|
||||||
|
# (change requires restart)
|
||||||
|
#max_standby_archive_delay = 30s # max delay before canceling queries
|
||||||
|
# when reading WAL from archive;
|
||||||
|
# -1 allows indefinite delay
|
||||||
|
#max_standby_streaming_delay = 30s # max delay before canceling queries
|
||||||
|
# when reading streaming WAL;
|
||||||
|
# -1 allows indefinite delay
|
||||||
|
#wal_receiver_status_interval = 10s # send replies at least this often
|
||||||
|
# 0 disables
|
||||||
|
#hot_standby_feedback = off # send info from standby to prevent
|
||||||
|
# query conflicts
|
||||||
|
#wal_receiver_timeout = 60s # time that receiver waits for
|
||||||
|
# communication from master
|
||||||
|
# in milliseconds; 0 disables
|
||||||
|
#wal_retrieve_retry_interval = 5s # time to wait before retrying to
|
||||||
|
# retrieve WAL after a failed attempt
|
||||||
|
|
||||||
|
# - Subscribers -
|
||||||
|
|
||||||
|
# These settings are ignored on a publisher.
|
||||||
|
|
||||||
|
#max_logical_replication_workers = 4 # taken from max_worker_processes
|
||||||
|
# (change requires restart)
|
||||||
|
#max_sync_workers_per_subscription = 2 # taken from max_logical_replication_workers
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# QUERY TUNING
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# - Planner Method Configuration -
|
||||||
|
|
||||||
|
#enable_bitmapscan = on
|
||||||
|
#enable_hashagg = on
|
||||||
|
#enable_hashjoin = on
|
||||||
|
#enable_indexscan = on
|
||||||
|
#enable_indexonlyscan = on
|
||||||
|
#enable_material = on
|
||||||
|
#enable_mergejoin = on
|
||||||
|
#enable_nestloop = on
|
||||||
|
#enable_seqscan = on
|
||||||
|
#enable_sort = on
|
||||||
|
#enable_tidscan = on
|
||||||
|
|
||||||
|
# - Planner Cost Constants -
|
||||||
|
|
||||||
|
#seq_page_cost = 1.0 # measured on an arbitrary scale
|
||||||
|
#random_page_cost = 4.0 # same scale as above
|
||||||
|
#cpu_tuple_cost = 0.01 # same scale as above
|
||||||
|
#cpu_index_tuple_cost = 0.005 # same scale as above
|
||||||
|
#cpu_operator_cost = 0.0025 # same scale as above
|
||||||
|
#parallel_tuple_cost = 0.1 # same scale as above
|
||||||
|
#parallel_setup_cost = 1000.0 # same scale as above
|
||||||
|
#min_parallel_table_scan_size = 8MB
|
||||||
|
#min_parallel_index_scan_size = 512kB
|
||||||
|
#effective_cache_size = 4GB
|
||||||
|
|
||||||
|
# - Genetic Query Optimizer -
|
||||||
|
|
||||||
|
#geqo = on
|
||||||
|
#geqo_threshold = 12
|
||||||
|
#geqo_effort = 5 # range 1-10
|
||||||
|
#geqo_pool_size = 0 # selects default based on effort
|
||||||
|
#geqo_generations = 0 # selects default based on effort
|
||||||
|
#geqo_selection_bias = 2.0 # range 1.5-2.0
|
||||||
|
#geqo_seed = 0.0 # range 0.0-1.0
|
||||||
|
|
||||||
|
# - Other Planner Options -
|
||||||
|
|
||||||
|
#default_statistics_target = 100 # range 1-10000
|
||||||
|
#constraint_exclusion = partition # on, off, or partition
|
||||||
|
#cursor_tuple_fraction = 0.1 # range 0.0-1.0
|
||||||
|
#from_collapse_limit = 8
|
||||||
|
#join_collapse_limit = 8 # 1 disables collapsing of explicit
|
||||||
|
# JOIN clauses
|
||||||
|
#force_parallel_mode = off
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# ERROR REPORTING AND LOGGING
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# - Where to Log -
|
||||||
|
|
||||||
|
#log_destination = 'stderr' # Valid values are combinations of
|
||||||
|
# stderr, csvlog, syslog, and eventlog,
|
||||||
|
# depending on platform. csvlog
|
||||||
|
# requires logging_collector to be on.
|
||||||
|
|
||||||
|
# This is used when logging to stderr:
|
||||||
|
logging_collector = on
|
||||||
|
# Enable capturing of stderr and csvlog
|
||||||
|
# into log files. Required to be on for
|
||||||
|
# csvlogs.
|
||||||
|
# (change requires restart)
|
||||||
|
|
||||||
|
# These are only used if logging_collector is on:
|
||||||
|
log_directory = '/var/log/postgresql'
|
||||||
|
# directory where log files are written,
|
||||||
|
# can be absolute or relative to PGDATA
|
||||||
|
log_filename = 'postgresql-%Y-%m-%d_%H%M%S.log'
|
||||||
|
# log file name pattern,
|
||||||
|
# can include strftime() escapes
|
||||||
|
#log_file_mode = 0600 # creation mode for log files,
|
||||||
|
# begin with 0 to use octal notation
|
||||||
|
log_file_mode = 0660
|
||||||
|
#log_truncate_on_rotation = off # If on, an existing log file with the
|
||||||
|
# same name as the new log file will be
|
||||||
|
# truncated rather than appended to.
|
||||||
|
# But such truncation only occurs on
|
||||||
|
# time-driven rotation, not on restarts
|
||||||
|
# or size-driven rotation. Default is
|
||||||
|
# off, meaning append to existing files
|
||||||
|
# in all cases.
|
||||||
|
log_rotation_age = 1d
|
||||||
|
# Automatic rotation of logfiles will
|
||||||
|
# happen after that time. 0 disables.
|
||||||
|
log_rotation_size = 100MB
|
||||||
|
# Automatic rotation of logfiles will
|
||||||
|
# happen after that much log output.
|
||||||
|
# 0 disables.
|
||||||
|
|
||||||
|
# These are relevant when logging to syslog:
|
||||||
|
#syslog_facility = 'LOCAL0'
|
||||||
|
#syslog_ident = 'postgres'
|
||||||
|
#syslog_sequence_numbers = on
|
||||||
|
#syslog_split_messages = on
|
||||||
|
|
||||||
|
# This is only relevant when logging to eventlog (win32):
|
||||||
|
# (change requires restart)
|
||||||
|
#event_source = 'PostgreSQL'
|
||||||
|
|
||||||
|
# - When to Log -
|
||||||
|
|
||||||
|
#client_min_messages = notice # values in order of decreasing detail:
|
||||||
|
# debug5
|
||||||
|
# debug4
|
||||||
|
# debug3
|
||||||
|
# debug2
|
||||||
|
# debug1
|
||||||
|
# log
|
||||||
|
# notice
|
||||||
|
# warning
|
||||||
|
# error
|
||||||
|
|
||||||
|
#log_min_messages = warning # values in order of decreasing detail:
|
||||||
|
# debug5
|
||||||
|
# debug4
|
||||||
|
# debug3
|
||||||
|
# debug2
|
||||||
|
# debug1
|
||||||
|
# info
|
||||||
|
# notice
|
||||||
|
# warning
|
||||||
|
# error
|
||||||
|
# log
|
||||||
|
# fatal
|
||||||
|
# panic
|
||||||
|
|
||||||
|
#log_min_error_statement = error # values in order of decreasing detail:
|
||||||
|
# debug5
|
||||||
|
# debug4
|
||||||
|
# debug3
|
||||||
|
# debug2
|
||||||
|
# debug1
|
||||||
|
# info
|
||||||
|
# notice
|
||||||
|
# warning
|
||||||
|
# error
|
||||||
|
# log
|
||||||
|
# fatal
|
||||||
|
# panic (effectively off)
|
||||||
|
|
||||||
|
#log_min_duration_statement = -1 # -1 is disabled, 0 logs all statements
|
||||||
|
# and their durations, > 0 logs only
|
||||||
|
# statements running at least this number
|
||||||
|
# of milliseconds
|
||||||
|
|
||||||
|
|
||||||
|
# - What to Log -
|
||||||
|
|
||||||
|
#debug_print_parse = off
|
||||||
|
#debug_print_rewritten = off
|
||||||
|
#debug_print_plan = off
|
||||||
|
#debug_pretty_print = on
|
||||||
|
#log_checkpoints = off
|
||||||
|
#log_connections = off
|
||||||
|
#log_disconnections = off
|
||||||
|
#log_duration = off
|
||||||
|
#log_error_verbosity = default # terse, default, or verbose messages
|
||||||
|
#log_hostname = off
|
||||||
|
#log_line_prefix = '%m [%p] ' # special values:
|
||||||
|
# %a = application name
|
||||||
|
# %u = user name
|
||||||
|
# %d = database name
|
||||||
|
# %r = remote host and port
|
||||||
|
# %h = remote host
|
||||||
|
# %p = process ID
|
||||||
|
# %t = timestamp without milliseconds
|
||||||
|
# %m = timestamp with milliseconds
|
||||||
|
# %n = timestamp with milliseconds (as a Unix epoch)
|
||||||
|
# %i = command tag
|
||||||
|
# %e = SQL state
|
||||||
|
# %c = session ID
|
||||||
|
# %l = session line number
|
||||||
|
# %s = session start timestamp
|
||||||
|
# %v = virtual transaction ID
|
||||||
|
# %x = transaction ID (0 if none)
|
||||||
|
# %q = stop here in non-session
|
||||||
|
# processes
|
||||||
|
# %% = '%'
|
||||||
|
# e.g. '<%u%%%d> '
|
||||||
|
#log_lock_waits = off # log lock waits >= deadlock_timeout
|
||||||
|
#log_statement = 'none' # none, ddl, mod, all
|
||||||
|
#log_replication_commands = off
|
||||||
|
#log_temp_files = -1 # log temporary files equal or larger
|
||||||
|
# than the specified size in kilobytes;
|
||||||
|
# -1 disables, 0 logs all temp files
|
||||||
|
#log_timezone = 'GMT'
|
||||||
|
|
||||||
|
|
||||||
|
# - Process Title -
|
||||||
|
|
||||||
|
#cluster_name = '' # added to process titles if nonempty
|
||||||
|
# (change requires restart)
|
||||||
|
#update_process_title = on
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# RUNTIME STATISTICS
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# - Query/Index Statistics Collector -
|
||||||
|
|
||||||
|
#track_activities = on
|
||||||
|
#track_counts = on
|
||||||
|
#track_io_timing = off
|
||||||
|
#track_functions = none # none, pl, all
|
||||||
|
#track_activity_query_size = 1024 # (change requires restart)
|
||||||
|
#stats_temp_directory = 'pg_stat_tmp'
|
||||||
|
|
||||||
|
|
||||||
|
# - Statistics Monitoring -
|
||||||
|
|
||||||
|
#log_parser_stats = off
|
||||||
|
#log_planner_stats = off
|
||||||
|
#log_executor_stats = off
|
||||||
|
#log_statement_stats = off
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# AUTOVACUUM PARAMETERS
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
#autovacuum = on # Enable autovacuum subprocess? 'on'
|
||||||
|
# requires track_counts to also be on.
|
||||||
|
#log_autovacuum_min_duration = -1 # -1 disables, 0 logs all actions and
|
||||||
|
# their durations, > 0 logs only
|
||||||
|
# actions running at least this number
|
||||||
|
# of milliseconds.
|
||||||
|
#autovacuum_max_workers = 3 # max number of autovacuum subprocesses
|
||||||
|
# (change requires restart)
|
||||||
|
#autovacuum_naptime = 1min # time between autovacuum runs
|
||||||
|
#autovacuum_vacuum_threshold = 50 # min number of row updates before
|
||||||
|
# vacuum
|
||||||
|
#autovacuum_analyze_threshold = 50 # min number of row updates before
|
||||||
|
# analyze
|
||||||
|
#autovacuum_vacuum_scale_factor = 0.2 # fraction of table size before vacuum
|
||||||
|
#autovacuum_analyze_scale_factor = 0.1 # fraction of table size before analyze
|
||||||
|
#autovacuum_freeze_max_age = 200000000 # maximum XID age before forced vacuum
|
||||||
|
# (change requires restart)
|
||||||
|
#autovacuum_multixact_freeze_max_age = 400000000 # maximum multixact age
|
||||||
|
# before forced vacuum
|
||||||
|
# (change requires restart)
|
||||||
|
#autovacuum_vacuum_cost_delay = 20ms # default vacuum cost delay for
|
||||||
|
# autovacuum, in milliseconds;
|
||||||
|
# -1 means use vacuum_cost_delay
|
||||||
|
#autovacuum_vacuum_cost_limit = -1 # default vacuum cost limit for
|
||||||
|
# autovacuum, -1 means use
|
||||||
|
# vacuum_cost_limit
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# CLIENT CONNECTION DEFAULTS
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# - Statement Behavior -
|
||||||
|
|
||||||
|
#search_path = '"$user", public' # schema names
|
||||||
|
#default_tablespace = '' # a tablespace name, '' uses the default
|
||||||
|
#temp_tablespaces = '' # a list of tablespace names, '' uses
|
||||||
|
# only default tablespace
|
||||||
|
#check_function_bodies = on
|
||||||
|
#default_transaction_isolation = 'read committed'
|
||||||
|
#default_transaction_read_only = off
|
||||||
|
#default_transaction_deferrable = off
|
||||||
|
#session_replication_role = 'origin'
|
||||||
|
#statement_timeout = 0 # in milliseconds, 0 is disabled
|
||||||
|
#lock_timeout = 0 # in milliseconds, 0 is disabled
|
||||||
|
#idle_in_transaction_session_timeout = 0 # in milliseconds, 0 is disabled
|
||||||
|
#vacuum_freeze_min_age = 50000000
|
||||||
|
#vacuum_freeze_table_age = 150000000
|
||||||
|
#vacuum_multixact_freeze_min_age = 5000000
|
||||||
|
#vacuum_multixact_freeze_table_age = 150000000
|
||||||
|
#bytea_output = 'hex' # hex, escape
|
||||||
|
#xmlbinary = 'base64'
|
||||||
|
#xmloption = 'content'
|
||||||
|
#gin_fuzzy_search_limit = 0
|
||||||
|
#gin_pending_list_limit = 4MB
|
||||||
|
|
||||||
|
# - Locale and Formatting -
|
||||||
|
|
||||||
|
#datestyle = 'iso, mdy'
|
||||||
|
#intervalstyle = 'postgres'
|
||||||
|
#timezone = 'GMT'
|
||||||
|
#timezone_abbreviations = 'Default' # Select the set of available time zone
|
||||||
|
# abbreviations. Currently, there are
|
||||||
|
# Default
|
||||||
|
# Australia (historical usage)
|
||||||
|
# India
|
||||||
|
# You can create your own file in
|
||||||
|
# share/timezonesets/.
|
||||||
|
#extra_float_digits = 0 # min -15, max 3
|
||||||
|
#client_encoding = sql_ascii # actually, defaults to database
|
||||||
|
# encoding
|
||||||
|
|
||||||
|
# These settings are initialized by initdb, but they can be changed.
|
||||||
|
#lc_messages = 'C' # locale for system error message
|
||||||
|
# strings
|
||||||
|
#lc_monetary = 'C' # locale for monetary formatting
|
||||||
|
#lc_numeric = 'C' # locale for number formatting
|
||||||
|
#lc_time = 'C' # locale for time formatting
|
||||||
|
|
||||||
|
# default configuration for text search
|
||||||
|
#default_text_search_config = 'pg_catalog.simple'
|
||||||
|
|
||||||
|
# - Other Defaults -
|
||||||
|
|
||||||
|
#dynamic_library_path = '$libdir'
|
||||||
|
#local_preload_libraries = ''
|
||||||
|
#session_preload_libraries = ''
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# LOCK MANAGEMENT
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
#deadlock_timeout = 1s
|
||||||
|
#max_locks_per_transaction = 64 # min 10
|
||||||
|
# (change requires restart)
|
||||||
|
#max_pred_locks_per_transaction = 64 # min 10
|
||||||
|
# (change requires restart)
|
||||||
|
#max_pred_locks_per_relation = -2 # negative values mean
|
||||||
|
# (max_pred_locks_per_transaction
|
||||||
|
# / -max_pred_locks_per_relation) - 1
|
||||||
|
#max_pred_locks_per_page = 2 # min 0
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# VERSION/PLATFORM COMPATIBILITY
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# - Previous PostgreSQL Versions -
|
||||||
|
|
||||||
|
#array_nulls = on
|
||||||
|
#backslash_quote = safe_encoding # on, off, or safe_encoding
|
||||||
|
#default_with_oids = off
|
||||||
|
#escape_string_warning = on
|
||||||
|
#lo_compat_privileges = off
|
||||||
|
#operator_precedence_warning = off
|
||||||
|
#quote_all_identifiers = off
|
||||||
|
#standard_conforming_strings = on
|
||||||
|
#synchronize_seqscans = on
|
||||||
|
|
||||||
|
# - Other Platforms and Clients -
|
||||||
|
|
||||||
|
#transform_null_equals = off
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# ERROR HANDLING
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
#exit_on_error = off # terminate session on any error?
|
||||||
|
#restart_after_crash = on # reinitialize after backend crash?
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# CONFIG FILE INCLUDES
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# These options allow settings to be loaded from files other than the
|
||||||
|
# default postgresql.conf.
|
||||||
|
|
||||||
|
#include_dir = 'conf.d' # include files ending in '.conf' from
|
||||||
|
# directory 'conf.d'
|
||||||
|
#include_if_exists = 'exists.conf' # include file only if it exists
|
||||||
|
#include = 'special.conf' # include file
|
||||||
|
|
||||||
|
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
# CUSTOMIZED OPTIONS
|
||||||
|
#------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# Add settings for extensions here
|
||||||
|
|
||||||
|
zdb.default_elasticsearch_url = 'http://elasticsearch:9200/'
|
||||||
4
docker/postgresql/setup.sh
Normal file
4
docker/postgresql/setup.sh
Normal file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/usr/bin/env bash
|
||||||
|
set -e
|
||||||
|
|
||||||
|
cat /tmp/postgresql.conf > /var/lib/postgresql/data/postgresql.conf
|
||||||
2
gradle/wrapper/gradle-wrapper.properties
vendored
2
gradle/wrapper/gradle-wrapper.properties
vendored
@@ -1,4 +1,4 @@
|
|||||||
distributionUrl=https\://services.gradle.org/distributions/gradle-5.6.2-all.zip
|
distributionUrl=https\://services.gradle.org/distributions/gradle-6.8-bin.zip
|
||||||
distributionBase=GRADLE_USER_HOME
|
distributionBase=GRADLE_USER_HOME
|
||||||
distributionPath=wrapper/dists
|
distributionPath=wrapper/dists
|
||||||
zipStorePath=wrapper/dists
|
zipStorePath=wrapper/dists
|
||||||
|
|||||||
2
jitpack.yml
Normal file
2
jitpack.yml
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
jdk:
|
||||||
|
- openjdk11
|
||||||
@@ -12,7 +12,7 @@ import fr.postgresjson.entity.Serializable
|
|||||||
import fr.postgresjson.serializer.Serializer
|
import fr.postgresjson.serializer.Serializer
|
||||||
import fr.postgresjson.utils.LoggerDelegate
|
import fr.postgresjson.utils.LoggerDelegate
|
||||||
import org.slf4j.Logger
|
import org.slf4j.Logger
|
||||||
import java.util.concurrent.*
|
import java.util.concurrent.CompletableFuture
|
||||||
|
|
||||||
typealias SelectOneCallback<T> = QueryResult.(T?) -> Unit
|
typealias SelectOneCallback<T> = QueryResult.(T?) -> Unit
|
||||||
typealias SelectCallback<T> = QueryResult.(List<T>) -> Unit
|
typealias SelectCallback<T> = QueryResult.(List<T>) -> Unit
|
||||||
@@ -25,17 +25,26 @@ class Connection(
|
|||||||
private val host: String = "localhost",
|
private val host: String = "localhost",
|
||||||
private val port: Int = 5432
|
private val port: Int = 5432
|
||||||
) : Executable {
|
) : Executable {
|
||||||
private lateinit var connection: ConnectionPool<PostgreSQLConnection>
|
private var connection: ConnectionPool<PostgreSQLConnection>? = null
|
||||||
private val serializer = Serializer()
|
private val serializer = Serializer()
|
||||||
private val logger: Logger? by LoggerDelegate()
|
private val logger: Logger? by LoggerDelegate()
|
||||||
|
|
||||||
internal fun connect(): ConnectionPool<PostgreSQLConnection> {
|
internal fun connect(): ConnectionPool<PostgreSQLConnection> {
|
||||||
if (!::connection.isInitialized || !connection.isConnected()) {
|
return connection.let { connectionPool ->
|
||||||
connection = PostgreSQLConnectionBuilder.createConnectionPool(
|
if (connectionPool == null || !connectionPool.isConnected()) {
|
||||||
"jdbc:postgresql://$host:$port/$database?user=$username&password=$password"
|
PostgreSQLConnectionBuilder.createConnectionPool(
|
||||||
)
|
"jdbc:postgresql://$host:$port/$database?user=$username&password=$password"
|
||||||
|
).also {
|
||||||
|
connection = it
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
connectionPool
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return connection
|
}
|
||||||
|
|
||||||
|
fun disconnect() {
|
||||||
|
connection?.run { disconnect() }
|
||||||
}
|
}
|
||||||
|
|
||||||
fun <A> inTransaction(f: (Connection) -> CompletableFuture<A>) = connect().inTransaction(f)
|
fun <A> inTransaction(f: (Connection) -> CompletableFuture<A>) = connect().inTransaction(f)
|
||||||
@@ -50,7 +59,7 @@ class Connection(
|
|||||||
it is EntityI && typeReference.type.typeName == it::class.java.name
|
it is EntityI && typeReference.type.typeName == it::class.java.name
|
||||||
} as R?
|
} as R?
|
||||||
val result = exec(sql, compileArgs(values))
|
val result = exec(sql, compileArgs(values))
|
||||||
val json = result.rows[0].getString(0)
|
val json = result.rows.firstOrNull()?.getString(0)
|
||||||
return if (json === null) {
|
return if (json === null) {
|
||||||
null
|
null
|
||||||
} else {
|
} else {
|
||||||
@@ -271,11 +280,14 @@ class Connection(
|
|||||||
logger?.debug("Query executed in $duration ms \n{}", args)
|
logger?.debug("Query executed in $duration ms \n{}", args)
|
||||||
return result
|
return result
|
||||||
} catch (e: Throwable) {
|
} catch (e: Throwable) {
|
||||||
logger?.info("""
|
logger?.info(
|
||||||
|
"""
|
||||||
Query Error:
|
Query Error:
|
||||||
${sql.prependIndent()},
|
${sql.prependIndent()},
|
||||||
${values.joinToString(", ").prependIndent()}
|
${values.joinToString(", ").prependIndent()}
|
||||||
""".trimIndent(), e)
|
""".trimIndent(),
|
||||||
|
e
|
||||||
|
)
|
||||||
throw e
|
throw e
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,7 +1,9 @@
|
|||||||
package fr.postgresjson.connexion
|
package fr.postgresjson.connexion
|
||||||
|
|
||||||
import java.io.File
|
import fr.postgresjson.utils.searchSqlFiles
|
||||||
|
import java.net.URI
|
||||||
import fr.postgresjson.definition.Function as DefinitionFunction
|
import fr.postgresjson.definition.Function as DefinitionFunction
|
||||||
|
import fr.postgresjson.definition.Query as QueryDefinition
|
||||||
|
|
||||||
class Requester(
|
class Requester(
|
||||||
private val connection: Connection,
|
private val connection: Connection,
|
||||||
@@ -13,17 +15,19 @@ class Requester(
|
|||||||
return this
|
return this
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fun addQuery(query: QueryDefinition): Requester = addQuery(query.name, query.script)
|
||||||
|
|
||||||
fun addQuery(name: String, sql: String): Requester {
|
fun addQuery(name: String, sql: String): Requester {
|
||||||
addQuery(Query(name, sql, connection))
|
addQuery(Query(name, sql, connection))
|
||||||
return this
|
return this
|
||||||
}
|
}
|
||||||
|
|
||||||
fun addQuery(queriesDirectory: File): Requester {
|
fun addQuery(queriesDirectory: URI): Requester {
|
||||||
queriesDirectory.walk()
|
queriesDirectory.searchSqlFiles()
|
||||||
.filter { it.isFile && it.extension == "sql" }
|
|
||||||
.forEach {
|
.forEach {
|
||||||
val path = it.parentFile.nameWithoutExtension
|
if (it is QueryDefinition) {
|
||||||
addQuery("$path/${it.nameWithoutExtension}", it.readText())
|
addQuery(it)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return this
|
return this
|
||||||
}
|
}
|
||||||
@@ -44,11 +48,12 @@ class Requester(
|
|||||||
return this
|
return this
|
||||||
}
|
}
|
||||||
|
|
||||||
fun addFunction(functionsDirectory: File): Requester {
|
fun addFunction(functionsDirectory: URI): Requester {
|
||||||
functionsDirectory.walk()
|
functionsDirectory.searchSqlFiles()
|
||||||
.filter { it.isFile && it.extension == "sql" }
|
|
||||||
.forEach {
|
.forEach {
|
||||||
addFunction(it.readText())
|
if (it is DefinitionFunction) {
|
||||||
|
addFunction(it)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return this
|
return this
|
||||||
}
|
}
|
||||||
@@ -69,8 +74,8 @@ class Requester(
|
|||||||
|
|
||||||
class RequesterFactory(
|
class RequesterFactory(
|
||||||
private val connection: Connection,
|
private val connection: Connection,
|
||||||
private val queriesDirectory: File? = null,
|
private val queriesDirectory: URI? = null,
|
||||||
private val functionsDirectory: File? = null
|
private val functionsDirectory: URI? = null
|
||||||
) {
|
) {
|
||||||
constructor(
|
constructor(
|
||||||
host: String = "localhost",
|
host: String = "localhost",
|
||||||
@@ -78,8 +83,8 @@ class Requester(
|
|||||||
database: String = "dc-project",
|
database: String = "dc-project",
|
||||||
username: String = "dc-project",
|
username: String = "dc-project",
|
||||||
password: String = "dc-project",
|
password: String = "dc-project",
|
||||||
queriesDirectory: File? = null,
|
queriesDirectory: URI? = null,
|
||||||
functionsDirectory: File? = null
|
functionsDirectory: URI? = null
|
||||||
) : this(
|
) : this(
|
||||||
Connection(host = host, port = port, database = database, username = username, password = password),
|
Connection(host = host, port = port, database = database, username = username, password = password),
|
||||||
queriesDirectory,
|
queriesDirectory,
|
||||||
|
|||||||
@@ -1,14 +1,15 @@
|
|||||||
package fr.postgresjson.definition
|
package fr.postgresjson.definition
|
||||||
|
|
||||||
import java.io.File
|
import java.io.File
|
||||||
|
import java.nio.file.Path
|
||||||
|
|
||||||
open class Function(
|
class Function(
|
||||||
override val script: String
|
override val script: String,
|
||||||
|
override var source: Path? = null
|
||||||
) : Resource, ParametersInterface {
|
) : Resource, ParametersInterface {
|
||||||
val returns: String
|
val returns: String
|
||||||
override val name: String
|
override val name: String
|
||||||
override val parameters: List<Parameter>
|
override val parameters: List<Parameter>
|
||||||
override var source: File? = null
|
|
||||||
|
|
||||||
init {
|
init {
|
||||||
val functionRegex =
|
val functionRegex =
|
||||||
@@ -46,8 +47,7 @@ open class Function(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
abstract class ParseException(message: String, cause: Throwable? = null) : Exception(message, cause)
|
class FunctionNotFound(cause: Throwable? = null) : Resource.ParseException("Function not found in script", cause)
|
||||||
class FunctionNotFound(cause: Throwable? = null) : ParseException("Function not found in script", cause)
|
|
||||||
|
|
||||||
fun getDefinition(): String {
|
fun getDefinition(): String {
|
||||||
return parameters
|
return parameters
|
||||||
|
|||||||
35
src/main/kotlin/fr/postgresjson/definition/Migration.kt
Normal file
35
src/main/kotlin/fr/postgresjson/definition/Migration.kt
Normal file
@@ -0,0 +1,35 @@
|
|||||||
|
package fr.postgresjson.definition
|
||||||
|
|
||||||
|
import java.nio.file.Path
|
||||||
|
|
||||||
|
class Migration(
|
||||||
|
override val script: String,
|
||||||
|
source: Path
|
||||||
|
) : Resource {
|
||||||
|
override val name: String
|
||||||
|
val direction: Direction
|
||||||
|
override var source: Path? = null
|
||||||
|
|
||||||
|
init {
|
||||||
|
this.source = source
|
||||||
|
this.direction = source.fileName.toString()
|
||||||
|
.let {
|
||||||
|
when {
|
||||||
|
it.endsWith(".down.sql") -> Direction.DOWN
|
||||||
|
it.endsWith(".up.sql") -> Direction.UP
|
||||||
|
else -> throw MigrationNotFound()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
this.name = source.fileName.toString()
|
||||||
|
.substringAfterLast("/")
|
||||||
|
.let {
|
||||||
|
when (direction) {
|
||||||
|
Direction.DOWN -> it.substringBefore(".down.sql")
|
||||||
|
Direction.UP -> it.substringBefore(".up.sql")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class MigrationNotFound(cause: Throwable? = null) : Resource.ParseException("Migration not found in script", cause)
|
||||||
|
enum class Direction { UP, DOWN }
|
||||||
|
}
|
||||||
25
src/main/kotlin/fr/postgresjson/definition/Query.kt
Normal file
25
src/main/kotlin/fr/postgresjson/definition/Query.kt
Normal file
@@ -0,0 +1,25 @@
|
|||||||
|
package fr.postgresjson.definition
|
||||||
|
|
||||||
|
import java.nio.file.Path
|
||||||
|
|
||||||
|
class Query(
|
||||||
|
override val script: String,
|
||||||
|
source: Path
|
||||||
|
) : Resource {
|
||||||
|
override var source: Path? = source
|
||||||
|
override val name: String = getNameFromComment(script) ?: getNameFromFile(source)
|
||||||
|
|
||||||
|
/** Try to get name from comment in file */
|
||||||
|
private fun getNameFromComment(script: String): String? =
|
||||||
|
"""-- *name ?: ?(?<name>[^ \n]+)"""
|
||||||
|
.toRegex(setOf(RegexOption.IGNORE_CASE, RegexOption.MULTILINE))
|
||||||
|
.find(script)?.let {
|
||||||
|
it.groups["name"]?.value?.trim()
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Try to get name from the filename */
|
||||||
|
private fun getNameFromFile(source: Path): String = source
|
||||||
|
.fileName.toString()
|
||||||
|
.substringAfterLast("/")
|
||||||
|
.substringBeforeLast(".sql")
|
||||||
|
}
|
||||||
@@ -1,11 +1,38 @@
|
|||||||
package fr.postgresjson.definition
|
package fr.postgresjson.definition
|
||||||
|
|
||||||
import java.io.File
|
import java.io.File
|
||||||
|
import java.net.URL
|
||||||
|
import java.nio.file.Path
|
||||||
|
|
||||||
interface Resource {
|
interface Resource {
|
||||||
val name: String
|
val name: String
|
||||||
val script: String
|
val script: String
|
||||||
var source: File?
|
var source: Path?
|
||||||
|
|
||||||
|
open class ParseException(message: String, cause: Throwable? = null) : Exception(message, cause)
|
||||||
|
|
||||||
|
companion object {
|
||||||
|
fun build(file: File): Resource =
|
||||||
|
build(file.readText(), Path.of(file.toURI()))
|
||||||
|
|
||||||
|
fun build(url: URL): Resource =
|
||||||
|
build(url.readText(), Path.of(url.toURI()))
|
||||||
|
|
||||||
|
fun build(resource: String, path: Path): Resource =
|
||||||
|
try {
|
||||||
|
Migration(resource, path)
|
||||||
|
} catch (e: ParseException) {
|
||||||
|
try {
|
||||||
|
Function(resource, path)
|
||||||
|
} catch (e: ParseException) {
|
||||||
|
try {
|
||||||
|
Query(resource, path)
|
||||||
|
} catch (e: ParseException) {
|
||||||
|
throw ParseException("No SQL resource found")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
interface ResourceCollection {
|
interface ResourceCollection {
|
||||||
|
|||||||
@@ -1,12 +1,7 @@
|
|||||||
package fr.postgresjson.entity.immutable
|
package fr.postgresjson.entity
|
||||||
|
|
||||||
import fr.postgresjson.entity.EntityI
|
|
||||||
import fr.postgresjson.entity.mutable.EntityDeletedAt
|
|
||||||
import fr.postgresjson.entity.mutable.EntityDeletedAtImp
|
|
||||||
import fr.postgresjson.entity.mutable.EntityDeletedBy
|
|
||||||
import fr.postgresjson.entity.mutable.EntityDeletedByImp
|
|
||||||
import org.joda.time.DateTime
|
import org.joda.time.DateTime
|
||||||
import java.util.*
|
import java.util.UUID
|
||||||
|
|
||||||
interface EntityRefI<T> : EntityI {
|
interface EntityRefI<T> : EntityI {
|
||||||
val id: T
|
val id: T
|
||||||
@@ -17,7 +12,7 @@ interface UuidEntityI : EntityRefI<UUID> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
abstract class Entity<T>(override val id: T) : EntityRefI<T>
|
abstract class Entity<T>(override val id: T) : EntityRefI<T>
|
||||||
open class UuidEntity(override val id: UUID = UUID.randomUUID()) : UuidEntityI, Entity<UUID>(id)
|
open class UuidEntity(id: UUID? = null) : UuidEntityI, Entity<UUID>(id ?: UUID.randomUUID())
|
||||||
|
|
||||||
/* Version */
|
/* Version */
|
||||||
interface EntityVersioning<ID, NUMBER> {
|
interface EntityVersioning<ID, NUMBER> {
|
||||||
@@ -26,16 +21,27 @@ interface EntityVersioning<ID, NUMBER> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
class UuidEntityVersioning(
|
class UuidEntityVersioning(
|
||||||
override var versionNumber: Int? = null,
|
override val versionNumber: Int,
|
||||||
override val versionId: UUID = UUID.randomUUID()
|
versionId: UUID? = null
|
||||||
) : EntityVersioning<UUID, Int?>
|
) : EntityVersioning<UUID, Int> {
|
||||||
|
override val versionId: UUID = versionId ?: UUID.randomUUID()
|
||||||
|
}
|
||||||
|
|
||||||
/* Dates */
|
/* Dates */
|
||||||
interface EntityCreatedAt {
|
interface EntityCreatedAt {
|
||||||
val createdAt: DateTime
|
val createdAt: DateTime
|
||||||
}
|
}
|
||||||
interface EntityUpdatedAt {
|
interface EntityUpdatedAt {
|
||||||
var updatedAt: DateTime
|
val updatedAt: DateTime
|
||||||
|
}
|
||||||
|
|
||||||
|
interface EntityDeletedAt {
|
||||||
|
val deletedAt: DateTime?
|
||||||
|
fun isDeleted(): Boolean {
|
||||||
|
return deletedAt?.let {
|
||||||
|
it < DateTime.now()
|
||||||
|
} ?: false
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
class EntityCreatedAtImp(
|
class EntityCreatedAtImp(
|
||||||
@@ -43,15 +49,23 @@ class EntityCreatedAtImp(
|
|||||||
) : EntityCreatedAt
|
) : EntityCreatedAt
|
||||||
|
|
||||||
class EntityUpdatedAtImp(
|
class EntityUpdatedAtImp(
|
||||||
override var updatedAt: DateTime = DateTime.now()
|
override val updatedAt: DateTime = DateTime.now()
|
||||||
) : EntityUpdatedAt
|
) : EntityUpdatedAt
|
||||||
|
|
||||||
|
class EntityDeletedAtImp(
|
||||||
|
override val deletedAt: DateTime? = null
|
||||||
|
) : EntityDeletedAt
|
||||||
|
|
||||||
/* Author */
|
/* Author */
|
||||||
interface EntityCreatedBy<T : EntityI> {
|
interface EntityCreatedBy<T : EntityI> {
|
||||||
val createdBy: T
|
val createdBy: T
|
||||||
}
|
}
|
||||||
interface EntityUpdatedBy<T : EntityI> {
|
interface EntityUpdatedBy<T : EntityI> {
|
||||||
var updatedBy: T
|
val updatedBy: T
|
||||||
|
}
|
||||||
|
|
||||||
|
interface EntityDeletedBy<T : EntityI> {
|
||||||
|
val deletedBy: T?
|
||||||
}
|
}
|
||||||
|
|
||||||
class EntityCreatedByImp<UserT : EntityI>(
|
class EntityCreatedByImp<UserT : EntityI>(
|
||||||
@@ -59,9 +73,13 @@ class EntityCreatedByImp<UserT : EntityI>(
|
|||||||
) : EntityCreatedBy<UserT>
|
) : EntityCreatedBy<UserT>
|
||||||
|
|
||||||
class EntityUpdatedByImp<UserT : EntityI>(
|
class EntityUpdatedByImp<UserT : EntityI>(
|
||||||
override var updatedBy: UserT
|
override val updatedBy: UserT
|
||||||
) : EntityUpdatedBy<UserT>
|
) : EntityUpdatedBy<UserT>
|
||||||
|
|
||||||
|
class EntityDeletedByImp<UserT : EntityI>(
|
||||||
|
override val deletedBy: UserT?
|
||||||
|
) : EntityDeletedBy<UserT>
|
||||||
|
|
||||||
/* Mixed */
|
/* Mixed */
|
||||||
class EntityCreatedImp<UserT : EntityI>(
|
class EntityCreatedImp<UserT : EntityI>(
|
||||||
override val createdAt: DateTime = DateTime.now(),
|
override val createdAt: DateTime = DateTime.now(),
|
||||||
@@ -71,10 +89,22 @@ class EntityCreatedImp<UserT : EntityI>(
|
|||||||
|
|
||||||
class EntityUpdatedImp<UserT : EntityI>(
|
class EntityUpdatedImp<UserT : EntityI>(
|
||||||
updatedAt: DateTime = DateTime.now(),
|
updatedAt: DateTime = DateTime.now(),
|
||||||
override var updatedBy: UserT
|
override val updatedBy: UserT
|
||||||
) : EntityUpdatedBy<UserT>,
|
) : EntityUpdatedBy<UserT>,
|
||||||
EntityUpdatedAt by EntityUpdatedAtImp(updatedAt)
|
EntityUpdatedAt by EntityUpdatedAtImp(updatedAt)
|
||||||
|
|
||||||
|
/* Published */
|
||||||
|
interface Published<UserT : EntityI> {
|
||||||
|
val publishedAt: DateTime?
|
||||||
|
val publishedBy: UserT?
|
||||||
|
}
|
||||||
|
|
||||||
|
class EntityPublishedImp<UserT : EntityI>(
|
||||||
|
override val publishedBy: UserT?
|
||||||
|
) : Published<UserT> {
|
||||||
|
override val publishedAt: DateTime? = null
|
||||||
|
}
|
||||||
|
|
||||||
/* Implementation */
|
/* Implementation */
|
||||||
abstract class EntityImp<T, UserT : EntityI>(
|
abstract class EntityImp<T, UserT : EntityI>(
|
||||||
updatedBy: UserT,
|
updatedBy: UserT,
|
||||||
@@ -86,3 +116,11 @@ abstract class EntityImp<T, UserT : EntityI>(
|
|||||||
EntityCreatedBy<UserT> by EntityCreatedByImp(updatedBy),
|
EntityCreatedBy<UserT> by EntityCreatedByImp(updatedBy),
|
||||||
EntityUpdatedBy<UserT> by EntityUpdatedByImp(updatedBy),
|
EntityUpdatedBy<UserT> by EntityUpdatedByImp(updatedBy),
|
||||||
EntityDeletedBy<UserT> by EntityDeletedByImp(updatedBy)
|
EntityDeletedBy<UserT> by EntityDeletedByImp(updatedBy)
|
||||||
|
|
||||||
|
abstract class UuidEntityExtended<T, UserT : EntityI>(
|
||||||
|
updatedBy: UserT,
|
||||||
|
publishedBy: UserT?
|
||||||
|
) :
|
||||||
|
EntityImp<T, UserT>(updatedBy),
|
||||||
|
EntityVersioning<UUID, Int> by UuidEntityVersioning(0),
|
||||||
|
Published<UserT> by EntityPublishedImp(publishedBy)
|
||||||
@@ -1,135 +0,0 @@
|
|||||||
package fr.postgresjson.entity.mutable
|
|
||||||
|
|
||||||
import fr.postgresjson.entity.EntityI
|
|
||||||
import org.joda.time.DateTime
|
|
||||||
import java.util.*
|
|
||||||
|
|
||||||
interface EntityRefI<T> : EntityI {
|
|
||||||
var id: T?
|
|
||||||
}
|
|
||||||
|
|
||||||
interface UuidEntityI : EntityRefI<UUID> {
|
|
||||||
override var id: UUID?
|
|
||||||
}
|
|
||||||
|
|
||||||
interface IdEntityI : EntityRefI<Int> {
|
|
||||||
override var id: Int?
|
|
||||||
}
|
|
||||||
|
|
||||||
abstract class Entity<T>(override var id: T? = null) : EntityRefI<T>
|
|
||||||
open class UuidEntity(override var id: UUID? = UUID.randomUUID()) : UuidEntityI, Entity<UUID>(id)
|
|
||||||
open class IdEntity(override var id: Int? = null) : IdEntityI, Entity<Int>(id)
|
|
||||||
|
|
||||||
/* Version */
|
|
||||||
interface EntityVersioning<ID, NUMBER> {
|
|
||||||
var versionId: ID
|
|
||||||
var versionNumber: NUMBER?
|
|
||||||
}
|
|
||||||
|
|
||||||
class UuidEntityVersioning(
|
|
||||||
override var versionNumber: Int? = null,
|
|
||||||
override var versionId: UUID = UUID.randomUUID()
|
|
||||||
) : EntityVersioning<UUID, Int>
|
|
||||||
|
|
||||||
/* Dates */
|
|
||||||
interface EntityCreatedAt {
|
|
||||||
var createdAt: DateTime?
|
|
||||||
}
|
|
||||||
|
|
||||||
interface EntityUpdatedAt {
|
|
||||||
var updatedAt: DateTime?
|
|
||||||
}
|
|
||||||
|
|
||||||
interface EntityDeletedAt {
|
|
||||||
var deletedAt: DateTime?
|
|
||||||
fun isDeleted(): Boolean {
|
|
||||||
val deletedAt = deletedAt
|
|
||||||
return deletedAt != null && deletedAt < DateTime.now()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
class EntityCreatedAtImp : EntityCreatedAt {
|
|
||||||
override var createdAt: DateTime? = null
|
|
||||||
}
|
|
||||||
|
|
||||||
class EntityUpdatedAtImp : EntityUpdatedAt {
|
|
||||||
override var updatedAt: DateTime? = null
|
|
||||||
}
|
|
||||||
|
|
||||||
class EntityDeletedAtImp : EntityDeletedAt {
|
|
||||||
override var deletedAt: DateTime? = null
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Author */
|
|
||||||
interface EntityCreatedBy<T : EntityI> {
|
|
||||||
var createdBy: T?
|
|
||||||
}
|
|
||||||
|
|
||||||
interface EntityUpdatedBy<T : EntityI> {
|
|
||||||
var updatedBy: T?
|
|
||||||
}
|
|
||||||
|
|
||||||
interface EntityDeletedBy<T : EntityI> {
|
|
||||||
var deletedBy: T?
|
|
||||||
}
|
|
||||||
|
|
||||||
class EntityCreatedByImp<UserT : EntityI>(
|
|
||||||
override var createdBy: UserT?
|
|
||||||
) : EntityCreatedBy<UserT>
|
|
||||||
|
|
||||||
class EntityUpdatedByImp<UserT : EntityI>(
|
|
||||||
override var updatedBy: UserT?
|
|
||||||
) : EntityUpdatedBy<UserT>
|
|
||||||
|
|
||||||
class EntityDeletedByImp<UserT : EntityI>(
|
|
||||||
override var deletedBy: UserT?
|
|
||||||
) : EntityDeletedBy<UserT>
|
|
||||||
|
|
||||||
/* Mixed */
|
|
||||||
class EntityDeletedImp<UserT : EntityI>(
|
|
||||||
override var deletedBy: UserT? = null
|
|
||||||
) : EntityDeletedBy<UserT>,
|
|
||||||
EntityDeletedAt by EntityDeletedAtImp()
|
|
||||||
|
|
||||||
class EntityUpdatedImp<UserT : EntityI>(
|
|
||||||
override var updatedAt: DateTime? = null,
|
|
||||||
override var updatedBy: UserT? = null
|
|
||||||
) : EntityUpdatedBy<UserT>,
|
|
||||||
EntityUpdatedAt by EntityUpdatedAtImp()
|
|
||||||
|
|
||||||
class EntityCreatedImp<UserT : EntityI>(
|
|
||||||
override var createdAt: DateTime? = null,
|
|
||||||
override var createdBy: UserT? = null
|
|
||||||
) : EntityCreatedBy<UserT>,
|
|
||||||
EntityCreatedAt by EntityCreatedAtImp()
|
|
||||||
|
|
||||||
/* Published */
|
|
||||||
interface Published<UserT : EntityI> {
|
|
||||||
var publishedAt: DateTime?
|
|
||||||
var publishedBy: UserT?
|
|
||||||
}
|
|
||||||
|
|
||||||
class EntityPublishedImp<UserT : EntityI>(
|
|
||||||
override var publishedBy: UserT?
|
|
||||||
) : Published<UserT> {
|
|
||||||
override var publishedAt: DateTime? = null
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Implementation */
|
|
||||||
abstract class EntityImp<T, UserT : EntityI>(
|
|
||||||
updatedBy: UserT?
|
|
||||||
) : Entity<T>(),
|
|
||||||
EntityCreatedAt by EntityCreatedAtImp(),
|
|
||||||
EntityUpdatedAt by EntityUpdatedAtImp(),
|
|
||||||
EntityDeletedAt by EntityDeletedAtImp(),
|
|
||||||
EntityCreatedBy<UserT> by EntityCreatedByImp(updatedBy),
|
|
||||||
EntityUpdatedBy<UserT> by EntityUpdatedByImp(updatedBy),
|
|
||||||
EntityDeletedBy<UserT> by EntityDeletedByImp(updatedBy)
|
|
||||||
|
|
||||||
abstract class UuidEntityExtended<T, UserT : EntityI>(
|
|
||||||
updatedBy: UserT?,
|
|
||||||
publishedBy: UserT?
|
|
||||||
) :
|
|
||||||
EntityImp<T, UserT>(updatedBy),
|
|
||||||
EntityVersioning<UUID, Int> by UuidEntityVersioning(),
|
|
||||||
Published<UserT> by EntityPublishedImp(publishedBy)
|
|
||||||
@@ -4,8 +4,8 @@ import com.github.jasync.sql.db.postgresql.exceptions.GenericDatabaseException
|
|||||||
import fr.postgresjson.connexion.Connection
|
import fr.postgresjson.connexion.Connection
|
||||||
import fr.postgresjson.migration.Migration.Action
|
import fr.postgresjson.migration.Migration.Action
|
||||||
import fr.postgresjson.migration.Migration.Status
|
import fr.postgresjson.migration.Migration.Status
|
||||||
import java.util.*
|
import java.util.Date
|
||||||
import java.util.concurrent.*
|
import java.util.concurrent.CompletionException
|
||||||
import fr.postgresjson.definition.Function as DefinitionFunction
|
import fr.postgresjson.definition.Function as DefinitionFunction
|
||||||
|
|
||||||
data class Function(
|
data class Function(
|
||||||
|
|||||||
@@ -1,11 +1,11 @@
|
|||||||
package fr.postgresjson.migration
|
package fr.postgresjson.migration
|
||||||
|
|
||||||
import fr.postgresjson.connexion.Connection
|
import fr.postgresjson.connexion.Connection
|
||||||
import fr.postgresjson.entity.mutable.Entity
|
import fr.postgresjson.entity.Entity
|
||||||
import fr.postgresjson.migration.Migration.Action
|
import fr.postgresjson.migration.Migration.Action
|
||||||
import java.util.*
|
import java.util.Date
|
||||||
|
|
||||||
data class Query(
|
data class MigrationScript(
|
||||||
val name: String,
|
val name: String,
|
||||||
val up: String,
|
val up: String,
|
||||||
val down: String,
|
val down: String,
|
||||||
@@ -57,7 +57,7 @@ data class Query(
|
|||||||
return Migration.Status.OK // TODO
|
return Migration.Status.OK // TODO
|
||||||
}
|
}
|
||||||
|
|
||||||
fun copy(): Query {
|
fun copy(): MigrationScript {
|
||||||
return this.copy(name = name, up = up, down = down, connection = connection, executedAt = executedAt).also {
|
return this.copy(name = name, up = up, down = down, connection = connection, executedAt = executedAt).also {
|
||||||
it.doExecute = this.doExecute
|
it.doExecute = this.doExecute
|
||||||
}
|
}
|
||||||
@@ -2,16 +2,17 @@ package fr.postgresjson.migration
|
|||||||
|
|
||||||
import com.fasterxml.jackson.core.type.TypeReference
|
import com.fasterxml.jackson.core.type.TypeReference
|
||||||
import fr.postgresjson.connexion.Connection
|
import fr.postgresjson.connexion.Connection
|
||||||
import fr.postgresjson.definition.Function.FunctionNotFound
|
import fr.postgresjson.entity.Entity
|
||||||
import fr.postgresjson.entity.mutable.Entity
|
|
||||||
import fr.postgresjson.migration.Migration.Action
|
import fr.postgresjson.migration.Migration.Action
|
||||||
import fr.postgresjson.migration.Migration.Status
|
import fr.postgresjson.migration.Migration.Status
|
||||||
import fr.postgresjson.utils.LoggerDelegate
|
import fr.postgresjson.utils.LoggerDelegate
|
||||||
|
import fr.postgresjson.utils.searchSqlFiles
|
||||||
import org.slf4j.Logger
|
import org.slf4j.Logger
|
||||||
import java.io.File
|
|
||||||
import java.io.FileNotFoundException
|
import java.io.FileNotFoundException
|
||||||
import java.util.*
|
import java.net.URI
|
||||||
|
import java.util.Date
|
||||||
import fr.postgresjson.definition.Function as DefinitionFunction
|
import fr.postgresjson.definition.Function as DefinitionFunction
|
||||||
|
import fr.postgresjson.definition.Migration as DefinitionMigration
|
||||||
|
|
||||||
class MigrationEntity(
|
class MigrationEntity(
|
||||||
val filename: String,
|
val filename: String,
|
||||||
@@ -35,27 +36,28 @@ interface Migration {
|
|||||||
|
|
||||||
data class Migrations private constructor(
|
data class Migrations private constructor(
|
||||||
private val connection: Connection,
|
private val connection: Connection,
|
||||||
private val queries: MutableMap<String, Query> = mutableMapOf(),
|
private val migrationsScripts: MutableMap<String, MigrationScript> = mutableMapOf(),
|
||||||
private val functions: MutableMap<String, Function> = mutableMapOf()
|
private val functions: MutableMap<String, Function> = mutableMapOf()
|
||||||
) {
|
) {
|
||||||
private var directories: List<File> = emptyList()
|
private var directories: List<URI> = emptyList()
|
||||||
private val logger: Logger? by LoggerDelegate()
|
private val logger: Logger? by LoggerDelegate()
|
||||||
constructor(directory: File, connection: Connection) : this(listOf(directory), connection)
|
constructor(directory: URI, connection: Connection) : this(listOf(directory), connection)
|
||||||
|
constructor(connection: Connection, vararg directory: URI) : this(directory.toList(), connection)
|
||||||
|
|
||||||
constructor(directories: List<File>, connection: Connection) : this(connection) {
|
constructor(directories: List<URI>, connection: Connection) : this(connection) {
|
||||||
initDB()
|
initDB()
|
||||||
this.directories = directories
|
this.directories = directories
|
||||||
reset()
|
reset()
|
||||||
}
|
}
|
||||||
|
|
||||||
fun reset() {
|
fun reset() {
|
||||||
queries.clear()
|
migrationsScripts.clear()
|
||||||
functions.clear()
|
functions.clear()
|
||||||
|
|
||||||
getMigrationFromDB()
|
getMigrationFromDB()
|
||||||
getMigrationFromDirectory(directories)
|
getMigrationFromDirectory(directories)
|
||||||
|
|
||||||
queries.forEach { (_, query) ->
|
migrationsScripts.forEach { (_, query) ->
|
||||||
if (query.doExecute === null) {
|
if (query.doExecute === null) {
|
||||||
query.doExecute = Action.DOWN
|
query.doExecute = Action.DOWN
|
||||||
}
|
}
|
||||||
@@ -84,7 +86,7 @@ data class Migrations private constructor(
|
|||||||
this::class.java.classLoader.getResource("sql/migration/findAllHistory.sql")!!.readText().let {
|
this::class.java.classLoader.getResource("sql/migration/findAllHistory.sql")!!.readText().let {
|
||||||
connection.select<MigrationEntity>(it, object : TypeReference<List<MigrationEntity>>() {})
|
connection.select<MigrationEntity>(it, object : TypeReference<List<MigrationEntity>>() {})
|
||||||
.map { query ->
|
.map { query ->
|
||||||
queries[query.filename] = Query(query.filename, query.up, query.down, connection, query.executedAt)
|
migrationsScripts[query.filename] = MigrationScript(query.filename, query.up, query.down, connection, query.executedAt)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -92,7 +94,7 @@ data class Migrations private constructor(
|
|||||||
/**
|
/**
|
||||||
* Get all migration from multiples Directories
|
* Get all migration from multiples Directories
|
||||||
*/
|
*/
|
||||||
private fun getMigrationFromDirectory(directory: List<File>) {
|
private fun getMigrationFromDirectory(directory: List<URI>) {
|
||||||
directory.forEach {
|
directory.forEach {
|
||||||
getMigrationFromDirectory(it)
|
getMigrationFromDirectory(it)
|
||||||
}
|
}
|
||||||
@@ -101,29 +103,26 @@ data class Migrations private constructor(
|
|||||||
/**
|
/**
|
||||||
* Get all migration from Directory
|
* Get all migration from Directory
|
||||||
*/
|
*/
|
||||||
private fun getMigrationFromDirectory(directory: File) {
|
private fun getMigrationFromDirectory(directory: URI) {
|
||||||
directory.walk().filter {
|
val downs: MutableMap<String, DefinitionMigration> = mutableMapOf()
|
||||||
it.isFile
|
|
||||||
}.forEach { file ->
|
directory.searchSqlFiles().apply {
|
||||||
if (file.name.endsWith(".up.sql")) {
|
/* Set Down Migration */
|
||||||
file.path.substring(0, file.path.length - 7).let {
|
forEach { migration ->
|
||||||
try {
|
if (migration is DefinitionMigration && migration.direction == DefinitionMigration.Direction.DOWN) {
|
||||||
val down = File("$it.down.sql").readText()
|
downs += migration.name to migration
|
||||||
val up = file.readText()
|
|
||||||
val name = file.name.substring(0, file.name.length - 7)
|
|
||||||
addQuery(name, up, down)
|
|
||||||
} catch (e: FileNotFoundException) {
|
|
||||||
throw DownMigrationNotDefined("$it.down.sql", e)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
} else if (file.name.endsWith(".down.sql")) {
|
}
|
||||||
// Nothing
|
|
||||||
} else {
|
/* Set up migrations and functions */
|
||||||
val fileContent = file.readText()
|
forEach { migration ->
|
||||||
try {
|
if (migration is DefinitionMigration && migration.direction == DefinitionMigration.Direction.UP) {
|
||||||
addFunction(fileContent)
|
val down = downs[migration.name] ?: throw DownMigrationNotDefined(migration.name + ".down.sql")
|
||||||
} catch (e: FunctionNotFound) {
|
downs -= migration.name
|
||||||
// Nothing
|
|
||||||
|
addMigrationScript(migration, down)
|
||||||
|
} else if (migration is DefinitionFunction) {
|
||||||
|
addFunction(migration)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -131,7 +130,7 @@ data class Migrations private constructor(
|
|||||||
|
|
||||||
enum class Direction { UP, DOWN }
|
enum class Direction { UP, DOWN }
|
||||||
|
|
||||||
internal class DownMigrationNotDefined(path: String, cause: FileNotFoundException) :
|
internal class DownMigrationNotDefined(path: String, cause: FileNotFoundException? = null) :
|
||||||
Throwable("The file $path whas not found", cause)
|
Throwable("The file $path whas not found", cause)
|
||||||
|
|
||||||
fun addFunction(newDefinition: DefinitionFunction, callback: (Function) -> Unit = {}): Migrations {
|
fun addFunction(newDefinition: DefinitionFunction, callback: (Function) -> Unit = {}): Migrations {
|
||||||
@@ -155,18 +154,21 @@ data class Migrations private constructor(
|
|||||||
return this
|
return this
|
||||||
}
|
}
|
||||||
|
|
||||||
fun addQuery(name: String, up: String, down: String, callback: (Query) -> Unit = {}): Migrations {
|
fun addMigrationScript(up: DefinitionMigration, down: DefinitionMigration, callback: (MigrationScript) -> Unit = {}): Migrations =
|
||||||
if (queries[name] === null) {
|
addMigrationScript(up.name, up.script, down.script, callback)
|
||||||
queries[name] = Query(name, up, down, connection).apply {
|
|
||||||
|
fun addMigrationScript(name: String, up: String, down: String, callback: (MigrationScript) -> Unit = {}): Migrations {
|
||||||
|
if (migrationsScripts[name] === null) {
|
||||||
|
migrationsScripts[name] = MigrationScript(name, up, down, connection).apply {
|
||||||
doExecute = Action.UP
|
doExecute = Action.UP
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
queries[name]!!.apply {
|
migrationsScripts[name]!!.apply {
|
||||||
doExecute = Action.OK
|
doExecute = Action.OK
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
callback(queries[name]!!)
|
callback(migrationsScripts[name]!!)
|
||||||
|
|
||||||
return this
|
return this
|
||||||
}
|
}
|
||||||
@@ -191,7 +193,7 @@ data class Migrations private constructor(
|
|||||||
|
|
||||||
internal fun up(): Map<String, Status> {
|
internal fun up(): Map<String, Status> {
|
||||||
val list: MutableMap<String, Status> = mutableMapOf()
|
val list: MutableMap<String, Status> = mutableMapOf()
|
||||||
queries.forEach {
|
migrationsScripts.forEach {
|
||||||
it.value.let { query ->
|
it.value.let { query ->
|
||||||
if (query.doExecute == Action.UP) {
|
if (query.doExecute == Action.UP) {
|
||||||
query.up().let { status ->
|
query.up().let { status ->
|
||||||
@@ -216,7 +218,7 @@ data class Migrations private constructor(
|
|||||||
|
|
||||||
internal fun down(force: Boolean = false): Map<String, Status> {
|
internal fun down(force: Boolean = false): Map<String, Status> {
|
||||||
val list: MutableMap<String, Status> = mutableMapOf()
|
val list: MutableMap<String, Status> = mutableMapOf()
|
||||||
queries.forEach {
|
migrationsScripts.forEach {
|
||||||
it.value.let { query ->
|
it.value.let { query ->
|
||||||
if (query.doExecute == Action.DOWN || force) {
|
if (query.doExecute == Action.DOWN || force) {
|
||||||
query.down().let { status ->
|
query.down().let { status ->
|
||||||
@@ -297,7 +299,7 @@ data class Migrations private constructor(
|
|||||||
}
|
}
|
||||||
|
|
||||||
fun copy(): Migrations {
|
fun copy(): Migrations {
|
||||||
val queriesCopy = queries.map {
|
val queriesCopy = migrationsScripts.map {
|
||||||
it.key to it.value.copy()
|
it.key to it.value.copy()
|
||||||
}.toMap().toMutableMap()
|
}.toMap().toMutableMap()
|
||||||
|
|
||||||
|
|||||||
51
src/main/kotlin/fr/postgresjson/utils/searchSqlFiles.kt
Normal file
51
src/main/kotlin/fr/postgresjson/utils/searchSqlFiles.kt
Normal file
@@ -0,0 +1,51 @@
|
|||||||
|
package fr.postgresjson.utils
|
||||||
|
|
||||||
|
import fr.postgresjson.definition.Resource
|
||||||
|
import org.slf4j.Logger
|
||||||
|
import org.slf4j.LoggerFactory
|
||||||
|
import java.net.URI
|
||||||
|
import java.net.URL
|
||||||
|
import java.nio.file.FileSystemNotFoundException
|
||||||
|
import java.nio.file.FileSystems
|
||||||
|
import java.nio.file.FileVisitOption
|
||||||
|
import java.nio.file.Files
|
||||||
|
import java.nio.file.Path
|
||||||
|
import kotlin.streams.asSequence
|
||||||
|
|
||||||
|
fun URL.searchSqlFiles() = this.toURI().searchSqlFiles()
|
||||||
|
|
||||||
|
fun URI.searchSqlFiles() = sequence<Resource> {
|
||||||
|
val logger: Logger = LoggerFactory.getLogger("sqlFilesSearch")
|
||||||
|
val uri: URI = this@searchSqlFiles
|
||||||
|
logger.debug("""SQL files found in "${uri.toString().substringAfter('!')}" :""")
|
||||||
|
if (uri.scheme == "jar") {
|
||||||
|
try {
|
||||||
|
FileSystems.getFileSystem(uri)
|
||||||
|
} catch (e: FileSystemNotFoundException) {
|
||||||
|
FileSystems.newFileSystem(uri, emptyMap<String, Any>())
|
||||||
|
}
|
||||||
|
|
||||||
|
uri
|
||||||
|
.walk(5)
|
||||||
|
.asSequence()
|
||||||
|
.filter { it.fileName.toString().endsWith(".sql") }
|
||||||
|
.map { it.toUri().toURL() }
|
||||||
|
.forEach {
|
||||||
|
logger.debug(it.toString())
|
||||||
|
yield(Resource.build(it))
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
uri
|
||||||
|
.walk(5)
|
||||||
|
.asSequence()
|
||||||
|
.map { it.toFile() }
|
||||||
|
.filter { it.isFile && it.extension == "sql" }
|
||||||
|
.forEach {
|
||||||
|
logger.debug(it.toString())
|
||||||
|
yield(Resource.build(it))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private fun Path.walk(maxDepth: Int = 2147483647, vararg options: FileVisitOption) = Files.walk(this, maxDepth, *options)
|
||||||
|
private fun URI.walk(maxDepth: Int = 2147483647, vararg options: FileVisitOption) = Files.walk(Path.of(this), maxDepth, *options)
|
||||||
@@ -1,70 +1,73 @@
|
|||||||
package fr.postgresjson
|
package fr.postgresjson
|
||||||
|
|
||||||
import fr.postgresjson.connexion.Paginated
|
import fr.postgresjson.connexion.Paginated
|
||||||
import fr.postgresjson.entity.mutable.IdEntity
|
|
||||||
import fr.postgresjson.entity.Parameter
|
import fr.postgresjson.entity.Parameter
|
||||||
import org.junit.Assert.*
|
import fr.postgresjson.entity.UuidEntity
|
||||||
|
import org.junit.Assert.assertEquals
|
||||||
|
import org.junit.Assert.assertNotNull
|
||||||
|
import org.junit.Assert.assertTrue
|
||||||
import org.junit.jupiter.api.Assertions
|
import org.junit.jupiter.api.Assertions
|
||||||
import org.junit.jupiter.api.Test
|
import org.junit.jupiter.api.Test
|
||||||
import org.junit.jupiter.api.TestInstance
|
import org.junit.jupiter.api.TestInstance
|
||||||
|
import java.util.UUID
|
||||||
|
|
||||||
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
||||||
class ConnectionTest() : TestAbstract() {
|
class ConnectionTest() : TestAbstract() {
|
||||||
private class ObjTest(var name: String) : IdEntity()
|
private class ObjTest(val name: String, id: UUID = UUID.fromString("2c0243ed-ff4d-4b9f-a52b-e38c71b0ed00")) : UuidEntity(id)
|
||||||
private class ObjTest2(var title: String, var test: ObjTest?) : IdEntity()
|
private class ObjTest2(val title: String, var test: ObjTest?) : UuidEntity()
|
||||||
private class ObjTest3(var first: String, var seconde: String, var third: Int) : IdEntity()
|
private class ObjTest3(val first: String, var seconde: String, var third: Int) : UuidEntity()
|
||||||
private class ObjTestWithParameterObject(var first: ParameterObject, var seconde: ParameterObject) : IdEntity()
|
private class ObjTestWithParameterObject(var first: ParameterObject, var seconde: ParameterObject) : UuidEntity()
|
||||||
private class ParameterObject(var third: String) : Parameter
|
private class ParameterObject(var third: String) : Parameter
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun getObject() {
|
fun getObject() {
|
||||||
val obj: ObjTest? = connection.selectOne("select to_json(a) from test a limit 1")
|
val obj: ObjTest? = connection.selectOne("select to_json(a) from test a limit 1")
|
||||||
assertTrue(obj is ObjTest)
|
assertTrue(obj is ObjTest)
|
||||||
assertTrue(obj!!.id == 1)
|
assertTrue(obj!!.id == UUID.fromString("1e5f5d41-6d14-4007-897b-0ed2616bec96"))
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun getExistingObject() {
|
fun getExistingObject() {
|
||||||
val objs: List<ObjTest2> = connection.select("""
|
val objs: List<ObjTest2> = connection.select(
|
||||||
select
|
"""
|
||||||
json_agg(j)
|
select
|
||||||
FROM (
|
json_agg(j)
|
||||||
SELECT
|
FROM (
|
||||||
t.id, t.title,
|
SELECT
|
||||||
t2 as test
|
t.id, t.title,
|
||||||
from test2 t
|
t2 as test
|
||||||
JOIN test t2 ON t.test_id = t2.id
|
from test2 t
|
||||||
) j;
|
JOIN test t2 ON t.test_id = t2.id
|
||||||
""".trimIndent()
|
) j;
|
||||||
|
""".trimIndent()
|
||||||
)
|
)
|
||||||
assertNotNull(objs)
|
assertNotNull(objs)
|
||||||
assertEquals(objs.size, 2)
|
assertEquals(objs.size, 2)
|
||||||
assertEquals(objs[0].id, 1)
|
assertEquals(objs[0].id, UUID.fromString("1e5f5d41-6d14-4007-897b-0ed2616bec96"))
|
||||||
assertEquals(objs[0].test!!.id, 1)
|
assertEquals(objs[0].test!!.id, UUID.fromString("1e5f5d41-6d14-4007-897b-0ed2616bec96"))
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun callRequestWithArgs() {
|
fun callRequestWithArgs() {
|
||||||
val result: ObjTest? = connection.selectOne("select json_build_object('id', 1, 'name', ?::text)", listOf("myName"))
|
val result: ObjTest? = connection.selectOne("select json_build_object('id', '2c0243ed-ff4d-4b9f-a52b-e38c71b0ed00', 'name', ?::text)", listOf("myName"))
|
||||||
assertNotNull(result)
|
assertNotNull(result)
|
||||||
assertEquals("myName", result!!.name)
|
assertEquals("myName", result!!.name)
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun callRequestWithArgsEntity() {
|
fun callRequestWithArgsEntity() {
|
||||||
val o = ObjTest("myName")
|
val o = ObjTest("myName", id = UUID.fromString("2c0243ed-ff4d-4b9f-a52b-e38c71b0ed00"))
|
||||||
o.id = 88
|
val obj: ObjTest? = connection.selectOne("select json_build_object('id', id, 'name', name) FROM json_to_record(?::json) as o(id uuid, name text);", listOf(o))
|
||||||
val obj: ObjTest? = connection.selectOne("select json_build_object('id', id, 'name', name) FROM json_to_record(?::json) as o(id int, name text);", listOf(o))
|
|
||||||
assertNotNull(obj)
|
assertNotNull(obj)
|
||||||
assertTrue(obj is ObjTest)
|
assertTrue(obj is ObjTest)
|
||||||
assertEquals(obj!!.id, 88)
|
assertEquals(obj!!.id, UUID.fromString("2c0243ed-ff4d-4b9f-a52b-e38c71b0ed00"))
|
||||||
assertEquals(obj.name, "myName")
|
assertEquals(obj.name, "myName")
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun callExec() {
|
fun callExec() {
|
||||||
val o = ObjTest("myName")
|
val o = ObjTest("myName")
|
||||||
val result = connection.exec("select json_build_object('id', 1, 'name', ?::json->>'name')", listOf(o))
|
val result = connection.exec("select json_build_object('id', '2c0243ed-ff4d-4b9f-a52b-e38c71b0ed00', 'name', ?::json->>'name')", listOf(o))
|
||||||
Assertions.assertEquals(1, result.rowsAffected)
|
Assertions.assertEquals(1, result.rowsAffected)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -141,8 +144,8 @@ class ConnectionTest() : TestAbstract() {
|
|||||||
val result: Paginated<ObjTest> = connection.select(
|
val result: Paginated<ObjTest> = connection.select(
|
||||||
"""
|
"""
|
||||||
SELECT json_build_array(
|
SELECT json_build_array(
|
||||||
json_build_object('id', 3, 'name', :name::text),
|
json_build_object('id', '417aaa7e-7bc6-49b7-9fe8-6c8433b3f430', 'name', :name::text),
|
||||||
json_build_object('id', 4, 'name', :name::text || '-2')
|
json_build_object('id', 'abd46e7a-e749-4ce4-8361-e7b64da89da6', 'name', :name::text || '-2')
|
||||||
), 10 as total
|
), 10 as total
|
||||||
LIMIT :limit OFFSET :offset
|
LIMIT :limit OFFSET :offset
|
||||||
""".trimIndent(),
|
""".trimIndent(),
|
||||||
|
|||||||
@@ -1,15 +1,22 @@
|
|||||||
package fr.postgresjson
|
package fr.postgresjson
|
||||||
|
|
||||||
|
import fr.postgresjson.entity.Entity
|
||||||
|
import fr.postgresjson.entity.EntityCreatedAt
|
||||||
|
import fr.postgresjson.entity.EntityCreatedBy
|
||||||
import fr.postgresjson.entity.EntityI
|
import fr.postgresjson.entity.EntityI
|
||||||
import fr.postgresjson.entity.mutable.*
|
import fr.postgresjson.entity.EntityUpdatedAt
|
||||||
|
import fr.postgresjson.entity.EntityUpdatedBy
|
||||||
|
import fr.postgresjson.entity.Published
|
||||||
|
import fr.postgresjson.entity.UuidEntityExtended
|
||||||
import org.junit.jupiter.api.Assertions.assertTrue
|
import org.junit.jupiter.api.Assertions.assertTrue
|
||||||
import org.junit.jupiter.api.Test
|
import org.junit.jupiter.api.Test
|
||||||
import org.junit.jupiter.api.TestInstance
|
import org.junit.jupiter.api.TestInstance
|
||||||
|
import java.util.UUID
|
||||||
|
|
||||||
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
||||||
class EntityTest() {
|
class EntityTest() {
|
||||||
private class User(id: Int?) : Entity<Int?>(id)
|
private class User(id: UUID = UUID.randomUUID()) : Entity<UUID>(id)
|
||||||
private class ObjTest(var name: String) : UuidEntityExtended<Int?, User>(User(1), User(2))
|
private class ObjTest(var name: String) : UuidEntityExtended<Int?, User>(User(), User())
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun getObject() {
|
fun getObject() {
|
||||||
@@ -17,7 +24,7 @@ class EntityTest() {
|
|||||||
assertTrue(obj is ObjTest)
|
assertTrue(obj is ObjTest)
|
||||||
assertTrue(obj is UuidEntityExtended<Int?, User>)
|
assertTrue(obj is UuidEntityExtended<Int?, User>)
|
||||||
assertTrue(obj is EntityI)
|
assertTrue(obj is EntityI)
|
||||||
assertTrue(obj is Entity<Int?>)
|
assertTrue(obj is Entity<UUID>)
|
||||||
assertTrue(obj is Published<User>)
|
assertTrue(obj is Published<User>)
|
||||||
assertTrue(obj is EntityCreatedBy<User>)
|
assertTrue(obj is EntityCreatedBy<User>)
|
||||||
assertTrue(obj is EntityUpdatedBy<User>)
|
assertTrue(obj is EntityUpdatedBy<User>)
|
||||||
|
|||||||
@@ -10,14 +10,14 @@ import org.amshove.kluent.shouldThrow
|
|||||||
import org.junit.jupiter.api.Assertions
|
import org.junit.jupiter.api.Assertions
|
||||||
import org.junit.jupiter.api.Test
|
import org.junit.jupiter.api.Test
|
||||||
import org.junit.jupiter.api.TestInstance
|
import org.junit.jupiter.api.TestInstance
|
||||||
import java.io.File
|
import java.util.UUID
|
||||||
|
|
||||||
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
||||||
class MigrationTest() : TestAbstract() {
|
class MigrationTest() : TestAbstract() {
|
||||||
@Test
|
@Test
|
||||||
fun `run up query`() {
|
fun `run up query`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/migrations").toURI())
|
val resources = this::class.java.getResource("/sql/migrations").toURI()
|
||||||
val m = Migrations(resources, connection)
|
val m = Migrations(connection, resources)
|
||||||
m.up().apply {
|
m.up().apply {
|
||||||
this `should contain` Pair("1", Migration.Status.OK)
|
this `should contain` Pair("1", Migration.Status.OK)
|
||||||
size `should be equal to` 1
|
size `should be equal to` 1
|
||||||
@@ -28,7 +28,7 @@ class MigrationTest() : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `migration up Query should throw error if no down`() {
|
fun `migration up Query should throw error if no down`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/migration_without_down").toURI())
|
val resources = this::class.java.getResource("/sql/migration_without_down").toURI()
|
||||||
invoking {
|
invoking {
|
||||||
Migrations(resources, connection)
|
Migrations(resources, connection)
|
||||||
} shouldThrow Migrations.DownMigrationNotDefined::class
|
} shouldThrow Migrations.DownMigrationNotDefined::class
|
||||||
@@ -36,7 +36,7 @@ class MigrationTest() : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `run forced down query`() {
|
fun `run forced down query`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/migrations").toURI())
|
val resources = this::class.java.getResource("/sql/migrations").toURI()
|
||||||
val m = Migrations(resources, connection)
|
val m = Migrations(resources, connection)
|
||||||
repeat(3) {
|
repeat(3) {
|
||||||
m.down(true).apply {
|
m.down(true).apply {
|
||||||
@@ -48,7 +48,7 @@ class MigrationTest() : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `run dry migrations`() {
|
fun `run dry migrations`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/real_migrations").toURI())
|
val resources = this::class.java.getResource("/sql/real_migrations").toURI()
|
||||||
Migrations(resources, connection).apply {
|
Migrations(resources, connection).apply {
|
||||||
runDry().size `should be equal to` 2
|
runDry().size `should be equal to` 2
|
||||||
}
|
}
|
||||||
@@ -59,7 +59,7 @@ class MigrationTest() : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `run dry migrations launch twice`() {
|
fun `run dry migrations launch twice`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/real_migrations").toURI())
|
val resources = this::class.java.getResource("/sql/real_migrations").toURI()
|
||||||
Migrations(resources, connection).apply {
|
Migrations(resources, connection).apply {
|
||||||
runDry().size `should be equal to` 2
|
runDry().size `should be equal to` 2
|
||||||
runDry().size `should be equal to` 2
|
runDry().size `should be equal to` 2
|
||||||
@@ -68,7 +68,7 @@ class MigrationTest() : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `run migrations`() {
|
fun `run migrations`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/real_migrations").toURI())
|
val resources = this::class.java.getResource("/sql/real_migrations").toURI()
|
||||||
Migrations(resources, connection).apply {
|
Migrations(resources, connection).apply {
|
||||||
run().apply {
|
run().apply {
|
||||||
size `should be equal to` 1
|
size `should be equal to` 1
|
||||||
@@ -78,8 +78,8 @@ class MigrationTest() : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `run migrations force down`() {
|
fun `run migrations force down`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/real_migrations").toURI())
|
val resources = this::class.java.getResource("/sql/real_migrations").toURI()
|
||||||
val resourcesFunctions = File(this::class.java.getResource("/sql/function/Test").toURI())
|
val resourcesFunctions = this::class.java.getResource("/sql/function/Test").toURI()
|
||||||
Migrations(listOf(resources, resourcesFunctions), connection).apply {
|
Migrations(listOf(resources, resourcesFunctions), connection).apply {
|
||||||
up().apply {
|
up().apply {
|
||||||
size `should be equal to` 6
|
size `should be equal to` 6
|
||||||
@@ -94,7 +94,7 @@ class MigrationTest() : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `run functions migrations`() {
|
fun `run functions migrations`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/function/Test").toURI())
|
val resources = this::class.java.getResource("/sql/function/Test").toURI()
|
||||||
Migrations(resources, connection).apply {
|
Migrations(resources, connection).apply {
|
||||||
run().size `should be equal to` 5
|
run().size `should be equal to` 5
|
||||||
}
|
}
|
||||||
@@ -104,13 +104,13 @@ class MigrationTest() : TestAbstract() {
|
|||||||
.getFunction("test_function")
|
.getFunction("test_function")
|
||||||
.selectOne(listOf("test", "plip"))
|
.selectOne(listOf("test", "plip"))
|
||||||
|
|
||||||
Assertions.assertEquals(objTest!!.id, 3)
|
Assertions.assertEquals(objTest!!.id, UUID.fromString("457daad5-4f1b-4eb7-80ec-6882adb8cc7d"))
|
||||||
Assertions.assertEquals(objTest.name, "test")
|
Assertions.assertEquals(objTest.name, "test")
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `run functions migrations and drop if exist`() {
|
fun `run functions migrations and drop if exist`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/function/Test1").toURI())
|
val resources = this::class.java.getResource("/sql/function/Test1").toURI()
|
||||||
Migrations(resources, connection).apply {
|
Migrations(resources, connection).apply {
|
||||||
run().size `should be equal to` 1
|
run().size `should be equal to` 1
|
||||||
}
|
}
|
||||||
@@ -120,10 +120,10 @@ class MigrationTest() : TestAbstract() {
|
|||||||
.getFunction("test_function_duplicate")
|
.getFunction("test_function_duplicate")
|
||||||
.selectOne(listOf("test"))
|
.selectOne(listOf("test"))
|
||||||
|
|
||||||
Assertions.assertEquals(objTest!!.id, 3)
|
Assertions.assertEquals(objTest!!.id, UUID.fromString("457daad5-4f1b-4eb7-80ec-6882adb8cc7d"))
|
||||||
Assertions.assertEquals(objTest.name, "test")
|
Assertions.assertEquals(objTest.name, "test")
|
||||||
|
|
||||||
val resources2 = File(this::class.java.getResource("/sql/function/Test2").toURI())
|
val resources2 = this::class.java.getResource("/sql/function/Test2").toURI()
|
||||||
Migrations(resources2, connection).apply {
|
Migrations(resources2, connection).apply {
|
||||||
run().size `should be equal to` 1
|
run().size `should be equal to` 1
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,45 +2,45 @@ package fr.postgresjson
|
|||||||
|
|
||||||
import fr.postgresjson.connexion.Paginated
|
import fr.postgresjson.connexion.Paginated
|
||||||
import fr.postgresjson.connexion.Requester
|
import fr.postgresjson.connexion.Requester
|
||||||
import fr.postgresjson.entity.mutable.IdEntity
|
import fr.postgresjson.entity.UuidEntity
|
||||||
import org.junit.Assert
|
import org.junit.Assert
|
||||||
import org.junit.jupiter.api.Assertions.assertEquals
|
import org.junit.jupiter.api.Assertions.assertEquals
|
||||||
import org.junit.jupiter.api.Test
|
import org.junit.jupiter.api.Test
|
||||||
import java.io.File
|
import java.util.UUID
|
||||||
|
|
||||||
class RequesterTest : TestAbstract() {
|
class RequesterTest : TestAbstract() {
|
||||||
class ObjTest(var name: String) : IdEntity(1)
|
class ObjTest(var name: String, id: UUID = UUID.fromString("5623d902-3067-42f3-bfd9-095dbb12c29f")) : UuidEntity(id)
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `get query from file`() {
|
fun `get query from file`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/query").toURI())
|
val resources = this::class.java.getResource("/sql/query").toURI()
|
||||||
val objTest: ObjTest? = Requester(connection)
|
val objTest: ObjTest? = Requester(connection)
|
||||||
.addQuery(resources)
|
.addQuery(resources)
|
||||||
.getQuery("Test/selectOne")
|
.getQuery("selectOne")
|
||||||
.selectOne()
|
.selectOne()
|
||||||
|
|
||||||
assertEquals(objTest!!.id, 2)
|
assertEquals(objTest!!.id, UUID.fromString("829b1a29-5db8-47f9-9562-961c561ac528"))
|
||||||
assertEquals(objTest.name, "test")
|
assertEquals(objTest.name, "test")
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `get function from file`() {
|
fun `get function from file`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/function/Test").toURI())
|
val resources = this::class.java.getResource("/sql/function/Test").toURI()
|
||||||
val objTest: ObjTest? = Requester(connection)
|
val objTest: ObjTest? = Requester(connection)
|
||||||
.addFunction(resources)
|
.addFunction(resources)
|
||||||
.getFunction("test_function")
|
.getFunction("test_function")
|
||||||
.selectOne(listOf("test", "plip"))
|
.selectOne(listOf("test", "plip"))
|
||||||
|
|
||||||
assertEquals(objTest!!.id, 3)
|
assertEquals(objTest!!.id, UUID.fromString("457daad5-4f1b-4eb7-80ec-6882adb8cc7d"))
|
||||||
assertEquals(objTest.name, "test")
|
assertEquals(objTest.name, "test")
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call exec on query`() {
|
fun `call exec on query`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/query").toURI())
|
val resources = this::class.java.getResource("/sql/query").toURI()
|
||||||
val result = Requester(connection)
|
val result = Requester(connection)
|
||||||
.addQuery(resources)
|
.addQuery(resources)
|
||||||
.getQuery("Test/selectOne")
|
.getQuery("selectOne")
|
||||||
.exec()
|
.exec()
|
||||||
|
|
||||||
assertEquals(1, result.rowsAffected)
|
assertEquals(1, result.rowsAffected)
|
||||||
@@ -48,7 +48,7 @@ class RequesterTest : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call exec on function`() {
|
fun `call exec on function`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/function/Test").toURI())
|
val resources = this::class.java.getResource("/sql/function/Test").toURI()
|
||||||
val result = Requester(connection)
|
val result = Requester(connection)
|
||||||
.addFunction(resources)
|
.addFunction(resources)
|
||||||
.getFunction("test_function")
|
.getFunction("test_function")
|
||||||
@@ -58,11 +58,11 @@ class RequesterTest : TestAbstract() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call sendQuery on query`() {
|
fun `call sendQuery on query with name`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/query").toURI())
|
val resources = this::class.java.getResource("/sql/query").toURI()
|
||||||
val result = Requester(connection)
|
val result = Requester(connection)
|
||||||
.addQuery(resources)
|
.addQuery(resources)
|
||||||
.getQuery("Test/exec")
|
.getQuery("DeleteTest")
|
||||||
.sendQuery()
|
.sendQuery()
|
||||||
|
|
||||||
assertEquals(0, result)
|
assertEquals(0, result)
|
||||||
@@ -70,7 +70,7 @@ class RequesterTest : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call sendQuery on function`() {
|
fun `call sendQuery on function`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/function/Test").toURI())
|
val resources = this::class.java.getResource("/sql/function/Test").toURI()
|
||||||
val result = Requester(connection)
|
val result = Requester(connection)
|
||||||
.addFunction(resources)
|
.addFunction(resources)
|
||||||
.getFunction("function_void")
|
.getFunction("function_void")
|
||||||
@@ -81,7 +81,7 @@ class RequesterTest : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call selectOne on function`() {
|
fun `call selectOne on function`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/function/Test").toURI())
|
val resources = this::class.java.getResource("/sql/function/Test").toURI()
|
||||||
val obj: ObjTest = Requester(connection)
|
val obj: ObjTest = Requester(connection)
|
||||||
.addFunction(resources)
|
.addFunction(resources)
|
||||||
.getFunction("test_function")
|
.getFunction("test_function")
|
||||||
@@ -92,7 +92,7 @@ class RequesterTest : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call selectOne on function with object`() {
|
fun `call selectOne on function with object`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/function/Test").toURI())
|
val resources = this::class.java.getResource("/sql/function/Test").toURI()
|
||||||
val obj2 = ObjTest("original")
|
val obj2 = ObjTest("original")
|
||||||
val obj: ObjTest = Requester(connection)
|
val obj: ObjTest = Requester(connection)
|
||||||
.addFunction(resources)
|
.addFunction(resources)
|
||||||
@@ -105,10 +105,10 @@ class RequesterTest : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call selectOne on query`() {
|
fun `call selectOne on query`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/query").toURI())
|
val resources = this::class.java.getResource("/sql/query").toURI()
|
||||||
val obj: ObjTest = Requester(connection)
|
val obj: ObjTest = Requester(connection)
|
||||||
.addQuery(resources)
|
.addQuery(resources)
|
||||||
.getQuery("Test/selectOneWithParameters")
|
.getQuery("selectOneWithParameters")
|
||||||
.selectOne(mapOf("name" to "myName"))!!
|
.selectOne(mapOf("name" to "myName"))!!
|
||||||
|
|
||||||
assertEquals("myName", obj.name)
|
assertEquals("myName", obj.name)
|
||||||
@@ -116,7 +116,7 @@ class RequesterTest : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call select (multiple) on function`() {
|
fun `call select (multiple) on function`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/function/Test").toURI())
|
val resources = this::class.java.getResource("/sql/function/Test").toURI()
|
||||||
val obj: List<ObjTest>? = Requester(connection)
|
val obj: List<ObjTest>? = Requester(connection)
|
||||||
.addFunction(resources)
|
.addFunction(resources)
|
||||||
.getFunction("test_function_multiple")
|
.getFunction("test_function_multiple")
|
||||||
@@ -127,10 +127,10 @@ class RequesterTest : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call select paginated on query`() {
|
fun `call select paginated on query`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/query").toURI())
|
val resources = this::class.java.getResource("/sql/query").toURI()
|
||||||
val result: Paginated<ObjTest> = Requester(connection)
|
val result: Paginated<ObjTest> = Requester(connection)
|
||||||
.addQuery(resources)
|
.addQuery(resources)
|
||||||
.getQuery("Test/selectPaginated")
|
.getQuery("selectPaginated")
|
||||||
.select(1, 2, mapOf("name" to "ff"))
|
.select(1, 2, mapOf("name" to "ff"))
|
||||||
Assert.assertNotNull(result)
|
Assert.assertNotNull(result)
|
||||||
Assert.assertEquals("ff", result.result[0].name)
|
Assert.assertEquals("ff", result.result[0].name)
|
||||||
@@ -141,7 +141,7 @@ class RequesterTest : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call select paginated on function`() {
|
fun `call select paginated on function`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/function").toURI())
|
val resources = this::class.java.getResource("/sql/function").toURI()
|
||||||
val result: Paginated<ObjTest> = Requester(connection)
|
val result: Paginated<ObjTest> = Requester(connection)
|
||||||
.addFunction(resources)
|
.addFunction(resources)
|
||||||
.getFunction("test_function_paginated")
|
.getFunction("test_function_paginated")
|
||||||
@@ -155,10 +155,10 @@ class RequesterTest : TestAbstract() {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun `call selectOne on query with extra parameter`() {
|
fun `call selectOne on query with extra parameter`() {
|
||||||
val resources = File(this::class.java.getResource("/sql/query").toURI())
|
val resources = this::class.java.getResource("/sql/query").toURI()
|
||||||
val obj: ObjTest = Requester(connection)
|
val obj: ObjTest = Requester(connection)
|
||||||
.addQuery(resources)
|
.addQuery(resources)
|
||||||
.getQuery("Test/selectOneWithParameters")
|
.getQuery("selectOneWithParameters")
|
||||||
.selectOne(mapOf("name" to "myName")) {
|
.selectOne(mapOf("name" to "myName")) {
|
||||||
assertEquals("myName", it!!.name)
|
assertEquals("myName", it!!.name)
|
||||||
Assert.assertEquals("plop", rows[0].getString("other"))
|
Assert.assertEquals("plop", rows[0].getString("other"))
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package fr.postgresjson
|
package fr.postgresjson
|
||||||
|
|
||||||
import fr.postgresjson.entity.mutable.IdEntity
|
import fr.postgresjson.entity.UuidEntity
|
||||||
import fr.postgresjson.serializer.Serializer
|
import fr.postgresjson.serializer.Serializer
|
||||||
import fr.postgresjson.serializer.deserialize
|
import fr.postgresjson.serializer.deserialize
|
||||||
import fr.postgresjson.serializer.serialize
|
import fr.postgresjson.serializer.serialize
|
||||||
@@ -10,23 +10,23 @@ import org.junit.jupiter.api.Assertions.assertTrue
|
|||||||
import org.junit.jupiter.api.BeforeEach
|
import org.junit.jupiter.api.BeforeEach
|
||||||
import org.junit.jupiter.api.Test
|
import org.junit.jupiter.api.Test
|
||||||
import org.junit.jupiter.api.TestInstance
|
import org.junit.jupiter.api.TestInstance
|
||||||
|
import java.util.UUID
|
||||||
|
|
||||||
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
||||||
internal class SerializerTest {
|
internal class SerializerTest {
|
||||||
private class ObjTest(var val1: String, var val2: Int) : IdEntity(1)
|
private class ObjTest(var val1: String, var val2: Int, id: UUID = UUID.fromString("1e5f5d41-6d14-4007-897b-0ed2616bec96")) : UuidEntity(id)
|
||||||
private class ObjTestDate(var val1: DateTime) : IdEntity(2)
|
private class ObjTestDate(var val1: DateTime, id: UUID = UUID.fromString("829b1a29-5db8-47f9-9562-961c561ac528")) : UuidEntity(id)
|
||||||
|
|
||||||
private val serializer = Serializer()
|
private val serializer = Serializer()
|
||||||
|
|
||||||
private val objSerialized: String = """{"val1":"plop","val2":123,"id":2}"""
|
private val objSerialized: String = """{"val1":"plop","val2":123,"id":"829b1a29-5db8-47f9-9562-961c561ac528"}"""
|
||||||
private val objSerializedWithExtra: String = """{"val1":"plop","val2":123,"id":2,"toto":"tata"}"""
|
private val objSerializedWithExtra: String = """{"val1":"plop","val2":123,"id":"829b1a29-5db8-47f9-9562-961c561ac528","toto":"tata"}"""
|
||||||
private val objSerializedUpdate = """{"val1":"update","val2":123}"""
|
private val objSerializedUpdate = """{"val1":"update","val2":123}"""
|
||||||
private lateinit var obj: ObjTest
|
private lateinit var obj: ObjTest
|
||||||
|
|
||||||
@BeforeEach
|
@BeforeEach
|
||||||
fun before() {
|
fun before() {
|
||||||
obj = ObjTest("plop", 123)
|
obj = ObjTest("plop", 123, UUID.fromString("829b1a29-5db8-47f9-9562-961c561ac528"))
|
||||||
obj.id = 2
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ import java.io.File
|
|||||||
|
|
||||||
@TestInstance(PER_CLASS)
|
@TestInstance(PER_CLASS)
|
||||||
abstract class TestAbstract {
|
abstract class TestAbstract {
|
||||||
protected val connection = Connection(database = "test_json", username = "test", password = "test")
|
protected val connection = Connection(database = "json_test", username = "test", password = "test", port = 5555)
|
||||||
|
|
||||||
@BeforeEach
|
@BeforeEach
|
||||||
fun beforeAll() {
|
fun beforeAll() {
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ create schema if not exists public;
|
|||||||
|
|
||||||
create table if not exists test
|
create table if not exists test
|
||||||
(
|
(
|
||||||
id serial not null
|
id uuid not null
|
||||||
constraint test_pk
|
constraint test_pk
|
||||||
primary key,
|
primary key,
|
||||||
name text
|
name text
|
||||||
@@ -16,24 +16,24 @@ create table if not exists test
|
|||||||
|
|
||||||
create table if not exists test2
|
create table if not exists test2
|
||||||
(
|
(
|
||||||
id serial not null,
|
id uuid not null,
|
||||||
title text,
|
title text,
|
||||||
test_id integer
|
test_id uuid
|
||||||
constraint test2_test_id_fk
|
constraint test2_test_id_fk
|
||||||
references test
|
references test
|
||||||
);
|
);
|
||||||
|
|
||||||
INSERT INTO test (id, name) VALUES (1, 'plop') ON CONFLICT DO NOTHING;
|
INSERT INTO test (id, name) VALUES ('1e5f5d41-6d14-4007-897b-0ed2616bec96', 'plop') ON CONFLICT DO NOTHING;
|
||||||
INSERT INTO test2 (id, title, test_id) VALUES (1, 'plop', 1) ON CONFLICT DO NOTHING;
|
INSERT INTO test2 (id, title, test_id) VALUES ('1e5f5d41-6d14-4007-897b-0ed2616bec96', 'plop', '1e5f5d41-6d14-4007-897b-0ed2616bec96') ON CONFLICT DO NOTHING;
|
||||||
INSERT INTO test2 (id, title, test_id) VALUES (2, 'plip', 1) ON CONFLICT DO NOTHING;
|
INSERT INTO test2 (id, title, test_id) VALUES ('829b1a29-5db8-47f9-9562-961c561ac528', 'plip', '1e5f5d41-6d14-4007-897b-0ed2616bec96') ON CONFLICT DO NOTHING;
|
||||||
INSERT INTO test2 (id, title, test_id) VALUES (3, 'ttt', null) ON CONFLICT DO NOTHING;
|
INSERT INTO test2 (id, title, test_id) VALUES ('457daad5-4f1b-4eb7-80ec-6882adb8cc7d', 'ttt', null) ON CONFLICT DO NOTHING;
|
||||||
|
|
||||||
CREATE OR REPLACE FUNCTION test_function (name text default 'plop', IN hi text default 'hello', out result json)
|
CREATE OR REPLACE FUNCTION test_function (name text default 'plop', IN hi text default 'hello', out result json)
|
||||||
LANGUAGE plpgsql
|
LANGUAGE plpgsql
|
||||||
AS
|
AS
|
||||||
$$
|
$$
|
||||||
BEGIN
|
BEGIN
|
||||||
result = json_build_object('id', 3, 'name', name);
|
result = json_build_object('id', '457daad5-4f1b-4eb7-80ec-6882adb8cc7d', 'name', name);
|
||||||
END;
|
END;
|
||||||
$$;
|
$$;
|
||||||
|
|
||||||
@@ -44,8 +44,8 @@ AS
|
|||||||
$$
|
$$
|
||||||
BEGIN
|
BEGIN
|
||||||
result = json_build_array(
|
result = json_build_array(
|
||||||
json_build_object('id', 3, 'name', name),
|
json_build_object('id', '457daad5-4f1b-4eb7-80ec-6882adb8cc7d', 'name', name),
|
||||||
json_build_object('id', 4, 'name', hi)
|
json_build_object('id', '8d20abb0-7f77-4b6c-9991-44acd3c88faa', 'name', hi)
|
||||||
);
|
);
|
||||||
END;
|
END;
|
||||||
$$;
|
$$;
|
||||||
@@ -56,8 +56,8 @@ AS
|
|||||||
$$
|
$$
|
||||||
BEGIN
|
BEGIN
|
||||||
SELECT json_build_array(
|
SELECT json_build_array(
|
||||||
json_build_object('id', 3, 'name', name::text),
|
json_build_object('id', '457daad5-4f1b-4eb7-80ec-6882adb8cc7d', 'name', name::text),
|
||||||
json_build_object('id', 4, 'name', name::text || '-2')
|
json_build_object('id', '8d20abb0-7f77-4b6c-9991-44acd3c88faa', 'name', name::text || '-2')
|
||||||
),
|
),
|
||||||
10
|
10
|
||||||
INTO result, total
|
INTO result, total
|
||||||
@@ -70,7 +70,7 @@ CREATE OR REPLACE FUNCTION test_function_object (inout resource json)
|
|||||||
AS
|
AS
|
||||||
$$
|
$$
|
||||||
BEGIN
|
BEGIN
|
||||||
resource = json_build_object('id', 1, 'name', 'changedName');
|
resource = json_build_object('id', '1e5f5d41-6d14-4007-897b-0ed2616bec96', 'name', 'changedName');
|
||||||
END;
|
END;
|
||||||
$$;
|
$$;
|
||||||
|
|
||||||
|
|||||||
@@ -3,6 +3,6 @@ LANGUAGE plpgsql
|
|||||||
AS
|
AS
|
||||||
$$
|
$$
|
||||||
BEGIN
|
BEGIN
|
||||||
result = json_build_object('id', 3, 'name', name);
|
result = json_build_object('id', '457daad5-4f1b-4eb7-80ec-6882adb8cc7d', 'name', name);
|
||||||
END;
|
END;
|
||||||
$$
|
$$
|
||||||
@@ -3,6 +3,6 @@ LANGUAGE plpgsql
|
|||||||
AS
|
AS
|
||||||
$$
|
$$
|
||||||
BEGIN
|
BEGIN
|
||||||
resource = json_build_object('id', 1, 'name', 'changedName');
|
resource = json_build_object('id', '1e5f5d41-6d14-4007-897b-0ed2616bec96', 'name', 'changedName');
|
||||||
END;
|
END;
|
||||||
$$
|
$$
|
||||||
@@ -4,8 +4,8 @@ AS
|
|||||||
$$
|
$$
|
||||||
BEGIN
|
BEGIN
|
||||||
result = json_build_array(
|
result = json_build_array(
|
||||||
json_build_object('id', 3, 'name', name),
|
json_build_object('id', '457daad5-4f1b-4eb7-80ec-6882adb8cc7d', 'name', name),
|
||||||
json_build_object('id', 4, 'name', hi)
|
json_build_object('id', '8d20abb0-7f77-4b6c-9991-44acd3c88faa', 'name', hi)
|
||||||
);
|
);
|
||||||
END;
|
END;
|
||||||
$$
|
$$
|
||||||
@@ -4,8 +4,8 @@ AS
|
|||||||
$$
|
$$
|
||||||
BEGIN
|
BEGIN
|
||||||
SELECT json_build_array(
|
SELECT json_build_array(
|
||||||
json_build_object('id', 3, 'name', name::text),
|
json_build_object('id', '457daad5-4f1b-4eb7-80ec-6882adb8cc7d', 'name', name::text),
|
||||||
json_build_object('id', 4, 'name', name::text || '-2')
|
json_build_object('id', '8d20abb0-7f77-4b6c-9991-44acd3c88faa', 'name', name::text || '-2')
|
||||||
),
|
),
|
||||||
10
|
10
|
||||||
INTO result, total
|
INTO result, total
|
||||||
|
|||||||
@@ -3,6 +3,6 @@ CREATE OR REPLACE FUNCTION test_function_duplicate (name text default 'plop') re
|
|||||||
AS
|
AS
|
||||||
$$
|
$$
|
||||||
BEGIN
|
BEGIN
|
||||||
return json_build_object('id', 3, 'name', name);
|
return json_build_object('id', '457daad5-4f1b-4eb7-80ec-6882adb8cc7d', 'name', name);
|
||||||
END;
|
END;
|
||||||
$$
|
$$
|
||||||
|
|||||||
@@ -1 +1,2 @@
|
|||||||
|
-- name: DeleteTest
|
||||||
delete FROM test where 2038538 = 2;
|
delete FROM test where 2038538 = 2;
|
||||||
@@ -1 +1 @@
|
|||||||
select json_build_object('id', 2, 'name', 'test');
|
select json_build_object('id', '829b1a29-5db8-47f9-9562-961c561ac528', 'name', 'test');
|
||||||
@@ -1 +1 @@
|
|||||||
select json_build_object('id', 2, 'name', :name::text), 'plop'::text as other;
|
select json_build_object('id', '829b1a29-5db8-47f9-9562-961c561ac528', 'name', :name::text), 'plop'::text as other;
|
||||||
@@ -1,5 +1,5 @@
|
|||||||
SELECT json_build_array(
|
SELECT json_build_array(
|
||||||
json_build_object('id', 3, 'name', :name::text),
|
json_build_object('id', '457daad5-4f1b-4eb7-80ec-6882adb8cc7d', 'name', :name::text),
|
||||||
json_build_object('id', 4, 'name', :name::text || '-2')
|
json_build_object('id', '6085c12e-e94d-4ae1-b7ad-23acc7a82a98', 'name', :name::text || '-2')
|
||||||
), 10 as total
|
), 10 as total
|
||||||
LIMIT :limit OFFSET :offset
|
LIMIT :limit OFFSET :offset
|
||||||
Reference in New Issue
Block a user