Skip to content

Commit 96fd60a

Browse files
authored
Merge pull request #1163 from cardstack/cs-6460-setup-node-pg-migrate-for-database-migrations
Add support for auto generating sqlite schema
2 parents db6f490 + 4d4c3df commit 96fd60a

11 files changed

+330
-32
lines changed

.gitignore

+1
Original file line numberDiff line numberDiff line change
@@ -6,3 +6,4 @@ node_modules
66
.dist
77
/junit
88
.eslintcache
9+
schema_tmp.sql

README.md

+39
Original file line numberDiff line numberDiff line change
@@ -109,6 +109,45 @@ The realm server uses the request accept header to determine the type of request
109109
| `text/html` | Card instance URL's should not include the `.json` file extension. This is considered a 404 | Used to request rendered card instance html (this serves the host application) |
110110
| `*/*` | We support node-like resolution, which means that the extension is optional | Used to request transpiled executable code modules |
111111

112+
### Database
113+
114+
Boxel uses a Postgres database. In development, the Postgres database runs within a docker container, `boxel-pg`, that is started as part of `pnpm start:all`. You can manually start and stop the `boxel-pg` docker container using `pnpm start:pg` and `pnpm stop:pg`. The postgres database runs on port 5435 so that it doesn't conflict with a natively installed postgres that may be running on your system.
115+
116+
When running tests we isolate the database between each test run by actually creating a new database for each test with a random database name (e.g. `test_db_1234567`). The test databases are dropped before the beginning of each test run.
117+
118+
If you wish to drop the development database you can execute:
119+
```
120+
pnpm drop-db
121+
```
122+
123+
You can then run `pnpm migrate up` or start the realm server to create the database again.
124+
125+
#### DB Migrations
126+
When the realm server starts up it will automatically run DB migrations that live in the `packages/realm-server/migrations` folder. As part of development you may wish to run migrations manually as well as to create a new migration.
127+
128+
To create a new migration, from `packages/realm-server`, execute:
129+
```
130+
pnpm migrate create name-of-migration
131+
```
132+
This creates a new migration file in `packages/realm-server/migrations`. You can then edit the newly created migration file with the details of your migration. We use `node-pg-migrate` to handle our migrations. You can find the API at https://salsita.github.io/node-pg-migrate.
133+
134+
To run the migration, execute:
135+
```
136+
pnpm migrate up
137+
```
138+
139+
To revert the migration, execute:
140+
```
141+
pnpm migrate down
142+
```
143+
144+
Boxel also uses SQLite in order to run the DB in the browser as part of running browser tests (and eventually we may run the realm server in the browser to provide a local index). We treat the Postgres database schema as the source of truth and derive the SQLite schema from it. Therefore, once you author and apply a migration, you should generate a new schema SQL file for SQLite. To generate a new SQLite schema, from `packages/realm-server`, execute:
145+
```
146+
pnpm make-schema
147+
```
148+
This will create a new SQLite schema based on the current postgres DB (the schema file will be placed in the `packages/host/config/schema` directory). This schema file will share the same timestamp as the latest migration file's timestamp. If you forget to generate a new schema file, the next time you start the host app, you will receive an error that the SQLite schema is out of date.
149+
150+
112151
### Matrix Server
113152

114153
The boxel platform leverages a Matrix server called Synapse in order to support identity, workflow, and chat behaviors. This project uses a dockerized Matrix server. We have multiple matrix server configurations (currently one for development that uses a persistent DB, and one for testing that uses an in-memory DB). You can find and configure these matrix servers at `packages/matrix/docker/synapse/*`.

packages/host/config/environment.js

+24-1
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
const fs = require('fs');
44
const path = require('path');
55

6-
let sqlSchema = fs.readFileSync(path.join(__dirname, 'schema.sql'), 'utf8');
6+
let sqlSchema = fs.readFileSync(getLatestSchemaFile(), 'utf8');
77

88
module.exports = function (environment) {
99
const ENV = {
@@ -94,3 +94,26 @@ module.exports = function (environment) {
9494

9595
return ENV;
9696
};
97+
98+
function getLatestSchemaFile() {
99+
const migrationsDir = path.resolve(
100+
path.join(__dirname, '..', '..', 'realm-server', 'migrations'),
101+
);
102+
let migrations = fs.readdirSync(migrationsDir);
103+
let lastMigration = migrations
104+
.filter((f) => f !== '.eslintrc.js')
105+
.sort()
106+
.pop();
107+
const schemaDir = path.join(__dirname, 'schema');
108+
let files = fs.readdirSync(schemaDir);
109+
let latestSchemaFile = files.sort().pop();
110+
if (
111+
lastMigration.replace(/_.*/, '') !== latestSchemaFile.replace(/_.*/, '') &&
112+
['development', 'test'].includes(process.env.EMBER_ENV)
113+
) {
114+
throw new Error(
115+
`The sqlite schema file is out of date--please regenerate the sqlite schema file`,
116+
);
117+
}
118+
return path.join(schemaDir, latestSchemaFile);
119+
}

packages/host/config/schema.sql

-29
This file was deleted.
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,24 @@
1+
-- This is auto-generated by packages/realm-server/scripts/convert-to-sqlite.ts
2+
-- Please don't directly modify this file
3+
4+
CREATE TABLE IF NOT EXISTS indexed_cards (
5+
card_url TEXT NOT NULL,
6+
realm_version INTEGER NOT NULL,
7+
realm_url TEXT NOT NULL,
8+
pristine_doc JSON,
9+
search_doc JSON,
10+
error_doc JSON,
11+
deps JSON,
12+
types JSON,
13+
embedded_html TEXT,
14+
isolated_html TEXT,
15+
indexed_at INTEGER,
16+
is_deleted BOOLEAN,
17+
PRIMARY KEY ( card_url, realm_version )
18+
);
19+
20+
CREATE TABLE IF NOT EXISTS realm_versions (
21+
realm_url TEXT NOT NULL,
22+
current_version INTEGER NOT NULL,
23+
PRIMARY KEY ( realm_url )
24+
);

packages/realm-server/migrations/1712771547705_initial.js

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
exports.up = (pgm) => {
22
pgm.createTable('indexed_cards', {
33
card_url: { type: 'varchar', notNull: true },
4-
realm_version: { type: 'varchar', notNull: true },
4+
realm_version: { type: 'integer', notNull: true },
55
realm_url: { type: 'varchar', notNull: true },
66
pristine_doc: 'jsonb',
77
search_doc: 'jsonb',

packages/realm-server/package.json

+4-1
Original file line numberDiff line numberDiff line change
@@ -56,6 +56,7 @@
5656
"qs": "^6.10.5",
5757
"qunit": "^2.20.0",
5858
"sane": "^5.0.1",
59+
"sql-parser-cst": "^0.28.0",
5960
"start-server-and-test": "^1.14.0",
6061
"supertest": "^6.2.4",
6162
"testem": "^3.10.1",
@@ -95,7 +96,9 @@
9596
"lint:js": "eslint . --cache",
9697
"lint:js:fix": "eslint . --fix",
9798
"lint:glint": "glint",
98-
"migrate": "node-pg-migrate"
99+
"migrate": "PGDATABASE=boxel ./scripts/ensure-db-exists.sh && PGPORT=5435 PGDATABASE=boxel PGUSER=postgres node-pg-migrate",
100+
"make-schema": "./scripts/schema-dump.sh",
101+
"drop-db": "docker exec boxel-pg dropdb -U postgres -w boxel"
99102
},
100103
"volta": {
101104
"extends": "../../package.json"
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,205 @@
1+
/* eslint-env node */
2+
import { readFileSync, readdirSync, writeFileSync } from 'fs-extra';
3+
import { resolve, join } from 'path';
4+
import {
5+
parse,
6+
type CreateTableStmt,
7+
type AlterTableStmt,
8+
type Program,
9+
} from 'sql-parser-cst';
10+
11+
// Currently this script only cares about CREATE TABLE statements and ALTER
12+
// TABLE statements that add primary key constraints. All the other schema aspects of the
13+
// pg_dump are generally beyond the capability of SQLite. Perhaps index creation
14+
// can be added but it will get really tricky fast since SQLite's indices are
15+
// much more simplistic than postgres.
16+
17+
const args = process.argv;
18+
const migrationsDir = resolve(join(__dirname, '..', 'migrations'));
19+
const sqliteSchemaDir = resolve(
20+
join(__dirname, '..', '..', 'host', 'config', 'schema'),
21+
);
22+
const INDENT = ' ';
23+
24+
let pgDumpFile = args[2];
25+
if (!pgDumpFile) {
26+
console.error(`please specify the path of the pg_dump file`);
27+
process.exit(-1);
28+
}
29+
let pgDump = readFileSync(pgDumpFile, 'utf8');
30+
31+
let cst = parse(prepareDump(pgDump), {
32+
dialect: 'postgresql',
33+
});
34+
35+
let sql: string[] = [
36+
`
37+
-- This is auto-generated by packages/realm-server/scripts/convert-to-sqlite.ts
38+
-- Please don't directly modify this file
39+
40+
`,
41+
];
42+
for (let statement of cst.statements) {
43+
if (statement.type !== 'create_table_stmt') {
44+
continue;
45+
}
46+
sql.push('CREATE TABLE IF NOT EXISTS');
47+
if (
48+
statement.name.type === 'member_expr' &&
49+
statement.name.property.type === 'identifier'
50+
) {
51+
let tableName = statement.name.property.name;
52+
sql.push(statement.name.property.name, '(\n');
53+
createColumns(cst, tableName, statement, sql);
54+
} else {
55+
throw new Error(`could not determine table name to be created`);
56+
}
57+
58+
sql.push('\n);\n\n');
59+
}
60+
61+
let result = sql.join(' ').trim();
62+
let filename = getSchemaFilename();
63+
let schemaFile = join(sqliteSchemaDir, filename);
64+
writeFileSync(schemaFile, result);
65+
console.log(`created SQLite schema file ${schemaFile}`);
66+
67+
function createColumns(
68+
cst: Program,
69+
tableName: string,
70+
statement: CreateTableStmt,
71+
sql: string[],
72+
) {
73+
if (!statement.columns) {
74+
return;
75+
}
76+
let columns: string[] = [];
77+
for (let [index, item] of statement.columns.expr.items.entries()) {
78+
if (item.type !== 'column_definition') {
79+
continue;
80+
}
81+
let column: string[] = [];
82+
column.push(index === 0 ? INDENT.substring(1) : INDENT, item.name.name);
83+
if (item.dataType?.type === 'named_data_type') {
84+
let dataTypeName = Array.isArray(item.dataType.nameKw)
85+
? item.dataType.nameKw[0]
86+
: item.dataType.nameKw;
87+
switch (dataTypeName.name) {
88+
case 'CHARACTER':
89+
column.push('TEXT');
90+
break;
91+
case 'JSONB':
92+
// TODO change this to 'BLOB' after we do the sqlite BLOB storage
93+
// support in CS-6668 for faster performance
94+
column.push('JSON');
95+
break;
96+
case 'BOOLEAN':
97+
column.push('BOOLEAN');
98+
break;
99+
case 'INTEGER':
100+
column.push('INTEGER');
101+
break;
102+
}
103+
}
104+
for (let constraint of item.constraints) {
105+
switch (constraint.type) {
106+
case 'constraint_not_null':
107+
column.push('NOT NULL');
108+
break;
109+
case 'constraint_primary_key':
110+
column.push('PRIMARY KEY');
111+
break;
112+
default:
113+
throw new Error(
114+
`Don't know how to serialize constraint ${constraint.type} for column '${item.name.name}'`,
115+
);
116+
}
117+
}
118+
119+
columns.push(column.join(' '));
120+
}
121+
let pkConstraint = makePrimaryKeyConstraint(cst, tableName);
122+
sql.push([...columns, ...(pkConstraint ? [pkConstraint] : [])].join(',\n'));
123+
}
124+
125+
function makePrimaryKeyConstraint(
126+
cst: Program,
127+
tableName: string,
128+
): string | undefined {
129+
let alterTableStmts = cst.statements.filter(
130+
(s) =>
131+
s.type === 'alter_table_stmt' &&
132+
s.table.type === 'table_without_inheritance' &&
133+
s.table.table.type === 'member_expr' &&
134+
s.table.table.property.type === 'identifier' &&
135+
s.table.table.property.name === tableName,
136+
) as AlterTableStmt[];
137+
let pkConstraint: string[] = [];
138+
for (let alterTableStmt of alterTableStmts) {
139+
for (let item of alterTableStmt.actions.items) {
140+
if (item.type === 'alter_action_add_constraint') {
141+
switch (item.constraint.type) {
142+
case 'constraint_primary_key': {
143+
if (pkConstraint.length > 0) {
144+
throw new Error(
145+
`encountered multiple primary key constraints for table ${tableName}`,
146+
);
147+
}
148+
if (item.constraint.columns) {
149+
let columns: string[] = [];
150+
if (item.constraint.columns.type === 'paren_expr') {
151+
for (let column of item.constraint.columns.expr.items) {
152+
if (
153+
column.type === 'index_specification' &&
154+
column.expr.type === 'identifier'
155+
) {
156+
columns.push(column.expr.name);
157+
}
158+
}
159+
} else {
160+
throw new Error(
161+
`Don't know how to serialize constraint ${item.constraint.type} for table '${tableName}'`,
162+
);
163+
}
164+
if (columns.length > 0) {
165+
pkConstraint.push(
166+
INDENT,
167+
'PRIMARY KEY (',
168+
columns.join(', '),
169+
')',
170+
);
171+
}
172+
}
173+
break;
174+
}
175+
default:
176+
throw new Error(
177+
`Don't know how to serialize constraint ${item.constraint.type} for table '${tableName}'`,
178+
);
179+
}
180+
}
181+
}
182+
}
183+
if (pkConstraint.length === 0) {
184+
return undefined;
185+
}
186+
return pkConstraint.join(' ');
187+
}
188+
189+
// This strips out all the things that our SQL AST chokes on (it's still in an
190+
// experimental phase for postgresql)
191+
function prepareDump(sql: string): string {
192+
let result = sql
193+
.replace(/\s*SET\s[^;].*;/gm, '')
194+
.replace(/\s*CREATE\sTYPE\s[^;]*;/gm, '');
195+
return result;
196+
}
197+
198+
function getSchemaFilename(): string {
199+
let files = readdirSync(migrationsDir);
200+
let lastFile = files
201+
.filter((f) => f !== '.eslintrc.js')
202+
.sort()
203+
.pop()!;
204+
return `${lastFile.replace(/_.*/, '')}_schema.sql`;
205+
}
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,8 @@
1+
#!/bin/sh
2+
3+
if docker exec boxel-pg psql -U postgres -w -lqt | cut -d \| -f 1 | grep -qw "$PGDATABASE"; then
4+
echo "Database $PGDATABASE exists"
5+
else
6+
docker exec boxel-pg psql -U postgres -w -c "CREATE DATABASE $PGDATABASE"
7+
echo "created database $PGDATABASE"
8+
fi

0 commit comments

Comments
 (0)