full database extracton
This commit is contained in:
@ -29,12 +29,15 @@ use std::collections::HashMap;
|
|||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use r#type::Type;
|
use r#type::Type;
|
||||||
|
|
||||||
|
#[derive(serde::Serialize)]
|
||||||
pub struct Database {
|
pub struct Database {
|
||||||
pub enums: HashMap<String, Enum>,
|
pub enums: HashMap<String, Enum>,
|
||||||
pub types: HashMap<String, Type>,
|
pub types: HashMap<String, Type>,
|
||||||
pub puncs: HashMap<String, Punc>,
|
pub puncs: HashMap<String, Punc>,
|
||||||
pub relations: HashMap<String, Relation>,
|
pub relations: HashMap<String, Relation>,
|
||||||
|
#[serde(skip)]
|
||||||
pub schemas: HashMap<String, Arc<Schema>>,
|
pub schemas: HashMap<String, Arc<Schema>>,
|
||||||
|
#[serde(skip)]
|
||||||
pub executor: Box<dyn DatabaseExecutor + Send + Sync>,
|
pub executor: Box<dyn DatabaseExecutor + Send + Sync>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -238,23 +241,30 @@ impl Database {
|
|||||||
|
|
||||||
// Phase 2: Synthesize Composed Filter References
|
// Phase 2: Synthesize Composed Filter References
|
||||||
let mut filter_schemas = Vec::new();
|
let mut filter_schemas = Vec::new();
|
||||||
for type_def in self.types.values() {
|
for (type_name, type_def) in &self.types {
|
||||||
for (id, schema_arc) in &type_def.schemas {
|
for (id, schema_arc) in &type_def.schemas {
|
||||||
// Only run synthesis on actual structured, table-backed boundaries. Exclude subschemas!
|
// Only run synthesis on actual structured, table-backed boundaries. Exclude subschemas!
|
||||||
let base_name = id.split('.').last().unwrap_or(id);
|
let base_name = id.split('.').last().unwrap_or(id);
|
||||||
let is_table_backed = base_name == type_def.name;
|
let is_table_backed = base_name == type_def.name;
|
||||||
if is_table_backed && !id.contains('/') {
|
if is_table_backed && !id.contains('/') {
|
||||||
if let Some(filter_schema) = schema_arc.compile_filter(self, id, errors) {
|
if let Some(filter_schema) = schema_arc.compile_filter(self, id, errors) {
|
||||||
filter_schemas.push((format!("{}.filter", id), Arc::new(filter_schema)));
|
filter_schemas.push((
|
||||||
|
type_name.clone(),
|
||||||
|
format!("{}.filter", id),
|
||||||
|
Arc::new(filter_schema),
|
||||||
|
));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut filter_ids = Vec::new();
|
let mut filter_ids = Vec::new();
|
||||||
for (id, filter_arc) in filter_schemas {
|
for (type_name, id, filter_arc) in filter_schemas {
|
||||||
filter_ids.push(id.clone());
|
filter_ids.push(id.clone());
|
||||||
self.schemas.insert(id, filter_arc);
|
self.schemas.insert(id.clone(), filter_arc.clone());
|
||||||
|
if let Some(t) = self.types.get_mut(&type_name) {
|
||||||
|
t.schemas.insert(id, filter_arc);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Now actively compile the newly injected filters to lock all nested compose references natively
|
// Now actively compile the newly injected filters to lock all nested compose references natively
|
||||||
@ -269,50 +279,88 @@ impl Database {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn collect_schemas(&mut self, errors: &mut Vec<crate::drop::Error>) {
|
fn collect_schemas(&mut self, errors: &mut Vec<crate::drop::Error>) {
|
||||||
let mut to_insert = Vec::new();
|
let mut type_insert = Vec::new();
|
||||||
|
let mut punc_insert = Vec::new();
|
||||||
|
let mut enum_insert = Vec::new();
|
||||||
|
let mut global_insert = Vec::new();
|
||||||
|
|
||||||
// Pass 1: Extract all Schemas structurally off top level definitions into the master registry.
|
// Pass 1: Extract all Schemas structurally off top level definitions into the master registry.
|
||||||
// Validate every node recursively via string filters natively!
|
// Validate every node recursively via string filters natively!
|
||||||
for type_def in self.types.values() {
|
for (type_name, type_def) in &self.types {
|
||||||
for (id, schema_arc) in &type_def.schemas {
|
for (id, schema_arc) in &type_def.schemas {
|
||||||
to_insert.push((id.clone(), Arc::clone(schema_arc)));
|
global_insert.push((id.clone(), Arc::clone(schema_arc)));
|
||||||
|
let mut local_insert = Vec::new();
|
||||||
crate::database::schema::Schema::collect_schemas(
|
crate::database::schema::Schema::collect_schemas(
|
||||||
schema_arc,
|
schema_arc,
|
||||||
id,
|
id,
|
||||||
id.clone(),
|
id.clone(),
|
||||||
&mut to_insert,
|
&mut local_insert,
|
||||||
errors,
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for punc_def in self.puncs.values() {
|
|
||||||
for (id, schema_arc) in &punc_def.schemas {
|
|
||||||
to_insert.push((id.clone(), Arc::clone(schema_arc)));
|
|
||||||
crate::database::schema::Schema::collect_schemas(
|
|
||||||
schema_arc,
|
|
||||||
id,
|
|
||||||
id.clone(),
|
|
||||||
&mut to_insert,
|
|
||||||
errors,
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for enum_def in self.enums.values() {
|
|
||||||
for (id, schema_arc) in &enum_def.schemas {
|
|
||||||
to_insert.push((id.clone(), Arc::clone(schema_arc)));
|
|
||||||
crate::database::schema::Schema::collect_schemas(
|
|
||||||
schema_arc,
|
|
||||||
id,
|
|
||||||
id.clone(),
|
|
||||||
&mut to_insert,
|
|
||||||
errors,
|
errors,
|
||||||
);
|
);
|
||||||
|
for entry in &local_insert {
|
||||||
|
type_insert.push((type_name.clone(), entry.0.clone(), Arc::clone(&entry.1)));
|
||||||
|
global_insert.push((entry.0.clone(), Arc::clone(&entry.1)));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
for (id, schema_arc) in to_insert {
|
for (punc_name, punc_def) in &self.puncs {
|
||||||
|
for (id, schema_arc) in &punc_def.schemas {
|
||||||
|
global_insert.push((id.clone(), Arc::clone(schema_arc)));
|
||||||
|
let mut local_insert = Vec::new();
|
||||||
|
crate::database::schema::Schema::collect_schemas(
|
||||||
|
schema_arc,
|
||||||
|
id,
|
||||||
|
id.clone(),
|
||||||
|
&mut local_insert,
|
||||||
|
errors,
|
||||||
|
);
|
||||||
|
for entry in &local_insert {
|
||||||
|
punc_insert.push((punc_name.clone(), entry.0.clone(), Arc::clone(&entry.1)));
|
||||||
|
global_insert.push((entry.0.clone(), Arc::clone(&entry.1)));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for (enum_name, enum_def) in &self.enums {
|
||||||
|
for (id, schema_arc) in &enum_def.schemas {
|
||||||
|
global_insert.push((id.clone(), Arc::clone(schema_arc)));
|
||||||
|
let mut local_insert = Vec::new();
|
||||||
|
crate::database::schema::Schema::collect_schemas(
|
||||||
|
schema_arc,
|
||||||
|
id,
|
||||||
|
id.clone(),
|
||||||
|
&mut local_insert,
|
||||||
|
errors,
|
||||||
|
);
|
||||||
|
for entry in &local_insert {
|
||||||
|
enum_insert.push((enum_name.clone(), entry.0.clone(), Arc::clone(&entry.1)));
|
||||||
|
global_insert.push((entry.0.clone(), Arc::clone(&entry.1)));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Apply global inserts
|
||||||
|
for (id, schema_arc) in global_insert {
|
||||||
self.schemas.insert(id, schema_arc);
|
self.schemas.insert(id, schema_arc);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Apply local scopes
|
||||||
|
for (origin_name, id, schema_arc) in type_insert {
|
||||||
|
if let Some(t) = self.types.get_mut(&origin_name) {
|
||||||
|
t.schemas.insert(id, schema_arc);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for (origin_name, id, schema_arc) in punc_insert {
|
||||||
|
if let Some(p) = self.puncs.get_mut(&origin_name) {
|
||||||
|
p.schemas.insert(id, schema_arc);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for (origin_name, id, schema_arc) in enum_insert {
|
||||||
|
if let Some(e) = self.enums.get_mut(&origin_name) {
|
||||||
|
e.schemas.insert(id, schema_arc);
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Inspects the Postgres pg_constraint relations catalog to securely identify
|
/// Inspects the Postgres pg_constraint relations catalog to securely identify
|
||||||
|
|||||||
@ -109,7 +109,7 @@ pub fn jspg_validate(schema_id: &str, instance: JsonB) -> JsonB {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[cfg_attr(not(test), pg_extern)]
|
#[cfg_attr(not(test), pg_extern)]
|
||||||
pub fn jspg_schemas() -> JsonB {
|
pub fn jspg_database() -> JsonB {
|
||||||
let engine_opt = {
|
let engine_opt = {
|
||||||
let lock = GLOBAL_JSPG.read().unwrap();
|
let lock = GLOBAL_JSPG.read().unwrap();
|
||||||
lock.clone()
|
lock.clone()
|
||||||
@ -117,9 +117,9 @@ pub fn jspg_schemas() -> JsonB {
|
|||||||
|
|
||||||
match engine_opt {
|
match engine_opt {
|
||||||
Some(engine) => {
|
Some(engine) => {
|
||||||
let schemas_json = serde_json::to_value(&engine.database.schemas)
|
let database_json = serde_json::to_value(&engine.database)
|
||||||
.unwrap_or(serde_json::Value::Object(serde_json::Map::new()));
|
.unwrap_or(serde_json::Value::Object(serde_json::Map::new()));
|
||||||
let drop = crate::drop::Drop::success_with_val(schemas_json);
|
let drop = crate::drop::Drop::success_with_val(database_json);
|
||||||
JsonB(serde_json::to_value(drop).unwrap())
|
JsonB(serde_json::to_value(drop).unwrap())
|
||||||
}
|
}
|
||||||
None => jspg_failure(),
|
None => jspg_failure(),
|
||||||
|
|||||||
142
src/tests/mod.rs
142
src/tests/mod.rs
@ -81,54 +81,114 @@ fn test_library_api() {
|
|||||||
})
|
})
|
||||||
);
|
);
|
||||||
|
|
||||||
// 3. Validate jspg_schemas
|
// 3. Validate jspg_database mapping natively!
|
||||||
let schemas_drop = jspg_schemas();
|
let db_drop = jspg_database();
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
schemas_drop.0,
|
db_drop.0,
|
||||||
json!({
|
json!({
|
||||||
"type": "drop",
|
"type": "drop",
|
||||||
"response": {
|
"response": {
|
||||||
"source_schema": {
|
"enums": {},
|
||||||
"type": "object",
|
"puncs": {},
|
||||||
"properties": {
|
"relations": {
|
||||||
"type": { "type": "string" },
|
"fk_test_target": {
|
||||||
"name": { "type": "string" },
|
"constraint": "fk_test_target",
|
||||||
"target": {
|
"destination_columns": ["id"],
|
||||||
"type": "target_schema",
|
"destination_type": "target_schema",
|
||||||
"compiledPropertyNames": ["value"]
|
"prefix": "target",
|
||||||
}
|
"source_columns": ["target_id"],
|
||||||
},
|
"source_type": "source_schema"
|
||||||
"required": ["name"],
|
|
||||||
"compiledPropertyNames": ["name", "target", "type"],
|
|
||||||
"compiledEdges": {
|
|
||||||
"target": {
|
|
||||||
"constraint": "fk_test_target",
|
|
||||||
"forward": true
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"source_schema.filter": {
|
"types": {
|
||||||
"type": "object",
|
"source_schema": {
|
||||||
"properties": {
|
"default_fields": [],
|
||||||
"type": { "type": ["string.condition", "null"] },
|
"field_types": null,
|
||||||
"name": { "type": ["string.condition", "null"] },
|
"fields": [],
|
||||||
"target": { "type": ["target_schema.filter", "null"] }
|
"grouped_fields": null,
|
||||||
|
"hierarchy": ["source_schema", "entity"],
|
||||||
|
"historical": false,
|
||||||
|
"id": "",
|
||||||
|
"longevity": null,
|
||||||
|
"lookup_fields": [],
|
||||||
|
"module": "",
|
||||||
|
"name": "source_schema",
|
||||||
|
"notify": false,
|
||||||
|
"null_fields": [],
|
||||||
|
"ownable": false,
|
||||||
|
"relationship": false,
|
||||||
|
"schemas": {
|
||||||
|
"source_schema": {
|
||||||
|
"compiledEdges": {
|
||||||
|
"target": {
|
||||||
|
"constraint": "fk_test_target",
|
||||||
|
"forward": true
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"compiledPropertyNames": ["name", "target", "type"],
|
||||||
|
"properties": {
|
||||||
|
"name": { "type": "string" },
|
||||||
|
"target": {
|
||||||
|
"compiledPropertyNames": ["value"],
|
||||||
|
"type": "target_schema"
|
||||||
|
},
|
||||||
|
"type": { "type": "string" }
|
||||||
|
},
|
||||||
|
"required": ["name"],
|
||||||
|
"type": "object"
|
||||||
|
},
|
||||||
|
"source_schema.filter": {
|
||||||
|
"compiledPropertyNames": ["name", "target", "type"],
|
||||||
|
"properties": {
|
||||||
|
"name": { "type": ["string.condition", "null"] },
|
||||||
|
"target": { "type": ["target_schema.filter", "null"] },
|
||||||
|
"type": { "type": ["string.condition", "null"] }
|
||||||
|
},
|
||||||
|
"type": "object"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"sensitive": false,
|
||||||
|
"source": "",
|
||||||
|
"type": "",
|
||||||
|
"variations": ["source_schema"]
|
||||||
},
|
},
|
||||||
"compiledPropertyNames": ["name", "target", "type"]
|
"target_schema": {
|
||||||
},
|
"default_fields": [],
|
||||||
"target_schema": {
|
"field_types": null,
|
||||||
"type": "object",
|
"fields": [],
|
||||||
"properties": {
|
"grouped_fields": null,
|
||||||
"value": { "type": "number" }
|
"hierarchy": ["target_schema", "entity"],
|
||||||
},
|
"historical": false,
|
||||||
"compiledPropertyNames": ["value"]
|
"id": "",
|
||||||
},
|
"longevity": null,
|
||||||
"target_schema.filter": {
|
"lookup_fields": [],
|
||||||
"type": "object",
|
"module": "",
|
||||||
"properties": {
|
"name": "target_schema",
|
||||||
"value": { "type": ["number.condition", "null"] }
|
"notify": false,
|
||||||
},
|
"null_fields": [],
|
||||||
"compiledPropertyNames": ["value"]
|
"ownable": false,
|
||||||
|
"relationship": false,
|
||||||
|
"schemas": {
|
||||||
|
"target_schema": {
|
||||||
|
"compiledPropertyNames": ["value"],
|
||||||
|
"properties": {
|
||||||
|
"value": { "type": "number" }
|
||||||
|
},
|
||||||
|
"type": "object"
|
||||||
|
},
|
||||||
|
"target_schema.filter": {
|
||||||
|
"compiledPropertyNames": ["value"],
|
||||||
|
"properties": {
|
||||||
|
"value": { "type": ["number.condition", "null"] }
|
||||||
|
},
|
||||||
|
"type": "object"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"sensitive": false,
|
||||||
|
"source": "",
|
||||||
|
"type": "",
|
||||||
|
"variations": ["target_schema"]
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|||||||
Reference in New Issue
Block a user