Compare commits
7 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| 59395a33ac | |||
| 92c0a6fc0b | |||
| 7f66a4a35a | |||
| d37aadb0dd | |||
| d0ccc47d97 | |||
| 2d19bf100e | |||
| fb333c6cbb |
1
rustfmt.toml
Normal file
1
rustfmt.toml
Normal file
@ -0,0 +1 @@
|
||||
tab_spaces = 2
|
||||
456
src/lib.rs
456
src/lib.rs
@ -2,16 +2,26 @@ use pgrx::*;
|
||||
|
||||
pg_module_magic!();
|
||||
|
||||
use serde_json::{json, Value};
|
||||
use std::{collections::HashMap, sync::RwLock};
|
||||
use boon::{Compiler, Schemas, ValidationError, SchemaIndex, CompileError};
|
||||
use boon::{CompileError, Compiler, ErrorKind, SchemaIndex, Schemas, ValidationError};
|
||||
use lazy_static::lazy_static;
|
||||
use serde_json::{json, Value};
|
||||
use std::collections::hash_map::Entry;
|
||||
use std::{collections::HashMap, sync::RwLock};
|
||||
|
||||
struct BoonCache {
|
||||
schemas: Schemas,
|
||||
id_to_index: HashMap<String, SchemaIndex>,
|
||||
}
|
||||
|
||||
// Structure to hold error information without lifetimes
|
||||
#[derive(Debug)]
|
||||
struct Error {
|
||||
path: String,
|
||||
code: String,
|
||||
message: String,
|
||||
cause: String,
|
||||
}
|
||||
|
||||
lazy_static! {
|
||||
static ref SCHEMA_CACHE: RwLock<BoonCache> = RwLock::new(BoonCache {
|
||||
schemas: Schemas::new(),
|
||||
@ -20,11 +30,17 @@ lazy_static! {
|
||||
}
|
||||
|
||||
#[pg_extern(strict)]
|
||||
fn cache_json_schema(schema_id: &str, schema: JsonB) -> JsonB {
|
||||
fn cache_json_schema(schema_id: &str, schema: JsonB, strict: bool) -> JsonB {
|
||||
let mut cache = SCHEMA_CACHE.write().unwrap();
|
||||
let schema_value: Value = schema.0;
|
||||
let mut schema_value: Value = schema.0;
|
||||
let schema_path = format!("urn:{}", schema_id);
|
||||
|
||||
// Apply strict validation to all objects in the schema if requested
|
||||
if strict {
|
||||
apply_strict_validation(&mut schema_value);
|
||||
}
|
||||
|
||||
// Create the boon compiler and enable format assertions
|
||||
let mut compiler = Compiler::new();
|
||||
compiler.enable_format_assertions();
|
||||
|
||||
@ -54,9 +70,9 @@ fn cache_json_schema(schema_id: &str, schema: JsonB) -> JsonB {
|
||||
CompileError::ValidationError { url: _url, src } => {
|
||||
// Collect leaf errors from the meta-schema validation failure
|
||||
let mut error_list = Vec::new();
|
||||
collect_validation_errors(src, &mut error_list);
|
||||
collect_errors(src, &mut error_list);
|
||||
// Filter and format errors properly - no instance for schema compilation
|
||||
format_drop_errors(error_list, &schema_value)
|
||||
format_errors(error_list, &schema_value)
|
||||
}
|
||||
_ => {
|
||||
// Other compilation errors
|
||||
@ -75,10 +91,34 @@ fn cache_json_schema(schema_id: &str, schema: JsonB) -> JsonB {
|
||||
}
|
||||
}
|
||||
|
||||
// Helper function to recursively apply strict validation to all objects in a schema
|
||||
fn apply_strict_validation(schema: &mut Value) {
|
||||
match schema {
|
||||
Value::Object(map) => {
|
||||
// If this is an object type schema, add additionalProperties: false
|
||||
if let Some(Value::String(t)) = map.get("type") {
|
||||
if t == "object" && !map.contains_key("additionalProperties") {
|
||||
map.insert("additionalProperties".to_string(), Value::Bool(false));
|
||||
}
|
||||
}
|
||||
// Recurse into all properties
|
||||
for (_, value) in map.iter_mut() {
|
||||
apply_strict_validation(value);
|
||||
}
|
||||
}
|
||||
Value::Array(arr) => {
|
||||
// Recurse into array items
|
||||
for item in arr.iter_mut() {
|
||||
apply_strict_validation(item);
|
||||
}
|
||||
}
|
||||
_ => {}
|
||||
}
|
||||
}
|
||||
|
||||
#[pg_extern(strict, parallel_safe)]
|
||||
fn validate_json_schema(schema_id: &str, instance: JsonB) -> JsonB {
|
||||
let cache = SCHEMA_CACHE.read().unwrap();
|
||||
|
||||
// Lookup uses the original schema_id
|
||||
match cache.id_to_index.get(schema_id) {
|
||||
None => JsonB(json!({
|
||||
@ -96,10 +136,9 @@ fn validate_json_schema(schema_id: &str, instance: JsonB) -> JsonB {
|
||||
Ok(_) => JsonB(json!({ "response": "success" })),
|
||||
Err(validation_error) => {
|
||||
let mut error_list = Vec::new();
|
||||
collect_validation_errors(&validation_error, &mut error_list);
|
||||
let errors = format_drop_errors(error_list, &instance_value);
|
||||
|
||||
JsonB(json!({ "errors": errors }))
|
||||
collect_errors(&validation_error, &mut error_list);
|
||||
let errors = format_errors(error_list, &instance_value);
|
||||
JsonB(json!({ "errors": errors }))
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -107,206 +146,244 @@ fn validate_json_schema(schema_id: &str, instance: JsonB) -> JsonB {
|
||||
}
|
||||
|
||||
// Recursively collects validation errors
|
||||
fn collect_validation_errors(error: &ValidationError, errors_list: &mut Vec<(String, String, String)>) {
|
||||
fn collect_errors(error: &ValidationError, errors_list: &mut Vec<Error>) {
|
||||
// Check if this is a structural error that we should skip
|
||||
let error_message = format!("{}", error.kind);
|
||||
let is_structural = error_message == "validation failed" ||
|
||||
error_message == "allOf failed" ||
|
||||
error_message == "anyOf failed" ||
|
||||
error_message == "not failed" ||
|
||||
error_message.starts_with("oneOf failed");
|
||||
let is_structural = matches!(
|
||||
&error.kind,
|
||||
ErrorKind::Group | ErrorKind::AllOf | ErrorKind::AnyOf | ErrorKind::Not | ErrorKind::OneOf(_)
|
||||
);
|
||||
|
||||
if error.causes.is_empty() && !is_structural {
|
||||
// This is a leaf error that's not structural
|
||||
// Format just the error kind, not the whole validation error
|
||||
let message = format!("{}", error.kind);
|
||||
let original_message = format!("{}", error.kind);
|
||||
let (error_code, human_message) = convert_error_kind(&error.kind);
|
||||
|
||||
errors_list.push((
|
||||
error.instance_location.to_string(),
|
||||
error.schema_url.to_string(),
|
||||
message
|
||||
));
|
||||
errors_list.push(Error {
|
||||
path: error.instance_location.to_string(),
|
||||
code: error_code,
|
||||
message: human_message,
|
||||
cause: original_message,
|
||||
});
|
||||
} else {
|
||||
// Recurse into causes
|
||||
for cause in &error.causes {
|
||||
collect_validation_errors(cause, errors_list);
|
||||
collect_errors(cause, errors_list);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Convert ErrorKind to error code and human message
|
||||
fn convert_error_kind(kind: &ErrorKind) -> (String, String) {
|
||||
match kind {
|
||||
ErrorKind::Type { .. } => (
|
||||
"TYPE_MISMATCH".to_string(),
|
||||
"Field type does not match the expected type".to_string(),
|
||||
),
|
||||
ErrorKind::Required { .. } => (
|
||||
"REQUIRED_FIELD_MISSING".to_string(),
|
||||
"Required field is missing".to_string(),
|
||||
),
|
||||
ErrorKind::DependentRequired { .. } => (
|
||||
"DEPENDENT_REQUIRED_MISSING".to_string(),
|
||||
"Dependent required fields are missing".to_string(),
|
||||
),
|
||||
ErrorKind::Dependency { .. } => (
|
||||
"DEPENDENCY_FAILED".to_string(),
|
||||
"Dependency requirement not met".to_string(),
|
||||
),
|
||||
ErrorKind::Enum { .. } => (
|
||||
"ENUM_VIOLATED".to_string(),
|
||||
"Value is not one of the allowed options".to_string(),
|
||||
),
|
||||
ErrorKind::Const { .. } => (
|
||||
"CONST_VIOLATED".to_string(),
|
||||
"Value does not match the required constant".to_string(),
|
||||
),
|
||||
ErrorKind::MinLength { .. } => (
|
||||
"MIN_LENGTH_VIOLATED".to_string(),
|
||||
"Field length is below the minimum required".to_string(),
|
||||
),
|
||||
ErrorKind::MaxLength { .. } => (
|
||||
"MAX_LENGTH_VIOLATED".to_string(),
|
||||
"Field length exceeds the maximum allowed".to_string(),
|
||||
),
|
||||
ErrorKind::Pattern { .. } => (
|
||||
"PATTERN_VIOLATED".to_string(),
|
||||
"Value does not match the required pattern".to_string(),
|
||||
),
|
||||
ErrorKind::Minimum { .. } => (
|
||||
"MINIMUM_VIOLATED".to_string(),
|
||||
"Value is below the minimum allowed".to_string(),
|
||||
),
|
||||
ErrorKind::Maximum { .. } => (
|
||||
"MAXIMUM_VIOLATED".to_string(),
|
||||
"Value exceeds the maximum allowed".to_string(),
|
||||
),
|
||||
ErrorKind::ExclusiveMinimum { .. } => (
|
||||
"EXCLUSIVE_MINIMUM_VIOLATED".to_string(),
|
||||
"Value must be greater than the minimum".to_string(),
|
||||
),
|
||||
ErrorKind::ExclusiveMaximum { .. } => (
|
||||
"EXCLUSIVE_MAXIMUM_VIOLATED".to_string(),
|
||||
"Value must be less than the maximum".to_string(),
|
||||
),
|
||||
ErrorKind::MultipleOf { .. } => (
|
||||
"MULTIPLE_OF_VIOLATED".to_string(),
|
||||
"Value is not a multiple of the required factor".to_string(),
|
||||
),
|
||||
ErrorKind::MinItems { .. } => (
|
||||
"MIN_ITEMS_VIOLATED".to_string(),
|
||||
"Array has fewer items than required".to_string(),
|
||||
),
|
||||
ErrorKind::MaxItems { .. } => (
|
||||
"MAX_ITEMS_VIOLATED".to_string(),
|
||||
"Array has more items than allowed".to_string(),
|
||||
),
|
||||
ErrorKind::UniqueItems { .. } => (
|
||||
"UNIQUE_ITEMS_VIOLATED".to_string(),
|
||||
"Array contains duplicate items".to_string(),
|
||||
),
|
||||
ErrorKind::MinProperties { .. } => (
|
||||
"MIN_PROPERTIES_VIOLATED".to_string(),
|
||||
"Object has fewer properties than required".to_string(),
|
||||
),
|
||||
ErrorKind::MaxProperties { .. } => (
|
||||
"MAX_PROPERTIES_VIOLATED".to_string(),
|
||||
"Object has more properties than allowed".to_string(),
|
||||
),
|
||||
ErrorKind::AdditionalProperties { .. } => (
|
||||
"ADDITIONAL_PROPERTIES_NOT_ALLOWED".to_string(),
|
||||
"Object contains properties that are not allowed".to_string(),
|
||||
),
|
||||
ErrorKind::AdditionalItems { .. } => (
|
||||
"ADDITIONAL_ITEMS_NOT_ALLOWED".to_string(),
|
||||
"Array contains additional items that are not allowed".to_string(),
|
||||
),
|
||||
ErrorKind::Format { want, .. } => (
|
||||
"FORMAT_INVALID".to_string(),
|
||||
format!("Invalid {} format", want),
|
||||
),
|
||||
ErrorKind::PropertyName { .. } => (
|
||||
"INVALID_PROPERTY_NAME".to_string(),
|
||||
"Property name is invalid".to_string(),
|
||||
),
|
||||
ErrorKind::Contains => (
|
||||
"CONTAINS_FAILED".to_string(),
|
||||
"No items match the required schema".to_string(),
|
||||
),
|
||||
ErrorKind::MinContains { .. } => (
|
||||
"MIN_CONTAINS_VIOLATED".to_string(),
|
||||
"Too few items match the required schema".to_string(),
|
||||
),
|
||||
ErrorKind::MaxContains { .. } => (
|
||||
"MAX_CONTAINS_VIOLATED".to_string(),
|
||||
"Too many items match the required schema".to_string(),
|
||||
),
|
||||
ErrorKind::ContentEncoding { .. } => (
|
||||
"CONTENT_ENCODING_INVALID".to_string(),
|
||||
"Content encoding is invalid".to_string(),
|
||||
),
|
||||
ErrorKind::ContentMediaType { .. } => (
|
||||
"CONTENT_MEDIA_TYPE_INVALID".to_string(),
|
||||
"Content media type is invalid".to_string(),
|
||||
),
|
||||
ErrorKind::FalseSchema => (
|
||||
"FALSE_SCHEMA".to_string(),
|
||||
"Schema validation always fails".to_string(),
|
||||
),
|
||||
ErrorKind::Not => (
|
||||
"NOT_VIOLATED".to_string(),
|
||||
"Value matched a schema it should not match".to_string(),
|
||||
),
|
||||
ErrorKind::RefCycle { .. } => (
|
||||
"REFERENCE_CYCLE".to_string(),
|
||||
"Schema contains a reference cycle".to_string(),
|
||||
),
|
||||
ErrorKind::Reference { .. } => (
|
||||
"REFERENCE_FAILED".to_string(),
|
||||
"Reference validation failed".to_string(),
|
||||
),
|
||||
ErrorKind::Schema { .. } => (
|
||||
"SCHEMA_FAILED".to_string(),
|
||||
"Schema validation failed".to_string(),
|
||||
),
|
||||
ErrorKind::ContentSchema => (
|
||||
"CONTENT_SCHEMA_FAILED".to_string(),
|
||||
"Content schema validation failed".to_string(),
|
||||
),
|
||||
// These shouldn't appear as leaf errors due to is_structural check
|
||||
ErrorKind::Group => (
|
||||
"VALIDATION_FAILED".to_string(),
|
||||
"Validation failed".to_string(),
|
||||
),
|
||||
ErrorKind::AllOf => (
|
||||
"ALL_OF_VIOLATED".to_string(),
|
||||
"Value does not match all required schemas".to_string(),
|
||||
),
|
||||
ErrorKind::AnyOf => (
|
||||
"ANY_OF_VIOLATED".to_string(),
|
||||
"Value does not match any of the allowed schemas".to_string(),
|
||||
),
|
||||
ErrorKind::OneOf(_) => (
|
||||
"ONE_OF_VIOLATED".to_string(),
|
||||
"Value must match exactly one schema".to_string(),
|
||||
),
|
||||
}
|
||||
}
|
||||
|
||||
// Formats errors according to DropError structure
|
||||
fn format_drop_errors(raw_errors: Vec<(String, String, String)>, instance: &Value) -> Vec<Value> {
|
||||
use std::collections::HashMap;
|
||||
use std::collections::hash_map::Entry;
|
||||
|
||||
// We don't filter structural paths from instance paths anymore
|
||||
// because instance paths shouldn't contain these segments anyway
|
||||
// The issue was likely with schema paths, not instance paths
|
||||
let plausible_errors = raw_errors;
|
||||
|
||||
// 2. Deduplicate by instance_path and format as DropError
|
||||
let mut unique_errors: HashMap<String, Value> = HashMap::new();
|
||||
for (instance_path, schema_path, message) in plausible_errors {
|
||||
if let Entry::Vacant(entry) = unique_errors.entry(instance_path.clone()) {
|
||||
// Convert message to error code and make it human readable
|
||||
let (code, human_message) = enhance_error_message(&message);
|
||||
|
||||
// Extract the failing value from the instance
|
||||
let failing_value = extract_value_at_path(instance, &instance_path);
|
||||
|
||||
entry.insert(json!({
|
||||
"code": code,
|
||||
"message": human_message,
|
||||
"details": {
|
||||
"path": schema_path,
|
||||
"context": json!({
|
||||
"instance_path": instance_path,
|
||||
"failing_value": failing_value
|
||||
}),
|
||||
"cause": message // Original error message
|
||||
}
|
||||
}));
|
||||
}
|
||||
fn format_errors(errors: Vec<Error>, instance: &Value) -> Vec<Value> {
|
||||
// Deduplicate by instance_path and format as DropError
|
||||
let mut unique_errors: HashMap<String, Value> = HashMap::new();
|
||||
for error in errors {
|
||||
if let Entry::Vacant(entry) = unique_errors.entry(error.path.clone()) {
|
||||
// Extract the failing value from the instance
|
||||
let failing_value = extract_value_at_path(instance, &error.path);
|
||||
entry.insert(json!({
|
||||
"code": error.code,
|
||||
"message": error.message,
|
||||
"details": {
|
||||
"path": error.path,
|
||||
"context": failing_value,
|
||||
"cause": error.cause
|
||||
}
|
||||
}));
|
||||
}
|
||||
}
|
||||
|
||||
unique_errors.into_values().collect()
|
||||
unique_errors.into_values().collect()
|
||||
}
|
||||
|
||||
// Helper function to extract value at a JSON pointer path
|
||||
fn extract_value_at_path(instance: &Value, path: &str) -> Value {
|
||||
let parts: Vec<&str> = path.split('/').filter(|s| !s.is_empty()).collect();
|
||||
let mut current = instance;
|
||||
let parts: Vec<&str> = path.split('/').filter(|s| !s.is_empty()).collect();
|
||||
let mut current = instance;
|
||||
|
||||
for part in parts {
|
||||
match current {
|
||||
Value::Object(map) => {
|
||||
if let Some(value) = map.get(part) {
|
||||
current = value;
|
||||
} else {
|
||||
return Value::Null;
|
||||
}
|
||||
}
|
||||
Value::Array(arr) => {
|
||||
if let Ok(index) = part.parse::<usize>() {
|
||||
if let Some(value) = arr.get(index) {
|
||||
current = value;
|
||||
} else {
|
||||
return Value::Null;
|
||||
}
|
||||
} else {
|
||||
return Value::Null;
|
||||
}
|
||||
}
|
||||
_ => return Value::Null,
|
||||
for part in parts {
|
||||
match current {
|
||||
Value::Object(map) => {
|
||||
if let Some(value) = map.get(part) {
|
||||
current = value;
|
||||
} else {
|
||||
return Value::Null;
|
||||
}
|
||||
}
|
||||
Value::Array(arr) => {
|
||||
if let Ok(index) = part.parse::<usize>() {
|
||||
if let Some(value) = arr.get(index) {
|
||||
current = value;
|
||||
} else {
|
||||
return Value::Null;
|
||||
}
|
||||
} else {
|
||||
return Value::Null;
|
||||
}
|
||||
}
|
||||
_ => return Value::Null,
|
||||
}
|
||||
}
|
||||
|
||||
current.clone()
|
||||
}
|
||||
|
||||
// Helper to convert validation messages to error codes and human-readable messages
|
||||
fn enhance_error_message(message: &str) -> (String, String) {
|
||||
// Match exact boon error message patterns
|
||||
let trimmed = message.trim();
|
||||
|
||||
if trimmed.contains("value must be one of") {
|
||||
("ENUM_VIOLATED".to_string(),
|
||||
"Value is not one of the allowed options".to_string())
|
||||
} else if trimmed.contains("length must be >=") && trimmed.contains("but got") {
|
||||
("MIN_LENGTH_VIOLATED".to_string(),
|
||||
"Field length is below the minimum required".to_string())
|
||||
} else if trimmed.contains("length must be <=") && trimmed.contains("but got") {
|
||||
("MAX_LENGTH_VIOLATED".to_string(),
|
||||
"Field length exceeds the maximum allowed".to_string())
|
||||
} else if trimmed.contains("must be >=") && trimmed.contains("but got") {
|
||||
("MINIMUM_VIOLATED".to_string(),
|
||||
"Value is below the minimum allowed".to_string())
|
||||
} else if trimmed.contains("must be <=") && trimmed.contains("but got") {
|
||||
("MAXIMUM_VIOLATED".to_string(),
|
||||
"Value exceeds the maximum allowed".to_string())
|
||||
} else if trimmed.contains("must be >") && trimmed.contains("but got") {
|
||||
("EXCLUSIVE_MINIMUM_VIOLATED".to_string(),
|
||||
"Value must be greater than the minimum".to_string())
|
||||
} else if trimmed.contains("must be <") && trimmed.contains("but got") {
|
||||
("EXCLUSIVE_MAXIMUM_VIOLATED".to_string(),
|
||||
"Value must be less than the maximum".to_string())
|
||||
} else if trimmed.contains("does not match pattern") {
|
||||
("PATTERN_VIOLATED".to_string(),
|
||||
"Value does not match the required pattern".to_string())
|
||||
} else if trimmed.contains("missing properties") {
|
||||
("REQUIRED_FIELD_MISSING".to_string(),
|
||||
"Required field is missing".to_string())
|
||||
} else if trimmed.contains("want") && trimmed.contains("but got") {
|
||||
("TYPE_MISMATCH".to_string(),
|
||||
"Field type does not match the expected type".to_string())
|
||||
} else if trimmed.starts_with("value must be") && !trimmed.contains("one of") {
|
||||
("CONST_VIOLATED".to_string(),
|
||||
"Value does not match the required constant".to_string())
|
||||
} else if trimmed.contains("is not valid") && trimmed.contains(":") {
|
||||
("FORMAT_INVALID".to_string(),
|
||||
extract_format_message(trimmed))
|
||||
} else if trimmed.contains("items at") && trimmed.contains("are equal") {
|
||||
("UNIQUE_ITEMS_VIOLATED".to_string(),
|
||||
"Array contains duplicate items".to_string())
|
||||
} else if trimmed.contains("additionalProperties") && trimmed.contains("not allowed") {
|
||||
("ADDITIONAL_PROPERTIES_NOT_ALLOWED".to_string(),
|
||||
"Object contains properties that are not allowed".to_string())
|
||||
} else if trimmed.contains("is not multipleOf") {
|
||||
("MULTIPLE_OF_VIOLATED".to_string(),
|
||||
"Value is not a multiple of the required factor".to_string())
|
||||
} else if trimmed.contains("minimum") && trimmed.contains("properties required") {
|
||||
("MIN_PROPERTIES_VIOLATED".to_string(),
|
||||
"Object has fewer properties than required".to_string())
|
||||
} else if trimmed.contains("maximum") && trimmed.contains("properties required") {
|
||||
("MAX_PROPERTIES_VIOLATED".to_string(),
|
||||
"Object has more properties than allowed".to_string())
|
||||
} else if trimmed.contains("minimum") && trimmed.contains("items required") {
|
||||
("MIN_ITEMS_VIOLATED".to_string(),
|
||||
"Array has fewer items than required".to_string())
|
||||
} else if trimmed.contains("maximum") && trimmed.contains("items required") {
|
||||
("MAX_ITEMS_VIOLATED".to_string(),
|
||||
"Array has more items than allowed".to_string())
|
||||
} else if trimmed == "false schema" {
|
||||
("FALSE_SCHEMA".to_string(),
|
||||
"Schema validation always fails".to_string())
|
||||
} else if trimmed == "not failed" {
|
||||
("NOT_VIOLATED".to_string(),
|
||||
"Value matched a schema it should not match".to_string())
|
||||
} else if trimmed == "allOf failed" {
|
||||
("ALL_OF_VIOLATED".to_string(),
|
||||
"Value does not match all required schemas".to_string())
|
||||
} else if trimmed == "anyOf failed" {
|
||||
("ANY_OF_VIOLATED".to_string(),
|
||||
"Value does not match any of the allowed schemas".to_string())
|
||||
} else if trimmed.contains("oneOf failed") {
|
||||
("ONE_OF_VIOLATED".to_string(),
|
||||
"Value must match exactly one schema".to_string())
|
||||
} else if trimmed == "validation failed" {
|
||||
("VALIDATION_FAILED".to_string(),
|
||||
"Validation failed".to_string())
|
||||
} else {
|
||||
// For any unmatched patterns, try to provide a generic human-readable message
|
||||
// while preserving the original error in details.cause
|
||||
("VALIDATION_FAILED".to_string(),
|
||||
"Validation failed".to_string())
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
// Extract a better format message
|
||||
fn extract_format_message(message: &str) -> String {
|
||||
if message.contains("date-time") {
|
||||
"Invalid date-time format".to_string()
|
||||
} else if message.contains("email") {
|
||||
"Invalid email format".to_string()
|
||||
} else if message.contains("uri") {
|
||||
"Invalid URI format".to_string()
|
||||
} else if message.contains("uuid") {
|
||||
"Invalid UUID format".to_string()
|
||||
} else {
|
||||
"Invalid format".to_string()
|
||||
}
|
||||
current.clone()
|
||||
}
|
||||
|
||||
#[pg_extern(strict, parallel_safe)]
|
||||
@ -347,7 +424,6 @@ pub mod pg_test {
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
#[cfg(any(test, feature = "pg_test"))]
|
||||
#[pg_schema]
|
||||
mod tests {
|
||||
|
||||
201
src/tests.rs
201
src/tests.rs
@ -143,7 +143,7 @@ fn test_cache_and_validate_json_schema() {
|
||||
let invalid_instance_type = json!({ "name": "Bob", "age": -5 });
|
||||
let invalid_instance_missing = json!({ "name": "Charlie" });
|
||||
|
||||
let cache_result = cache_json_schema(schema_id, jsonb(schema.clone()));
|
||||
let cache_result = cache_json_schema(schema_id, jsonb(schema.clone()), false);
|
||||
assert_success_with_json!(cache_result, "Cache operation should succeed.");
|
||||
|
||||
let valid_result = validate_json_schema(schema_id, jsonb(valid_instance));
|
||||
@ -153,16 +153,15 @@ fn test_cache_and_validate_json_schema() {
|
||||
let invalid_result_type = validate_json_schema(schema_id, jsonb(invalid_instance_type));
|
||||
assert_failure_with_json!(invalid_result_type, 1, "Value is below the minimum allowed", "Validation with invalid type should fail.");
|
||||
let errors_type = invalid_result_type.0["errors"].as_array().unwrap();
|
||||
assert_eq!(errors_type[0]["details"]["context"]["instance_path"], "/age");
|
||||
assert_eq!(errors_type[0]["details"]["path"], "urn:my_schema#/properties/age");
|
||||
assert_eq!(errors_type[0]["details"]["path"], "/age");
|
||||
assert_eq!(errors_type[0]["details"]["context"], -5);
|
||||
assert_eq!(errors_type[0]["code"], "MINIMUM_VIOLATED");
|
||||
|
||||
// Missing field
|
||||
let invalid_result_missing = validate_json_schema(schema_id, jsonb(invalid_instance_missing));
|
||||
assert_failure_with_json!(invalid_result_missing, 1, "Required field is missing", "Validation with missing field should fail.");
|
||||
let errors_missing = invalid_result_missing.0["errors"].as_array().unwrap();
|
||||
assert_eq!(errors_missing[0]["details"]["context"]["instance_path"], "");
|
||||
assert_eq!(errors_missing[0]["details"]["path"], "urn:my_schema#");
|
||||
assert_eq!(errors_missing[0]["details"]["path"], "");
|
||||
assert_eq!(errors_missing[0]["code"], "REQUIRED_FIELD_MISSING");
|
||||
|
||||
// Schema not found
|
||||
@ -191,7 +190,7 @@ fn test_cache_invalid_json_schema() {
|
||||
"type": ["invalid_type_value"]
|
||||
});
|
||||
|
||||
let cache_result = cache_json_schema(schema_id, jsonb(invalid_schema));
|
||||
let cache_result = cache_json_schema(schema_id, jsonb(invalid_schema), false);
|
||||
|
||||
// Expect 2 leaf errors because the meta-schema validation fails at the type value
|
||||
// and within the type array itself.
|
||||
@ -208,9 +207,9 @@ fn test_cache_invalid_json_schema() {
|
||||
// Both errors should have ENUM_VIOLATED code
|
||||
assert_eq!(errors_array[0]["code"], "ENUM_VIOLATED");
|
||||
assert_eq!(errors_array[1]["code"], "ENUM_VIOLATED");
|
||||
// Check instance paths are preserved in context
|
||||
// Check instance paths are preserved in path field
|
||||
let paths: Vec<&str> = errors_array.iter()
|
||||
.map(|e| e["details"]["context"]["instance_path"].as_str().unwrap())
|
||||
.map(|e| e["details"]["path"].as_str().unwrap())
|
||||
.collect();
|
||||
assert!(paths.contains(&"/type"));
|
||||
assert!(paths.contains(&"/type/0"));
|
||||
@ -234,7 +233,7 @@ fn test_validate_json_schema_detailed_validation_errors() {
|
||||
},
|
||||
"required": ["address"]
|
||||
});
|
||||
let _ = cache_json_schema(schema_id, jsonb(schema));
|
||||
let _ = cache_json_schema(schema_id, jsonb(schema), false);
|
||||
|
||||
let invalid_instance = json!({
|
||||
"address": {
|
||||
@ -273,7 +272,7 @@ fn test_validate_json_schema_oneof_validation_errors() {
|
||||
]
|
||||
});
|
||||
|
||||
let _ = cache_json_schema(schema_id, jsonb(schema));
|
||||
let _ = cache_json_schema(schema_id, jsonb(schema), false);
|
||||
|
||||
// --- Test case 1: Fails string maxLength (in branch 0) AND missing number_prop (in branch 1) ---
|
||||
let invalid_string_instance = json!({ "string_prop": "toolongstring" });
|
||||
@ -283,11 +282,11 @@ fn test_validate_json_schema_oneof_validation_errors() {
|
||||
// Explicitly check that both expected errors are present, ignoring order
|
||||
let errors_string = result_invalid_string.0["errors"].as_array().expect("Expected error array for invalid string");
|
||||
assert!(errors_string.iter().any(|e|
|
||||
e["details"]["context"]["instance_path"] == "/string_prop" &&
|
||||
e["details"]["path"] == "/string_prop" &&
|
||||
e["code"] == "MAX_LENGTH_VIOLATED"
|
||||
), "Missing maxLength error");
|
||||
assert!(errors_string.iter().any(|e|
|
||||
e["details"]["context"]["instance_path"] == "" &&
|
||||
e["details"]["path"] == "" &&
|
||||
e["code"] == "REQUIRED_FIELD_MISSING"
|
||||
), "Missing number_prop required error");
|
||||
|
||||
@ -299,11 +298,11 @@ fn test_validate_json_schema_oneof_validation_errors() {
|
||||
// Explicitly check that both expected errors are present, ignoring order
|
||||
let errors_number = result_invalid_number.0["errors"].as_array().expect("Expected error array for invalid number");
|
||||
assert!(errors_number.iter().any(|e|
|
||||
e["details"]["context"]["instance_path"] == "/number_prop" &&
|
||||
e["details"]["path"] == "/number_prop" &&
|
||||
e["code"] == "MINIMUM_VIOLATED"
|
||||
), "Missing minimum error");
|
||||
assert!(errors_number.iter().any(|e|
|
||||
e["details"]["context"]["instance_path"] == "" &&
|
||||
e["details"]["path"] == "" &&
|
||||
e["code"] == "REQUIRED_FIELD_MISSING"
|
||||
), "Missing string_prop required error");
|
||||
|
||||
@ -317,7 +316,7 @@ fn test_validate_json_schema_oneof_validation_errors() {
|
||||
let errors_bool = result_invalid_bool.0["errors"].as_array().expect("Expected error array for invalid bool");
|
||||
assert_eq!(errors_bool.len(), 1, "Expected exactly one error after deduplication");
|
||||
assert_eq!(errors_bool[0]["code"], "TYPE_MISMATCH");
|
||||
assert_eq!(errors_bool[0]["details"]["context"]["instance_path"], "");
|
||||
assert_eq!(errors_bool[0]["details"]["path"], "");
|
||||
|
||||
// --- Test case 4: Fails missing required for both branches ---
|
||||
// Input: empty object, expected string_prop (branch 0) OR number_prop (branch 1)
|
||||
@ -329,7 +328,7 @@ fn test_validate_json_schema_oneof_validation_errors() {
|
||||
let errors_empty = result_empty_obj.0["errors"].as_array().expect("Expected error array for empty object");
|
||||
assert_eq!(errors_empty.len(), 1, "Expected exactly one error after filtering empty object");
|
||||
assert_eq!(errors_empty[0]["code"], "REQUIRED_FIELD_MISSING");
|
||||
assert_eq!(errors_empty[0]["details"]["context"]["instance_path"], "");
|
||||
assert_eq!(errors_empty[0]["details"]["path"], "");
|
||||
// The human message should be generic
|
||||
assert_eq!(errors_empty[0]["message"], "Required field is missing");
|
||||
}
|
||||
@ -341,7 +340,7 @@ fn test_clear_json_schemas() {
|
||||
|
||||
let schema_id = "schema_to_clear";
|
||||
let schema = json!({ "type": "string" });
|
||||
let cache_result = cache_json_schema(schema_id, jsonb(schema.clone()));
|
||||
let cache_result = cache_json_schema(schema_id, jsonb(schema.clone()), false);
|
||||
assert_success_with_json!(cache_result);
|
||||
|
||||
let show_result1 = show_json_schemas();
|
||||
@ -367,8 +366,8 @@ fn test_show_json_schemas() {
|
||||
let schema_id2 = "schema2";
|
||||
let schema = json!({ "type": "boolean" });
|
||||
|
||||
let _ = cache_json_schema(schema_id1, jsonb(schema.clone()));
|
||||
let _ = cache_json_schema(schema_id2, jsonb(schema.clone()));
|
||||
let _ = cache_json_schema(schema_id1, jsonb(schema.clone()), false);
|
||||
let _ = cache_json_schema(schema_id2, jsonb(schema.clone()), false);
|
||||
|
||||
let result = show_json_schemas();
|
||||
let schemas = result.0["response"].as_array().unwrap();
|
||||
@ -376,3 +375,167 @@ fn test_show_json_schemas() {
|
||||
assert!(schemas.contains(&json!(schema_id1)));
|
||||
assert!(schemas.contains(&json!(schema_id2)));
|
||||
}
|
||||
|
||||
#[pg_test]
|
||||
fn test_auto_strict_validation() {
|
||||
clear_json_schemas();
|
||||
let schema_id = "strict_test";
|
||||
let schema_id_non_strict = "non_strict_test";
|
||||
|
||||
// Schema without explicit additionalProperties: false
|
||||
let schema = json!({
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"name": { "type": "string" },
|
||||
"profile": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"age": { "type": "number" },
|
||||
"preferences": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"theme": { "type": "string" }
|
||||
}
|
||||
}
|
||||
}
|
||||
},
|
||||
"tags": {
|
||||
"type": "array",
|
||||
"items": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"id": { "type": "string" },
|
||||
"value": { "type": "string" }
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
// Cache the same schema twice - once with strict=true, once with strict=false
|
||||
let cache_result_strict = cache_json_schema(schema_id, jsonb(schema.clone()), true);
|
||||
assert_success_with_json!(cache_result_strict, "Schema caching with strict=true should succeed");
|
||||
|
||||
let cache_result_non_strict = cache_json_schema(schema_id_non_strict, jsonb(schema.clone()), false);
|
||||
assert_success_with_json!(cache_result_non_strict, "Schema caching with strict=false should succeed");
|
||||
|
||||
// Test 1: Valid instance with no extra properties (should pass for both)
|
||||
let valid_instance = json!({
|
||||
"name": "John",
|
||||
"profile": {
|
||||
"age": 30,
|
||||
"preferences": {
|
||||
"theme": "dark"
|
||||
}
|
||||
},
|
||||
"tags": [
|
||||
{"id": "1", "value": "rust"},
|
||||
{"id": "2", "value": "postgres"}
|
||||
]
|
||||
});
|
||||
|
||||
let valid_result_strict = validate_json_schema(schema_id, jsonb(valid_instance.clone()));
|
||||
assert_success_with_json!(valid_result_strict, "Valid instance should pass with strict schema");
|
||||
|
||||
let valid_result_non_strict = validate_json_schema(schema_id_non_strict, jsonb(valid_instance));
|
||||
assert_success_with_json!(valid_result_non_strict, "Valid instance should pass with non-strict schema");
|
||||
|
||||
// Test 2: Root level extra property
|
||||
let invalid_root_extra = json!({
|
||||
"name": "John",
|
||||
"extraField": "should fail" // Extra property at root
|
||||
});
|
||||
|
||||
// Should fail with strict schema
|
||||
let result_root_strict = validate_json_schema(schema_id, jsonb(invalid_root_extra.clone()));
|
||||
assert_failure_with_json!(result_root_strict, 1, "Object contains properties that are not allowed");
|
||||
let errors_root = result_root_strict.0["errors"].as_array().unwrap();
|
||||
assert_eq!(errors_root[0]["code"], "ADDITIONAL_PROPERTIES_NOT_ALLOWED");
|
||||
assert_eq!(errors_root[0]["details"]["path"], "");
|
||||
|
||||
// Should pass with non-strict schema
|
||||
let result_root_non_strict = validate_json_schema(schema_id_non_strict, jsonb(invalid_root_extra));
|
||||
assert_success_with_json!(result_root_non_strict, "Extra property should be allowed with non-strict schema");
|
||||
|
||||
// Test 3: Nested object extra property
|
||||
let invalid_nested_extra = json!({
|
||||
"name": "John",
|
||||
"profile": {
|
||||
"age": 30,
|
||||
"extraNested": "should fail" // Extra property in nested object
|
||||
}
|
||||
});
|
||||
|
||||
// Should fail with strict schema
|
||||
let result_nested_strict = validate_json_schema(schema_id, jsonb(invalid_nested_extra.clone()));
|
||||
assert_failure_with_json!(result_nested_strict, 1, "Object contains properties that are not allowed");
|
||||
let errors_nested = result_nested_strict.0["errors"].as_array().unwrap();
|
||||
assert_eq!(errors_nested[0]["code"], "ADDITIONAL_PROPERTIES_NOT_ALLOWED");
|
||||
assert_eq!(errors_nested[0]["details"]["path"], "/profile");
|
||||
|
||||
// Should pass with non-strict schema
|
||||
let result_nested_non_strict = validate_json_schema(schema_id_non_strict, jsonb(invalid_nested_extra));
|
||||
assert_success_with_json!(result_nested_non_strict, "Extra nested property should be allowed with non-strict schema");
|
||||
|
||||
// Test 4: Deeply nested object extra property
|
||||
let invalid_deep_extra = json!({
|
||||
"name": "John",
|
||||
"profile": {
|
||||
"age": 30,
|
||||
"preferences": {
|
||||
"theme": "dark",
|
||||
"extraDeep": "should fail" // Extra property in deeply nested object
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
// Should fail with strict schema
|
||||
let result_deep_strict = validate_json_schema(schema_id, jsonb(invalid_deep_extra.clone()));
|
||||
assert_failure_with_json!(result_deep_strict, 1, "Object contains properties that are not allowed");
|
||||
let errors_deep = result_deep_strict.0["errors"].as_array().unwrap();
|
||||
assert_eq!(errors_deep[0]["code"], "ADDITIONAL_PROPERTIES_NOT_ALLOWED");
|
||||
assert_eq!(errors_deep[0]["details"]["path"], "/profile/preferences");
|
||||
|
||||
// Should pass with non-strict schema
|
||||
let result_deep_non_strict = validate_json_schema(schema_id_non_strict, jsonb(invalid_deep_extra));
|
||||
assert_success_with_json!(result_deep_non_strict, "Extra deep property should be allowed with non-strict schema");
|
||||
|
||||
// Test 5: Array item extra property
|
||||
let invalid_array_item_extra = json!({
|
||||
"name": "John",
|
||||
"tags": [
|
||||
{"id": "1", "value": "rust", "extraInArray": "should fail"} // Extra property in array item
|
||||
]
|
||||
});
|
||||
|
||||
// Should fail with strict schema
|
||||
let result_array_strict = validate_json_schema(schema_id, jsonb(invalid_array_item_extra.clone()));
|
||||
assert_failure_with_json!(result_array_strict, 1, "Object contains properties that are not allowed");
|
||||
let errors_array = result_array_strict.0["errors"].as_array().unwrap();
|
||||
assert_eq!(errors_array[0]["code"], "ADDITIONAL_PROPERTIES_NOT_ALLOWED");
|
||||
assert_eq!(errors_array[0]["details"]["path"], "/tags/0");
|
||||
|
||||
// Should pass with non-strict schema
|
||||
let result_array_non_strict = validate_json_schema(schema_id_non_strict, jsonb(invalid_array_item_extra));
|
||||
assert_success_with_json!(result_array_non_strict, "Extra array item property should be allowed with non-strict schema");
|
||||
|
||||
// Test 6: Schema with explicit additionalProperties: true should allow extras even with strict=true
|
||||
let schema_id_permissive = "permissive_test";
|
||||
let permissive_schema = json!({
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"name": { "type": "string" }
|
||||
},
|
||||
"additionalProperties": true // Explicitly allow additional properties
|
||||
});
|
||||
|
||||
let _ = cache_json_schema(schema_id_permissive, jsonb(permissive_schema), true); // Note: strict=true
|
||||
|
||||
let instance_with_extra = json!({
|
||||
"name": "John",
|
||||
"extraAllowed": "should pass"
|
||||
});
|
||||
|
||||
let result_permissive = validate_json_schema(schema_id_permissive, jsonb(instance_with_extra));
|
||||
assert_success_with_json!(result_permissive, "Instance with extra property should pass when additionalProperties is explicitly true, even with strict=true");
|
||||
}
|
||||
Reference in New Issue
Block a user