defmodule DataSchema do
@external_resource Path.expand("./README.md")
@moduledoc File.read!(Path.expand("./README.md"))
|> String.split("<!-- README START -->")
|> Enum.at(1)
|> String.split("<!-- README END -->")
|> List.first()
@available_types [:field, :has_one, :has_many, :aggregate, :list_of]
alias DataSchema.CastFunctionError
@doc """
Accepts a the module of a compile time schema and will expand it into a runtime schema
recursively. This can be useful for tooling around generating schemas or for schema
reflection.
"""
def to_runtime_schema(schema) when is_atom(schema) do
if Code.ensure_loaded?(schema) &&
!function_exported?(schema, :__data_schema_fields, 0) do
raise "Provided schema is not a valid DataSchema: #{inspect(schema)}"
end
to_runtime_schema(schema.__data_schema_fields())
end
def to_runtime_schema([_ | _] = fields) do
Enum.reduce(fields, [], fn
{:has_one, {key, path, child_module}}, acc ->
child_schema = to_runtime_schema(child_module)
[{:has_one, {key, path, {child_module, child_schema}}} | acc]
{:has_one, {key, path, child_module, opts}}, acc ->
child_schema = to_runtime_schema(child_module)
[{:has_one, {key, path, {child_module, child_schema}, opts}} | acc]
{:has_many, {key, path, child_module}}, acc ->
child_schema = to_runtime_schema(child_module)
[{:has_many, {key, path, {child_module, child_schema}}} | acc]
{:has_many, {key, path, child_module, opts}}, acc ->
child_schema = to_runtime_schema(child_module)
[{:has_many, {key, path, {child_module, child_schema}, opts}} | acc]
# The aggregate schema already is runtime, but it may include "has_one"s etc inside it.
{:aggregate, {key, nested_fields, cast_fn}}, acc ->
nested_fields = to_runtime_schema(nested_fields)
[{:aggregate, {key, nested_fields, cast_fn}} | acc]
{:aggregate, {key, nested_fields, cast_fn, opts}}, acc ->
nested_fields = to_runtime_schema(nested_fields)
[{:aggregate, {key, nested_fields, cast_fn, opts}} | acc]
field, acc ->
[field | acc]
end)
|> :lists.reverse()
end
@doc """
A macro that creates a data schema. By default all struct fields are required but you
can specify that a field be optional by passing the correct option in. See the Options
section below for more.
### Field Types
There are 5 kinds of struct fields we can have:
1. `field` - The value will be a casted value from the source data.
2. `list_of` - The value will be a list of casted values created from the source data.
3. `has_one` - The value will be created from a nested data schema (so will be a struct)
4. `has_many` - The value will be created by casting a list of values into a data schema.
(You end up with a list of structs defined by the provided schema). Similar to has_many in ecto
5. `aggregate` - The value will a casted value formed from multiple bits of data in the source.
### Options
Available options are:
- `:optional?` - specifies whether or not the field in the struct should be included in
the `@enforce_keys` for the struct. By default all fields are required but you can mark
them as optional by setting this to `true`. This will also be checked when creating a
struct with `DataSchema.to_struct/2` returning an error if the required field is null.
- `:empty_values` - allows you to define what values should be used as "empty" for a
given field. If either the value returned from the data accessor or the casted value are
equivalent to any element in this list, that field is deemed to be empty. Defaults to `[nil]`.
- `:default` - specifies a 0 arity function that will be called to produce a default value for a field
when casting. This function will only be called if a field is found to be empty AND optional.
If it's empty and not optional we will error.
For example:
defmodule Sandwich do
require DataSchema
DataSchema.data_schema([
field: {:type, "the_type", &{:ok, String.upcase(&1)}, optional?: true},
])
end
And:
defmodule Sandwich do
require DataSchema
DataSchema.data_schema([
field: {:list, "list", &{:ok, &1}, optional?: true, empty_values: [[]]},
])
end
And:
defmodule Sandwich do
require DataSchema
@options [optional?: true, empty_values: [nil], default: &DateTime.utc_now/0]
DataSchema.data_schema([
field: {:created_at, "inserted_at", &{:ok, &1}, @options},
])
end
### Examples
See the guides for more in depth examples but below you can see how we create a schema
that will take a map of data and create a struct out of it. Given the following schema:
defmodule Sandwich do
require DataSchema
DataSchema.data_schema([
field: {:type, "the_type", &{:ok, String.upcase.(&1)}},
list_of: {:fillings, "the_fillings", &({:ok, String.downcase(&1["name"])})}
])
end
input_data = %{
"the_type" => "fake steak",
"the_fillings" => [
%{"name" => "fake stake", "good?" => true},
%{"name" => "SAUCE"},
%{"name" => "sweetcorn"},
]
}
DataSchema.to_struct(input_data, Sandwich)
# outputs the following:
%Sandwich{
type: "FAKE STEAK",
fillings: ["fake stake", "sauce", "sweetcorn"],
}
"""
# TODO - add a validate options thing that checks for names of options and errors
# if a field option is invalid. This would be a handy compile time error.
defmacro data_schema(fields) do
quote do
@doc false
def __data_schema_fields, do: unquote(fields)
# __MODULE__ refers to the module that this macro is used in - ie the schema module.
# We add a __data_accessor function so to_struct can call it, we default to a Map
# accessor if no accessor is provided.
if Module.has_attribute?(__MODULE__, :data_accessor) do
@doc false
def __data_accessor do
@data_accessor
end
else
@doc false
def __data_accessor, do: DataSchema.MapAccessor
end
@enforce_keys Enum.reduce(
unquote(fields),
[],
fn
# Validates the shape of the fields at compile time.
{type, {_, _, _, _}}, _acc
when type not in unquote(@available_types) ->
message = """
Field #{inspect(type)} is not a valid field type.
Check the docs in DataSchema for more information on how fields should be written.
The available types are: #{inspect(unquote(@available_types))}
"""
raise DataSchema.InvalidSchemaError, message: message
{type, {_, _, _}}, _acc
when type not in unquote(@available_types) ->
message = """
Field #{inspect(type)} is not a valid field type.
Check the docs in DataSchema for more information on how fields should be written.
The available types are: #{inspect(unquote(@available_types))}
"""
raise DataSchema.InvalidSchemaError, message: message
{:aggregate, {_, schema, _, _}}, _acc
when not is_atom(schema) and not is_list(schema) ->
raise DataSchema.InvalidSchemaError, """
An :aggregate field should provide a nested schema to describe the data to be extracted.
This can be a module of another DataSchema or a list of schema fields:
defmodule Thing do
import DataSchema, only: [data_schema: 1]
@fields [
field: {:date, "date", &Date.from_iso8601/1},
field: {:time, "time", &Time.from_iso8601/1}
]
data_schema([
aggregate: {:datetime, @fields, NaiveDateTime.new(&1.date, &1.time)}
])
end
Or:
defmodule Thing do
import DataSchema, only: [data_schema: 1]
defmodule DateTime do
import DataSchema, only: [data_schema: 1]
data_schema([
field: {:date, "date", &Date.from_iso8601/1},
field: {:time, "time", &Time.from_iso8601/1}
])
end
data_schema([
aggregate: {:datetime, DateTime, &NaiveDateTime.new(&1.date, &1.time)}
])
end
Provided schema: #{inspect(schema)}
"""
{:aggregate, {_, schema, _}}, _acc
when not is_atom(schema) and not is_list(schema) ->
raise DataSchema.InvalidSchemaError, """
An :aggregate field should provide a nested schema to describe the data to be extracted.
This can be a module of another DataSchema or a list of schema fields:
defmodule Thing do
import DataSchema, only: [data_schema: 1]
@fields [
field: {:date, "date", &Date.from_iso8601/1},
field: {:time, "time", &Time.from_iso8601/1}
]
data_schema([
aggregate: {:datetime, @fields, NaiveDateTime.new(&1.date, &1.time)}
])
end
Or:
defmodule Thing do
import DataSchema, only: [data_schema: 1]
defmodule DateTime do
import DataSchema, only: [data_schema: 1]
data_schema([
field: {:date, "date", &Date.from_iso8601/1},
field: {:time, "time", &Time.from_iso8601/1}
])
end
data_schema([
aggregate: {:datetime, DateTime, &NaiveDateTime.new(&1.date, &1.time)}
])
end
Provided schema: #{inspect(schema)}
"""
{type, {_, _, {module, _}, _opts}}, _acc
when type in [:has_one, :has_many] and not is_atom(module) and
not is_map(module) ->
message = """
#{type} fields require a DataSchema module as their casting function:
data_schema([
#{type}: {:foo, "path", Foo}
# ^^
# Should be a DataSchema module
])
Or an inline list of fields like so:
@foo_fields [
field: {:bar, "bar", &{:ok, to_string(&1)}}
]
data_schema([
#{type}: {:foo, "path", {%{}, @foo_fields}}
])
Or for an inline struct:
@foo_fields [
field: {:bar, "bar", &{:ok, to_string(&1)}}
]
data_schema([
#{type}: {:foo, "path", {SomeStructModule, @foo_fields}}
])
You provided the following as a schema: #{inspect(module)}.
Ensure you haven't used the wrong field type.
"""
raise DataSchema.InvalidSchemaError, message: message
# the tuple case is handed above so we need to make sure we _dont
# error if the module is a tuple. So a tuple is allowed
{type, {_, _, module, _opts}}, _acc
when type in [:has_one, :has_many] and not is_atom(module) and
not is_tuple(module) ->
message = """
#{type} fields require a DataSchema module as their casting function:
data_schema([
#{type}: {:foo, "path", Foo}
# ^^
# Should be a DataSchema module
])
Or an inline list of fields like so:
@foo_fields [
field: {:bar, "bar", &{:ok, to_string(&1)}}
]
data_schema([
#{type}: {:foo, "path", {%{}, @foo_fields}}
])
Or for an inline struct:
@foo_fields [
field: {:bar, "bar", &{:ok, to_string(&1)}}
]
data_schema([
#{type}: {:foo, "path", {SomeStructModule, @foo_fields}}
])
You provided the following as a schema: #{inspect(module)}.
Ensure you haven't used the wrong field type.
"""
raise DataSchema.InvalidSchemaError, message: message
{type, {_, _, {module, _inline}}}, _acc
when type in [:has_one, :has_many] and not is_atom(module) and
not is_map(module) ->
message = """
#{type} fields require a DataSchema module as their casting function:
data_schema([
#{type}: {:foo, "path", Foo}
# ^^
# Should be a DataSchema module
])
Or an inline list of fields like so:
@foo_fields [
field: {:bar, "bar", &{:ok, to_string(&1)}}
]
data_schema([
#{type}: {:foo, "path", {%{}, @foo_fields}}
])
Or for an inline struct:
@foo_fields [
field: {:bar, "bar", &{:ok, to_string(&1)}}
]
data_schema([
#{type}: {:foo, "path", {SomeStructModule, @foo_fields}}
])
You provided the following as a schema: #{inspect(module)}.
Ensure you haven't used the wrong field type.
"""
raise DataSchema.InvalidSchemaError, message: message
# the tuple case is handed above so we need to make sure we _dont
# error if the module is a tuple. So a tuple is allowed
{type, {_, _, module}}, _acc
when type in [:has_one, :has_many] and not is_atom(module) and
not is_tuple(module) ->
message = """
#{type} fields require a DataSchema module as their casting function:
data_schema([
#{type}: {:foo, "path", Foo}
# ^^
# Should be a DataSchema module
])
Or an inline list of fields like so:
@foo_fields [
field: {:bar, "bar", &{:ok, to_string(&1)}}
]
data_schema([
#{type}: {:foo, "path", {%{}, @foo_fields}}
])
Or for an inline struct:
@foo_fields [
field: {:bar, "bar", &{:ok, to_string(&1)}}
]
data_schema([
#{type}: {:foo, "path", {SomeStructModule, @foo_fields}}
])
You provided the following as a schema: #{inspect(module)}.
Ensure you haven't used the wrong field type.
"""
raise DataSchema.InvalidSchemaError, message: message
{_, {field, _, _, opts}}, acc ->
# By default fields are required but they can be marked as optional.
if Keyword.get(opts, :optional?, false) do
acc
else
[field | acc]
end
# If no options are provided the field is enforced.
{_, {field, _, _}}, acc ->
[field | acc]
_, _ ->
raise DataSchema.InvalidSchemaError
end
)
defstruct Enum.map(unquote(fields), fn
{_, {field, _, _}} when not is_atom(field) ->
message = """
The provided struct keys must be atoms. See docs for more information:
data_schema([
field: {:foo, "foo", &{:ok, &1}}
# ^^^
# must be an atom!
])
"""
raise DataSchema.InvalidSchemaError, message: message
{_, {field, _, _, _}} when not is_atom(field) ->
message = """
The provided struct keys must be atoms. See docs for more information:
data_schema([
field: {:foo, "foo", &{:ok, &1}}
# ^^^
# must be an atom!
])
"""
raise DataSchema.InvalidSchemaError, message: message
{_, {field, _, _}} ->
field
{_, {field, _, _, _}} ->
field
_ ->
raise DataSchema.InvalidSchemaError
end)
end
end
@doc """
Accepts an data schema module and some source data and attempts to create the struct
defined in the schema from the source data recursively.
We essentially visit each field in the schema and extract the data the field points to
from the source data, passing it to the field's casting function before setting the
result of that as the value on the struct.
This function takes a simple approach to creating the struct - whatever you return from
a casting function will be set as the value of the struct field. You should raise if
you want casting to fail.
### Examples
data = %{ "spice" => "enables space travel" }
defmodule Foo do
require DataSchema
DataSchema.data_schema(
field: {:a_rocket, "spice", &({:ok, &1})}
)
end
DataSchema.to_struct(data, Foo)
# => Outputs the following:
%Foo{a_rocket: "enables space travel"}
"""
def to_struct(data, schema) do
do_to_struct(data, schema, [])
|> unwrap()
end
def to_struct(data, %schema{} = struct, opts) do
if is_compile_time_schema?(schema) do
fields = schema.__data_schema_fields()
accessor = schema.__data_accessor()
do_to_struct(data, struct, fields, accessor, opts)
|> unwrap()
else
raise "Provided schema is not a valid DataSchema: #{inspect(schema)}"
end
end
def to_struct(data, schema, opts) when is_atom(schema) do
if is_compile_time_schema?(schema) do
fields = schema.__data_schema_fields()
accessor = schema.__data_accessor()
struct = struct(schema, %{})
do_to_struct(data, struct, fields, accessor, opts)
|> unwrap()
else
raise "Provided schema is not a valid DataSchema: #{inspect(schema)}"
end
end
def to_struct(data, struct, fields, accessor) do
do_to_struct(data, struct, fields, accessor, [])
|> unwrap()
end
defp do_to_struct(data, schema) do
do_to_struct(data, schema, [])
end
defp do_to_struct(data, %schema{} = struct, opts) do
if is_compile_time_schema?(schema) do
fields = schema.__data_schema_fields()
accessor = schema.__data_accessor()
do_to_struct(data, struct, fields, accessor, opts)
else
raise "Provided schema is not a valid DataSchema: #{inspect(schema)}"
end
end
defp do_to_struct(data, schema, opts) when is_atom(schema) do
if is_compile_time_schema?(schema) do
fields = schema.__data_schema_fields()
accessor = schema.__data_accessor()
struct = struct(schema, %{})
do_to_struct(data, struct, fields, accessor, opts)
else
raise "Provided schema is not a valid DataSchema: #{inspect(schema)}"
end
end
@doc """
Creates a struct or map from the provided arguments. This function can be used to define
runtime schemas for the most dynamic of cases. This means you don't have to define a schema
at compile time using the `DataShema.data_schema/1` macro.
### Examples
Creating a struct:
defmodule Run do
defstruct [:time]
end
input = %{"time" => "10:00"}
fields = [
field: {:time, "time", &{:ok, to_string(&1)}}
]
DataSchema.to_struct(input, Run, fields, DataSchema.MapAccessor)
{:ok, %Run{time: "10:00"}}
Creating a map:
input = %{"time" => "10:00"}
fields = [
field: {:time, "time", &{:ok, to_string(&1)}}
]
DataSchema.to_struct(input, %{}, fields, DataSchema.MapAccessor)
{:ok, %{time: "10:00"}}
"""
# If we are passed a Module we assume it's a struct and create an empty one to reduce over.
def to_struct(data, struct, fields, accessor, opts) when is_atom(struct) do
do_to_struct(data, struct(struct, %{}), fields, accessor, opts)
|> unwrap()
end
def to_struct(data, %{} = struct, fields, accessor, opts) when is_list(fields) do
do_to_struct(data, %{} = struct, fields, accessor, opts)
|> unwrap()
end
defp do_to_struct(data, struct, fields, accessor, opts) when is_atom(struct) do
do_to_struct(data, struct(struct, %{}), fields, accessor, opts)
|> unwrap()
end
defp do_to_struct(data, %{} = struct, fields, accessor, _opts) when is_list(fields) do
default_opts = [optional?: false, empty_values: [nil]]
fields
|> Enum.reduce_while(struct, fn
{:aggregate, {field, schema_mod, cast_fn}} = original_field, struct
when is_atom(schema_mod) ->
fields = schema_mod.__data_schema_fields()
accessor = schema_mod.__data_accessor()
aggregate = struct(schema_mod, %{})
aggregate(
{field, cast_fn, default_opts},
accessor,
data,
aggregate,
struct,
fields,
original_field
)
{:aggregate, {field, schema_mod, cast_fn, field_opts}} = original_field, struct
when is_atom(schema_mod) ->
fields = schema_mod.__data_schema_fields()
accessor = schema_mod.__data_accessor()
aggregate = struct(schema_mod, %{})
field_opts = Keyword.merge(default_opts, field_opts)
aggregate(
{field, cast_fn, field_opts},
accessor,
data,
aggregate,
struct,
fields,
original_field
)
{:aggregate, {field, fields, cast_fn}} = original_field, struct when is_list(fields) ->
aggregate(
{field, cast_fn, default_opts},
accessor,
data,
%{},
struct,
fields,
original_field
)
{:aggregate, {field, fields, cast_fn, field_opts}} = original_field, struct
when is_list(fields) ->
# User supplied take precedence
field_opts = Keyword.merge(default_opts, field_opts)
aggregate(
{field, cast_fn, field_opts},
accessor,
data,
%{},
struct,
fields,
original_field
)
{type, {field, path, cast, opts}} = schema_field, struct ->
opts = Keyword.merge(default_opts, opts)
process_field({type, {field, path, cast, opts}}, struct, accessor, data, schema_field)
{type, {field, path, cast}} = schema_field, struct ->
process_field(
{type, {field, path, cast, default_opts}},
struct,
accessor,
data,
schema_field
)
end)
|> case do
{:error, error_message} ->
{:error, error_message}
struct_or_map ->
{:ok, struct_or_map}
end
end
defp process_field(
{:field, {field, path, cast_fn, opts}},
struct,
accessor,
data,
original_field
) do
value = accessor.field(data, path)
empty_values = Keyword.get(opts, :empty_values)
optional? = Keyword.get(opts, :optional?)
default = Keyword.get(opts, :default, :no_default)
do_cast = &call_cast_fn(cast_fn, &1, original_field)
case cast_and_validate(value, do_cast, empty_values, optional?, default) do
{:ok, value} ->
{:cont, update_struct(struct, field, value)}
{:error, :empty_required_value} ->
{:halt, {:error, DataSchema.Errors.empty_required_value_error(field)}}
{:error, %CastFunctionError{} = cast_error} ->
{:halt, {:error, add_field_to_error(cast_error, field, struct)}}
{:error, error} ->
{:halt, {:error, DataSchema.Errors.new({field, error})}}
:error ->
{:halt, {:error, DataSchema.Errors.default_error(field)}}
other_value ->
raise_incorrect_cast_function_error(field, other_value)
end
end
defp process_field(
{:has_one, {field, path, {cast_module, inline_fields}, opts}},
struct,
accessor,
data,
_original_field
) do
value = accessor.has_one(data, path)
empty_values = Keyword.get(opts, :empty_values)
optional? = Keyword.get(opts, :optional?)
default = Keyword.get(opts, :default, :no_default)
do_cast = &do_to_struct(&1, cast_module, inline_fields, accessor, [])
process_has_one(field, value, do_cast, struct, empty_values, optional?, default)
end
defp process_field(
{:has_one, {field, path, cast_module, opts}},
struct,
accessor,
data,
_original_field
) do
value = accessor.has_one(data, path)
empty_values = Keyword.get(opts, :empty_values)
optional? = Keyword.get(opts, :optional?)
default = Keyword.get(opts, :default, :no_default)
do_cast = &do_to_struct(&1, cast_module)
process_has_one(field, value, do_cast, struct, empty_values, optional?, default)
end
defp process_field(
{:has_many, {field, path, {cast_module, inline_fields}, opts}},
struct,
accessor,
data,
_original_field
) do
values = accessor.has_many(data, path)
empty_values = Keyword.get(opts, :empty_values)
optional? = Keyword.get(opts, :optional?)
default = Keyword.get(opts, :default, :no_default)
do_cast =
&Enum.reduce_while(&1, {:ok, []}, fn value, {:ok, acc} ->
case do_to_struct(value, cast_module, inline_fields, accessor, []) do
{:ok, struct} ->
{:cont, {:ok, [struct | acc]}}
{:error, _} = error ->
{:halt, error}
end
end)
process_has_many(field, values, do_cast, struct, empty_values, optional?, default)
end
defp process_field(
{:has_many, {field, path, cast_module, opts}},
struct,
accessor,
data,
_original_field
) do
values = accessor.has_many(data, path)
empty_values = Keyword.get(opts, :empty_values)
optional? = Keyword.get(opts, :optional?)
default = Keyword.get(opts, :default, :no_default)
# TODO: Enum.with_index so that we can identify the precise problematic element
# when it errors? Cost paid for when there are no errors. But is nice to be able to
# pin point the exact element that went wrong.
do_cast =
&Enum.reduce_while(&1, {:ok, []}, fn value, {:ok, acc} ->
case do_to_struct(value, cast_module) do
{:ok, struct} ->
{:cont, {:ok, [struct | acc]}}
{:error, _} = error ->
{:halt, error}
end
end)
process_has_many(field, values, do_cast, struct, empty_values, optional?, default)
end
defp process_field(
{:list_of, {field, path, cast_fn, opts}},
struct,
accessor,
data,
original_field
) do
values = accessor.list_of(data, path)
empty_values = Keyword.get(opts, :empty_values)
optional? = Keyword.get(opts, :optional?)
default = Keyword.get(opts, :default, :no_default)
do_cast =
&Enum.reduce_while(&1, {:ok, []}, fn value, {:ok, acc} ->
cast_value = fn val -> call_cast_fn(cast_fn, val, original_field) end
case cast_and_validate(value, cast_value, empty_values, optional?, default) do
{:ok, value} ->
{:cont, {:ok, [value | acc]}}
{:error, :empty_required_value} ->
{:halt, {:error, DataSchema.Errors.empty_required_value_error(field)}}
{:error, %CastFunctionError{} = cast_error} ->
{:halt, {:error, add_field_to_error(cast_error, field, struct)}}
{:error, error} ->
{:halt, {:error, DataSchema.Errors.new({field, error})}}
:error ->
{:halt, {:error, DataSchema.Errors.default_error(field)}}
other_value ->
raise_incorrect_cast_function_error(field, other_value)
end
end)
# More testing needed here!!
case cast_and_validate(values, do_cast, empty_values, optional?, default) do
{:ok, list} when is_list(list) ->
{:cont, update_struct(struct, field, :lists.reverse(list))}
{:ok, empty_value} ->
{:cont, update_struct(struct, field, empty_value)}
{:error, :empty_required_value} ->
{:halt, {:error, DataSchema.Errors.empty_required_value_error(field)}}
{:error, error} ->
{:halt, {:error, error}}
end
end
defp process_has_many(field, values, do_cast, struct, empty_values, optional?, default_value) do
case cast_and_validate(values, do_cast, empty_values, optional?, default_value) do
{:ok, list} when is_list(list) ->
{:cont, update_struct(struct, field, :lists.reverse(list))}
{:ok, empty_or_default_value} ->
{:cont, update_struct(struct, field, empty_or_default_value)}
{:error, :empty_required_value} ->
{:halt, {:error, DataSchema.Errors.empty_required_value_error(field)}}
{:error, %CastFunctionError{} = cast_error} ->
{:halt, {:error, add_field_to_error(cast_error, field, struct)}}
{:error, error} ->
{:halt, {:error, DataSchema.Errors.new({field, error})}}
:error ->
{:halt, {:error, DataSchema.Errors.default_error(field)}}
other_value ->
raise_incorrect_cast_function_error(field, other_value)
end
end
defp process_has_one(field, value, do_cast, struct, empty_values, optional?, default) do
case cast_and_validate(value, do_cast, empty_values, optional?, default) do
{:ok, casted} ->
{:cont, update_struct(struct, field, casted)}
{:error, :empty_required_value} ->
{:halt, {:error, DataSchema.Errors.empty_required_value_error(field)}}
{:error, %CastFunctionError{} = cast_error} ->
{:halt, {:error, add_field_to_error(cast_error, field, struct)}}
{:error, error} ->
{:halt, {:error, DataSchema.Errors.new({field, error})}}
:error ->
{:halt, {:error, DataSchema.Errors.default_error(field)}}
other_value ->
raise_incorrect_cast_function_error(field, other_value)
end
end
defp aggregate(
{field, cast_fn, field_opts},
accessor,
data,
aggregate,
parent,
fields,
original_field
) do
empty_values = Keyword.get(field_opts, :empty_values)
optional? = Keyword.get(field_opts, :optional?)
default = Keyword.get(field_opts, :default, :no_default)
do_cast = &call_cast_fn(cast_fn, &1, original_field)
case do_to_struct(data, aggregate, fields, accessor, []) do
{:error, %CastFunctionError{} = cast_error} ->
{:halt, {:error, add_field_to_error(cast_error, field, parent)}}
{:error, %DataSchema.Errors{} = error} ->
{:halt, {:error, DataSchema.Errors.new({field, error})}}
{:ok, values_map} ->
case cast_and_validate(values_map, do_cast, empty_values, optional?, default) do
{:ok, casted_value} ->
{:cont, update_struct(parent, field, casted_value)}
{:error, :empty_required_value} ->
{:halt, {:error, DataSchema.Errors.empty_required_value_error(field)}}
{:error, %CastFunctionError{} = cast_error} ->
{:halt, {:error, add_field_to_error(cast_error, field, parent)}}
{:error, error} ->
{:halt, {:error, DataSchema.Errors.new({field, error})}}
:error ->
{:halt, {:error, DataSchema.Errors.default_error(field)}}
other_value ->
raise_incorrect_cast_function_error(field, other_value)
end
end
end
defp cast_and_validate(value, do_cast, empty_values, optional?, default_value) do
with {:empty?, false} <- {:empty?, value in empty_values},
{:ok, casted_value} <- do_cast.(value),
{:empty?, _, false} <- {:empty?, casted_value, casted_value in empty_values} do
{:ok, casted_value}
else
{:empty?, casted_value, true} ->
if optional? do
if default_value == :no_default do
{:ok, casted_value}
else
{:ok, default_value.()}
end
else
{:error, :empty_required_value}
end
{:empty?, true} ->
if optional? do
if default_value == :no_default do
{:ok, value}
else
{:ok, default_value.()}
end
else
{:error, :empty_required_value}
end
error ->
error
end
end
defp raise_incorrect_cast_function_error(field, value) do
message = """
Casting error for field #{field}, cast function should return one of the following:
{:ok, any()} | :error | {:error, any()}
Cast function returned #{inspect(value)}
"""
raise DataSchema.InvalidCastFunction, message: message
end
def unwrap({:error, %CastFunctionError{} = cast_error}) do
message = CastFunctionError.error_message(cast_error)
reraise CastFunctionError,
[message: message, wrapped_error: cast_error.wrapped_error],
cast_error.stacktrace_of_wrapped_error
end
def unwrap(value), do: value
defp is_compile_time_schema?(schema) when is_atom(schema) do
Code.ensure_loaded?(schema) && function_exported?(schema, :__data_schema_fields, 0)
end
defp is_compile_time_schema?(%schema{}) do
is_compile_time_schema?(schema)
end
defp is_compile_time_schema?(%{}) do
false
end
defp add_field_to_error(cast_error, field, %schema{}) do
if is_compile_time_schema?(schema) do
%{cast_error | path: [{schema, field} | cast_error.path]}
else
%{cast_error | path: [field | cast_error.path]}
end
end
defp add_field_to_error(cast_error, field, _) do
%{cast_error | path: [field | cast_error.path]}
end
# Sometimes the data we are creating is a map, sometimes a struct. When it is a struct
# we want to know the field exists before we add it.
defp update_struct(%_struct_name{} = struct, field, item) do
%{struct | field => item}
end
defp update_struct(%{} = map, field, item) do
Map.put(map, field, item)
end
defp call_cast_fn({module, fun, args}, value, original_field) do
try do
apply(module, fun, [value | args])
rescue
error ->
cast_error = %CastFunctionError{
casted_value: value,
path: [],
leaf_field: original_field,
wrapped_error: error,
stacktrace_of_wrapped_error: __STACKTRACE__
}
{:error, cast_error}
end
end
defp call_cast_fn(module, value, original_field) when is_atom(module) do
try do
module.cast(value)
rescue
error ->
cast_error = %CastFunctionError{
casted_value: value,
path: [],
leaf_field: original_field,
wrapped_error: error,
stacktrace_of_wrapped_error: __STACKTRACE__
}
{:error, cast_error}
end
end
defp call_cast_fn(fun, value, original_field) when is_function(fun, 1) do
try do
fun.(value)
rescue
error ->
cast_error = %CastFunctionError{
casted_value: value,
path: [],
leaf_field: original_field,
wrapped_error: error,
stacktrace_of_wrapped_error: __STACKTRACE__
}
{:error, cast_error}
end
end
@doc """
A private function that aims to return a flat list of all of the absolute paths
in a given schema. This only works if the path is a list though, as would be the case
if you had an Access data accessor for example. This is useful groundwork for the same
idea with other paths though - you would just have to modify it slightly I think to get
what you want.
""" && false
def absolute_paths_for_schema(schema) when is_atom(schema) do
if Code.ensure_loaded?(schema) &&
!function_exported?(schema, :__data_schema_fields, 0) do
raise "Provided schema is not a valid DataSchema: #{inspect(schema)}"
end
schema
|> to_runtime_schema()
|> absolute_paths_for_schema()
end
def absolute_paths_for_schema(runtime_schema) when is_list(runtime_schema) do
absolute_paths_for_schema(runtime_schema, [])
end
def absolute_paths_for_schema(runtime_schema, acc) when is_list(runtime_schema) do
runtime_schema
|> Enum.reduce(acc, fn
{:has_many, {_key, path, {_child_schema, child_fields}, _opts}}, acc ->
child_fields
|> absolute_paths_for_schema([])
|> Enum.reduce(acc, fn
{p, modifier}, accu ->
[{path ++ p, modifier} | accu]
end)
{:has_many, {_key, path, {_child_schema, child_fields}}}, acc ->
child_fields
|> absolute_paths_for_schema([])
|> Enum.reduce(acc, fn
{p, modifier}, accu ->
[{path ++ p, modifier} | accu]
end)
{:has_one, {_key, path, {_child_schema, child_fields}, _opts}}, acc ->
child_fields
|> absolute_paths_for_schema([])
|> Enum.reduce(acc, fn
{p, modifier}, accu ->
[{path ++ p, modifier} | accu]
end)
{:has_one, {_key, path, {_child_schema, child_fields}}}, acc ->
child_fields
|> absolute_paths_for_schema([])
|> Enum.reduce(acc, fn
{p, modifier}, accu ->
[{path ++ p, modifier} | accu]
end)
{:aggregate, {_key, child_fields, _cast_fn}}, acc ->
absolute_paths_for_schema(child_fields, acc)
{:aggregate, {_key, child_fields, _cast_fn, _opts}}, acc ->
absolute_paths_for_schema(child_fields, acc)
{_field, {_key, path, _cast_fn}}, acc ->
[path | acc]
{_field, {_key, path, _cast_fn, _opts}}, acc ->
[path | acc]
end)
end
end