defmodule Explorer.DataFrame do
@moduledoc """
The DataFrame struct and API.
Dataframes are two-dimensional tabular data structures similar to a spreadsheet.
For example, the Iris dataset:
iex> Explorer.Datasets.iris()
#Explorer.DataFrame<
Polars[150 x 5]
sepal_length float [5.1, 4.9, 4.7, 4.6, 5.0, ...]
sepal_width float [3.5, 3.0, 3.2, 3.1, 3.6, ...]
petal_length float [1.4, 1.4, 1.3, 1.5, 1.4, ...]
petal_width float [0.2, 0.2, 0.2, 0.2, 0.2, ...]
species string ["Iris-setosa", "Iris-setosa", "Iris-setosa", "Iris-setosa", "Iris-setosa", ...]
>
This dataframe has 150 rows and five columns. Each column is an `Explorer.Series`
of the same size (150):
iex> df = Explorer.Datasets.iris()
iex> df["sepal_length"]
#Explorer.Series<
float[150]
[5.1, 4.9, 4.7, 4.6, 5.0, 5.4, 4.6, 5.0, 4.4, 4.9, 5.4, 4.8, 4.8, 4.3, 5.8, 5.7, 5.4, 5.1, 5.7, 5.1, 5.4, 5.1, 4.6, 5.1, 4.8, 5.0, 5.0, 5.2, 5.2, 4.7, 4.8, 5.4, 5.2, 5.5, 4.9, 5.0, 5.5, 4.9, 4.4, 5.1, 5.0, 4.5, 4.4, 5.0, 5.1, 4.8, 5.1, 4.6, 5.3, 5.0, ...]
>
## Creating dataframes
Dataframes can be created from normal Elixir objects. The main way you might do this is
with the `new/1` function. For example:
iex> Explorer.DataFrame.new(a: ["a", "b"], b: [1, 2])
#Explorer.DataFrame<
Polars[2 x 2]
a string ["a", "b"]
b integer [1, 2]
>
Or with a list of maps:
iex> Explorer.DataFrame.new([%{"col1" => "a", "col2" => 1}, %{"col1" => "b", "col2" => 2}])
#Explorer.DataFrame<
Polars[2 x 2]
col1 string ["a", "b"]
col2 integer [1, 2]
>
## Verbs
Explorer uses the idea of a consistent set of SQL-like `verbs` like
[`dplyr`](https://dplyr.tidyverse.org) which can help solve common
data manipulation challenges. These are split into single table verbs,
multiple table verbs, and row-based verbs:
### Single table verbs
Single table verbs are (unsurprisingly) used for manipulating a single dataframe.
Those operations typically driven by column names. These are:
- `select/3` for picking variables
- `filter/2` for picking rows based on predicates
- `mutate/2` for adding or replacing columns that are functions of existing columns
- `arrange/2` for changing the ordering of rows
- `distinct/2` for picking unique rows
- `summarise/2` for reducing multiple rows down to a single summary
- `pivot_longer/3` and `pivot_wider/4` for massaging dataframes into longer or wider forms, respectively
Each of these combine with `Explorer.DataFrame.group_by/2` for operating by group.
For more flexibility we also have functions that accept callback functions:
- `filter_with/2`
- `summarise_with/2`
- `mutate_with/2`
- `arrange_with/2`
Those functions work by having a "lazy" representation of the dataframe and series,
which adds the possibility to perform complex operations that are optimized by the backend.
### Multiple table verbs
Multiple table verbs are used for combining tables. These are:
- `join/3` for performing SQL-like joins
- `concat_rows/1` for vertically "stacking" dataframes
- `concat_columns/1` for horizontally "stacking" dataframes
### Row-based verbs
Those operations are driven by the row index. These are:
- `head/2` for picking the first rows
- `tail/2` for picking the last rows
- `slice/2` for slicing the dataframe by row indexes or a range
- `slice/3` for slicing a section by an offset
- `sample/2` for sampling the data-frame by row
## IO
Explorer supports reading and writing of:
- delimited files (such as CSV)
- [Parquet](https://databricks.com/glossary/what-is-parquet)
- [Arrow IPC](https://arrow.apache.org/docs/format/Columnar.html#ipc-file-format)
- [Newline Delimited JSON](http://ndjson.org)
The convention Explorer uses is to have `from_*` and `to_*` functions to read and write
to files in the formats above. `load_*` and `dump_*` versions are also available to read
and write those formats directly in memory.
## Access
In addition to this "grammar" of data manipulation, you'll find useful functions for
slicing and dicing dataframes such as `pull/2`, `head/2`, `sample/3`, `slice/2`, and
`slice/3`.
`Explorer.DataFrame` also implements the `Access` behaviour (also known as the brackets
syntax). This should be familiar for users coming from other language with dataframes
such as R or Python. For example:
iex> df = Explorer.Datasets.wine()
iex> df["class"]
#Explorer.Series<
integer[178]
[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...]
>
"""
alias __MODULE__, as: DataFrame
alias Explorer.Series
alias Explorer.Shared
alias Explorer.Backend.LazySeries
@valid_dtypes Explorer.Shared.dtypes()
@enforce_keys [:data, :groups, :names, :dtypes]
defstruct [:data, :groups, :names, :dtypes]
@typedoc """
Represents a column name as atom or string.
"""
@type column_name :: atom() | String.t()
@typedoc """
Represents a column name or its index.
"""
@type column :: column_name() | non_neg_integer()
@typedoc """
Represents multiple columns.
The columns may be specified as one of:
* a list of columns indexes or names as atoms and strings
* a range
* a one-arity function that receives column names and returns
true for column names to keep
* a two-arity function that receives column names and types and
returns true for column names to keep
"""
@type columns ::
[column]
| Range.t()
| (String.t() -> boolean())
| (String.t(), Explorer.Series.dtype() -> boolean())
@typedoc """
Represents multiple column names as atoms or strings.
"""
@type column_names :: [column_name]
@typedoc """
Represents a column pair where the value is a column name or
a column index, and the value is of type `value`.
"""
@type column_pairs(value) :: [{column(), value}] | %{column() => value}
@typedoc """
Represents a dataframe.
"""
@type t :: %DataFrame{
data: Explorer.Backend.DataFrame.t(),
groups: [String.t()],
names: [String.t()],
dtypes: %{String.t() => Explorer.Series.dtype()}
}
@default_infer_schema_length 1000
@default_sample_nrows 5
# Guards and helpers for columns
defguardp is_column(column) when is_binary(column) or is_atom(column) or is_integer(column)
defguardp is_column_name(column) when is_binary(column) or is_atom(column)
defguardp is_column_pairs(columns) when is_list(columns) or is_map(columns)
# Normalize a column name to string
defp to_column_name(column) when is_binary(column), do: column
defp to_column_name(column) when is_atom(column), do: Atom.to_string(column)
# Normalize pairs of `{column, value}` where value can be anything.
# The `column` is only validated if it's an integer. We check that the index is present.
defp to_column_pairs(df, pairs), do: to_column_pairs(df, pairs, & &1)
# The function allows to change the `value` for each pair.
defp to_column_pairs(df, pairs, value_fun)
when is_column_pairs(pairs) and is_function(value_fun, 1) do
existing_columns = df.names
pairs
|> Enum.map_reduce(nil, fn
{column, value}, maybe_map when is_integer(column) ->
map = maybe_map || column_index_map(existing_columns)
existing_column = fetch_column_at!(map, column)
{{existing_column, value_fun.(value)}, map}
{column, value}, maybe_map when is_atom(column) ->
column = Atom.to_string(column)
{{column, value_fun.(value)}, maybe_map}
{column, value}, maybe_map when is_binary(column) ->
{{column, value_fun.(value)}, maybe_map}
end)
|> then(fn {pairs, _} -> pairs end)
end
defp fetch_column_at!(map, index) do
normalized = if index < 0, do: index + map_size(map), else: index
case map do
%{^normalized => column} -> column
%{} -> raise ArgumentError, "no column exists at index #{index}"
end
end
defp column_index_map(names),
do: for({name, idx} <- Enum.with_index(names), into: %{}, do: {idx, name})
# Normalize column names and raise if column does not exist.
defp to_existing_columns(df, columns) when is_list(columns) do
{columns, _cache} =
Enum.map_reduce(columns, nil, fn
column, maybe_map when is_integer(column) ->
map = maybe_map || column_index_map(df.names)
existing_column = fetch_column_at!(map, column)
{existing_column, map}
column, maybe_map when is_atom(column) ->
column = Atom.to_string(column)
maybe_raise_column_not_found(df, column)
{column, maybe_map}
column, maybe_map when is_binary(column) ->
maybe_raise_column_not_found(df, column)
{column, maybe_map}
end)
columns
end
defp to_existing_columns(%{names: names}, %Range{} = columns) do
Enum.slice(names, columns)
end
defp to_existing_columns(%{names: names}, callback) when is_function(callback, 1) do
Enum.filter(names, callback)
end
defp to_existing_columns(%{names: names, dtypes: dtypes}, callback)
when is_function(callback, 2) do
Enum.filter(names, fn name -> callback.(name, dtypes[name]) end)
end
defp to_existing_columns(_, other) do
raise ArgumentError, """
invalid columns specification. Columns may be specified as one of:
* a list of columns indexes or names as atoms and strings
* a range
* a one-arity function that receives column names and returns
true for column names to keep
* a two-arity function that receives column names and types and
returns true for column names to keep
Got: #{inspect(other)}
"""
end
defp check_dtypes!(dtypes) do
Enum.map(dtypes, fn
{key, value} when is_atom(key) ->
{Atom.to_string(key), check_dtype!(key, value)}
{key, value} when is_binary(key) ->
{key, check_dtype!(key, value)}
_ ->
raise ArgumentError,
"dtypes must be a list/map of column names as keys and types as values, " <>
"where the keys are atoms or binaries. Got: #{inspect(dtypes)}"
end)
end
defp check_dtype!(_key, value) when value in @valid_dtypes, do: value
defp check_dtype!(key, value) do
raise ArgumentError,
"invalid dtype #{inspect(value)} for #{inspect(key)} (expected one of #{inspect(@valid_dtypes)})"
end
# Access
@behaviour Access
@impl true
def fetch(df, column) when is_column(column) do
{:ok, pull(df, column)}
end
def fetch(df, columns) do
columns = to_existing_columns(df, columns)
{:ok, select(df, columns)}
end
@impl true
def pop(df, column) when is_column(column) do
[column] = to_existing_columns(df, [column])
{pull(df, column), select(df, [column], :drop)}
end
def pop(df, columns) do
columns = to_existing_columns(df, columns)
{select(df, columns), select(df, columns, :drop)}
end
@impl true
def get_and_update(df, column, fun) when is_column(column) do
[column] = to_existing_columns(df, [column])
value = pull(df, column)
{current_value, new_value} = fun.(value)
new_data = mutate(df, %{column => new_value})
{current_value, new_data}
end
# IO
@doc """
Reads a delimited file into a dataframe.
If the CSV is compressed, it is automatically decompressed.
## Options
* `delimiter` - A single character used to separate fields within a record. (default: `","`)
* `dtypes` - A list/map of `{"column_name", dtype}` tuples. Any non-specified column has its type
imputed from the first 1000 rows. (default: `[]`)
* `header` - Does the file have a header of column names as the first row or not? (default: `true`)
* `max_rows` - Maximum number of lines to read. (default: `nil`)
* `null_character` - The string that should be interpreted as a nil value. (default: `"NA"`)
* `skip_rows` - The number of lines to skip at the beginning of the file. (default: `0`)
* `columns` - A list of column names or indexes to keep. If present, only these columns are read into the dataframe. (default: `nil`)
* `infer_schema_length` Maximum number of rows read for schema inference. Setting this to nil will do a full table scan and will be slow (default: `1000`).
* `parse_dates` - Automatically try to parse dates/ datetimes and time. If parsing fails, columns remain of dtype `string`
"""
@doc type: :io
@spec from_csv(filename :: String.t(), opts :: Keyword.t()) ::
{:ok, DataFrame.t()} | {:error, term()}
def from_csv(filename, opts \\ []) do
opts =
Keyword.validate!(opts,
delimiter: ",",
dtypes: [],
encoding: "utf8",
header: true,
max_rows: nil,
null_character: "NA",
skip_rows: 0,
columns: nil,
infer_schema_length: @default_infer_schema_length,
parse_dates: false
)
backend = backend_from_options!(opts)
backend.from_csv(
filename,
check_dtypes!(opts[:dtypes]),
opts[:delimiter],
opts[:null_character],
opts[:skip_rows],
opts[:header],
opts[:encoding],
opts[:max_rows],
opts[:columns],
opts[:infer_schema_length],
opts[:parse_dates]
)
end
@doc """
Similar to `from_csv/2` but raises if there is a problem reading the CSV.
"""
@doc type: :io
@spec from_csv!(filename :: String.t(), opts :: Keyword.t()) :: DataFrame.t()
def from_csv!(filename, opts \\ []) do
case from_csv(filename, opts) do
{:ok, df} -> df
{:error, error} -> raise "#{error}"
end
end
@doc """
Reads a parquet file into a dataframe.
"""
@doc type: :io
@spec from_parquet(filename :: String.t(), opts :: Keyword.t()) ::
{:ok, DataFrame.t()} | {:error, term()}
def from_parquet(filename, opts \\ []) do
backend = backend_from_options!(opts)
backend.from_parquet(filename)
end
@doc """
Writes a dataframe to a parquet file.
## Options
* `compression` - The compression algorithm to use when writing files.
Where a compression level is available, this can be passed as a tuple,
such as `{:zstd, 3}`. Supported options are:
* `nil` (uncompressed, default)
* `:snappy`
* `:gzip` (with levels 1-9)
* `:brotli` (with levels 1-11)
* `:zstd` (with levels -7-22)
* `:lz4raw`.
"""
@doc type: :io
@spec to_parquet(df :: DataFrame.t(), filename :: String.t()) ::
{:ok, String.t()} | {:error, term()}
def to_parquet(df, filename, opts \\ []) do
opts = Keyword.validate!(opts, compression: nil)
compression = parquet_compression(opts[:compression])
Shared.apply_impl(df, :to_parquet, [filename, compression])
end
defp parquet_compression(nil), do: {nil, nil}
defp parquet_compression(algorithm) when algorithm in ~w(snappy gzip brotli zstd lz4raw)a do
{algorithm, nil}
end
for {algorithm, min, max} <- [{:gzip, 1, 9}, {:brotli, 1, 11}, {:zstd, -7, 22}] do
defp parquet_compression({unquote(algorithm), level}) do
if level in unquote(min)..unquote(max) or is_nil(level) do
{unquote(algorithm), level}
else
raise ArgumentError,
"#{unquote(algorithm)} compression level must be between #{unquote(min)} and #{unquote(max)} inclusive or nil, got #{level}"
end
end
end
defp parquet_compression(other) do
raise ArgumentError, "unsupported :compression #{inspect(other)} for Parquet"
end
@doc """
Reads an IPC file into a dataframe.
## Options
* `columns` - List with the name or index of columns to be selected. Defaults to all columns.
"""
@doc type: :io
@spec from_ipc(filename :: String.t()) :: {:ok, DataFrame.t()} | {:error, term()}
def from_ipc(filename, opts \\ []) do
opts =
Keyword.validate!(opts,
columns: nil
)
backend = backend_from_options!(opts)
backend.from_ipc(
filename,
opts[:columns]
)
end
@doc """
Similar to `from_ipc/2` but raises if there is a problem reading the IPC file.
"""
@doc type: :io
@spec from_ipc!(filename :: String.t(), opts :: Keyword.t()) :: DataFrame.t()
def from_ipc!(filename, opts \\ []) do
case from_ipc(filename, opts) do
{:ok, df} -> df
{:error, error} -> raise "#{error}"
end
end
@doc """
Writes a dataframe to a IPC file.
Apache IPC is a language-agnostic columnar data structure that can be used to store data frames.
It excels as a format for quickly exchange data between different programming languages.
## Options
* `compression` - Sets the algorithm used to compress the IPC file.
It accepts `:zstd` or `:lz4` compression. (default: `nil`)
"""
@doc type: :io
@spec to_ipc(df :: DataFrame.t(), filename :: String.t()) ::
{:ok, String.t()} | {:error, term()}
def to_ipc(df, filename, opts \\ []) do
opts = Keyword.validate!(opts, compression: nil)
backend = backend_from_options!(opts)
compression = opts[:compression]
unless is_nil(compression) or compression in ~w(zstd lz4) do
raise ArgumentError, "unsupported :compression #{inspect(compression)} for IPC"
end
backend.to_ipc(df, filename, {compression, nil})
end
@doc """
Writes a dataframe to a delimited file.
## Options
* `header` - Should the column names be written as the first line of the file? (default: `true`)
* `delimiter` - A single character used to separate fields within a record. (default: `","`)
"""
@doc type: :io
@spec to_csv(df :: DataFrame.t(), filename :: String.t(), opts :: Keyword.t()) ::
{:ok, String.t()} | {:error, term()}
def to_csv(df, filename, opts \\ []) do
opts = Keyword.validate!(opts, header: true, delimiter: ",")
Shared.apply_impl(df, :to_csv, [filename, opts[:header], opts[:delimiter]])
end
@doc """
Similar to `to_csv/3` but raises if there is a problem reading the CSV.
"""
@doc type: :io
@spec to_csv!(df :: DataFrame.t(), filename :: String.t(), opts :: Keyword.t()) :: String.t()
def to_csv!(df, filename, opts \\ []) do
case to_csv(df, filename, opts) do
{:ok, filename} -> filename
{:error, error} -> raise "#{error}"
end
end
@doc """
Read a file of JSON objects or lists separated by new lines
## Options
* `batch_size` - Sets the batch size for reading rows.
This value may have significant impact in performance, so adjust it for your needs (default: `1000`).
* `infer_schema_length` - Maximum number of rows read for schema inference.
Setting this to nil will do a full table scan and will be slow (default: `1000`).
"""
@doc type: :io
@spec from_ndjson(filename :: String.t(), opts :: Keyword.t()) ::
{:ok, DataFrame.t()} | {:error, term()}
def from_ndjson(filename, opts \\ []) do
opts =
Keyword.validate!(opts,
batch_size: 1000,
infer_schema_length: @default_infer_schema_length
)
backend = backend_from_options!(opts)
backend.from_ndjson(
filename,
opts[:infer_schema_length],
opts[:batch_size]
)
end
@doc """
Writes a dataframe to a ndjson file.
"""
@doc type: :io
@spec to_ndjson(df :: DataFrame.t(), filename :: String.t()) ::
{:ok, String.t()} | {:error, term()}
def to_ndjson(df, filename) do
Shared.apply_impl(df, :to_ndjson, [filename])
end
@doc """
Writes a dataframe to a binary representation of a delimited file.
## Options
* `header` - Should the column names be written as the first line of the file? (default: `true`)
* `delimiter` - A single character used to separate fields within a record. (default: `","`)
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> df |> Explorer.DataFrame.head(2) |> Explorer.DataFrame.dump_csv()
"year,country,total,solid_fuel,liquid_fuel,gas_fuel,cement,gas_flaring,per_capita,bunker_fuels\\n2010,AFGHANISTAN,2308,627,1601,74,5,0,0.08,9\\n2010,ALBANIA,1254,117,953,7,177,0,0.43,7\\n"
"""
@doc type: :io
@spec dump_csv(df :: DataFrame.t(), opts :: Keyword.t()) :: String.t()
def dump_csv(df, opts \\ []) do
opts = Keyword.validate!(opts, header: true, delimiter: ",")
Shared.apply_impl(df, :dump_csv, [opts[:header], opts[:delimiter]])
end
## Conversion
@doc """
Converts the dataframe to the lazy version of the current backend.
If already lazy, this is a noop.
"""
@doc type: :single
@spec to_lazy(df :: DataFrame.t()) :: DataFrame.t()
def to_lazy(df), do: Shared.apply_impl(df, :to_lazy)
@doc """
This collects the lazy data frame into an eager one, computing the query.
If already eager, this is a noop.
"""
@doc type: :single
@spec collect(df :: DataFrame.t()) :: {:ok, DataFrame.t()} | {:error, term()}
def collect(df), do: Shared.apply_impl(df, :collect)
@doc """
Creates a new dataframe.
Accepts any tabular data adhering to the `Table.Reader` protocol, as well as a map or a keyword list with series.
## Options
* `backend` - The Explorer backend to use. Defaults to the value returned by `Explorer.Backend.get/0`.
## Examples
From series:
iex> Explorer.DataFrame.new(%{floats: Explorer.Series.from_list([1.0, 2.0]), ints: Explorer.Series.from_list([1, nil])})
#Explorer.DataFrame<
Polars[2 x 2]
floats float [1.0, 2.0]
ints integer [1, nil]
>
From columnar data:
iex> Explorer.DataFrame.new(%{floats: [1.0, 2.0], ints: [1, nil]})
#Explorer.DataFrame<
Polars[2 x 2]
floats float [1.0, 2.0]
ints integer [1, nil]
>
iex> Explorer.DataFrame.new(floats: [1.0, 2.0], ints: [1, nil])
#Explorer.DataFrame<
Polars[2 x 2]
floats float [1.0, 2.0]
ints integer [1, nil]
>
iex> Explorer.DataFrame.new(%{floats: [1.0, 2.0], ints: [1, "wrong"]})
** (ArgumentError) cannot create series "ints": the value "wrong" does not match the inferred series dtype :integer
From row data:
iex> rows = [%{id: 1, name: "José"}, %{id: 2, name: "Christopher"}, %{id: 3, name: "Cristine"}]
iex> Explorer.DataFrame.new(rows)
#Explorer.DataFrame<
Polars[3 x 2]
id integer [1, 2, 3]
name string ["José", "Christopher", "Cristine"]
>
iex> rows = [[id: 1, name: "José"], [id: 2, name: "Christopher"], [id: 3, name: "Cristine"]]
iex> Explorer.DataFrame.new(rows)
#Explorer.DataFrame<
Polars[3 x 2]
id integer [1, 2, 3]
name string ["José", "Christopher", "Cristine"]
>
"""
@doc type: :single
@spec new(
Table.Reader.t() | series_pairs,
opts :: Keyword.t()
) :: DataFrame.t()
when series_pairs: %{column_name() => Series.t()} | [{column_name(), Series.t()}]
def new(data, opts \\ []) do
backend = backend_from_options!(opts)
case data do
%DataFrame{data: %^backend{}} ->
data
data ->
case classify_data(data) do
{:series, series} -> backend.from_series(series)
{:other, tabular} -> backend.from_tabular(tabular)
end
end
end
defp classify_data([{_, %Series{}} | _] = data), do: {:series, data}
defp classify_data(%_{} = data), do: {:other, data}
defp classify_data(data) when is_map(data) do
case :maps.next(:maps.iterator(data)) do
{_key, %Series{}, _} -> {:series, data}
_ -> {:other, data}
end
end
defp classify_data(data), do: {:other, data}
@doc """
Converts a dataframe to a list of columns with lists as values.
See `to_series/2` if you want a list of columns with series as values.
## Options
* `:atom_keys` - Configure if the resultant map should have atom keys. (default: `false`)
## Examples
iex> df = Explorer.DataFrame.new(ints: [1, nil], floats: [1.0, 2.0])
iex> Explorer.DataFrame.to_columns(df)
%{"floats" => [1.0, 2.0], "ints" => [1, nil]}
iex> df = Explorer.DataFrame.new(floats: [1.0, 2.0], ints: [1, nil])
iex> Explorer.DataFrame.to_columns(df, atom_keys: true)
%{floats: [1.0, 2.0], ints: [1, nil]}
"""
@doc type: :single
@spec to_columns(df :: DataFrame.t(), Keyword.t()) :: map()
def to_columns(df, opts \\ []) do
opts = Keyword.validate!(opts, atom_keys: false)
atom_keys = opts[:atom_keys]
for name <- df.names, into: %{} do
series = Shared.apply_impl(df, :pull, [name])
key = if atom_keys, do: String.to_atom(name), else: name
{key, Series.to_list(series)}
end
end
@doc """
Converts a dataframe to a list of columns with series as values.
See `to_columns/2` if you want a list of columns with lists as values.
## Options
* `:atom_keys` - Configure if the resultant map should have atom keys. (default: `false`)
## Examples
iex> df = Explorer.DataFrame.new(ints: [1, nil], floats: [1.0, 2.0])
iex> map = Explorer.DataFrame.to_series(df)
iex> Explorer.Series.to_list(map["floats"])
[1.0, 2.0]
iex> Explorer.Series.to_list(map["ints"])
[1, nil]
"""
@doc type: :single
@spec to_series(df :: DataFrame.t(), Keyword.t()) :: map()
def to_series(df, opts \\ []) do
opts = Keyword.validate!(opts, atom_keys: false)
atom_keys = opts[:atom_keys]
for name <- df.names, into: %{} do
key = if atom_keys, do: String.to_atom(name), else: name
{key, Shared.apply_impl(df, :pull, [name])}
end
end
@doc """
Converts a dataframe to a list of maps (rows).
> #### Warning {: .warning}
>
> This may be an expensive operation because `polars` stores data in columnar format.
## Options
* `:atom_keys` - Configure if the resultant maps should have atom keys. (default: `false`)
## Examples
iex> df = Explorer.DataFrame.new(floats: [1.0, 2.0], ints: [1, nil])
iex> Explorer.DataFrame.to_rows(df)
[%{"floats" => 1.0, "ints" => 1}, %{"floats" => 2.0 ,"ints" => nil}]
iex> df = Explorer.DataFrame.new(floats: [1.0, 2.0], ints: [1, nil])
iex> Explorer.DataFrame.to_rows(df, atom_keys: true)
[%{floats: 1.0, ints: 1}, %{floats: 2.0, ints: nil}]
"""
@doc type: :single
@spec to_rows(df :: DataFrame.t(), Keyword.t()) :: [map()]
def to_rows(df, opts \\ []) do
opts = Keyword.validate!(opts, atom_keys: false)
Shared.apply_impl(df, :to_rows, [opts[:atom_keys]])
end
# Introspection
@doc """
Gets the names of the dataframe columns.
## Examples
iex> df = Explorer.DataFrame.new(floats: [1.0, 2.0], ints: [1, 2])
iex> Explorer.DataFrame.names(df)
["floats", "ints"]
"""
@doc type: :introspection
@spec names(df :: DataFrame.t()) :: [String.t()]
def names(df), do: df.names
@doc """
Gets the dtypes of the dataframe columns.
## Examples
iex> df = Explorer.DataFrame.new(floats: [1.0, 2.0], ints: [1, 2])
iex> Explorer.DataFrame.dtypes(df)
%{"floats" => :float, "ints" => :integer}
"""
@doc type: :introspection
@spec dtypes(df :: DataFrame.t()) :: %{String.t() => atom()}
def dtypes(df), do: df.dtypes
@doc """
Gets the shape of the dataframe as a `{height, width}` tuple.
## Examples
iex> df = Explorer.DataFrame.new(floats: [1.0, 2.0, 3.0], ints: [1, 2, 3])
iex> Explorer.DataFrame.shape(df)
{3, 2}
"""
@doc type: :introspection
@spec shape(df :: DataFrame.t()) :: {integer(), integer()}
def shape(df), do: {n_rows(df), n_columns(df)}
@doc """
Returns the number of rows in the dataframe.
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.n_rows(df)
1094
"""
@doc type: :introspection
@spec n_rows(df :: DataFrame.t()) :: integer()
def n_rows(df), do: Shared.apply_impl(df, :n_rows)
@doc """
Returns the number of columns in the dataframe.
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.n_columns(df)
10
"""
@doc type: :introspection
@spec n_columns(df :: DataFrame.t()) :: integer()
def n_columns(df), do: map_size(df.dtypes)
@doc """
Returns the groups of a dataframe.
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> df = Explorer.DataFrame.group_by(df, "country")
iex> Explorer.DataFrame.groups(df)
["country"]
"""
@doc type: :introspection
@spec groups(df :: DataFrame.t()) :: list(String.t())
def groups(%DataFrame{groups: groups}), do: groups
# Single table verbs
@doc """
Returns the first *n* rows of the dataframe.
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.head(df)
#Explorer.DataFrame<
Polars[5 x 10]
year integer [2010, 2010, 2010, 2010, 2010]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA"]
total integer [2308, 1254, 32500, 141, 7924]
solid_fuel integer [627, 117, 332, 0, 0]
liquid_fuel integer [1601, 953, 12381, 141, 3649]
gas_fuel integer [74, 7, 14565, 0, 374]
cement integer [5, 177, 2598, 0, 204]
gas_flaring integer [0, 0, 2623, 0, 3697]
per_capita float [0.08, 0.43, 0.9, 1.68, 0.37]
bunker_fuels integer [9, 7, 663, 0, 321]
>
"""
@doc type: :rows
@spec head(df :: DataFrame.t(), nrows :: integer()) :: DataFrame.t()
def head(df, nrows \\ @default_sample_nrows), do: Shared.apply_impl(df, :head, [nrows])
@doc """
Returns the last *n* rows of the dataframe.
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.tail(df)
#Explorer.DataFrame<
Polars[5 x 10]
year integer [2014, 2014, 2014, 2014, 2014]
country string ["VIET NAM", "WALLIS AND FUTUNA ISLANDS", "YEMEN", "ZAMBIA", "ZIMBABWE"]
total integer [45517, 6, 6190, 1228, 3278]
solid_fuel integer [19246, 0, 137, 132, 2097]
liquid_fuel integer [12694, 6, 5090, 797, 1005]
gas_fuel integer [5349, 0, 581, 0, 0]
cement integer [8229, 0, 381, 299, 177]
gas_flaring integer [0, 0, 0, 0, 0]
per_capita float [0.49, 0.44, 0.24, 0.08, 0.22]
bunker_fuels integer [761, 1, 153, 33, 9]
>
"""
@doc type: :rows
@spec tail(df :: DataFrame.t(), nrows :: integer()) :: DataFrame.t()
def tail(df, nrows \\ @default_sample_nrows), do: Shared.apply_impl(df, :tail, [nrows])
@doc """
Selects a subset of columns by name.
Can optionally return all *but* the named columns if `:drop` is passed as the last argument.
## Examples
You can select a single column:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.select(df, "a")
#Explorer.DataFrame<
Polars[3 x 1]
a string ["a", "b", "c"]
>
Or a list of names:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.select(df, ["a"])
#Explorer.DataFrame<
Polars[3 x 1]
a string ["a", "b", "c"]
>
You can also use a range or a list of integers:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3], c: [4, 5, 6])
iex> Explorer.DataFrame.select(df, [0, 1])
#Explorer.DataFrame<
Polars[3 x 2]
a string ["a", "b", "c"]
b integer [1, 2, 3]
>
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3], c: [4, 5, 6])
iex> Explorer.DataFrame.select(df, 0..1)
#Explorer.DataFrame<
Polars[3 x 2]
a string ["a", "b", "c"]
b integer [1, 2, 3]
>
Or you can use a callback function that takes the dataframe's names as its first argument:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.select(df, &String.starts_with?(&1, "b"))
#Explorer.DataFrame<
Polars[3 x 1]
b integer [1, 2, 3]
>
Or a callback function that takes names and types:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.select(df, fn _name, type -> type == :integer end)
#Explorer.DataFrame<
Polars[3 x 1]
b integer [1, 2, 3]
>
If you pass `:drop` as the third argument, it will return all but the named columns:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.select(df, ["b"], :drop)
#Explorer.DataFrame<
Polars[3 x 1]
a string ["a", "b", "c"]
>
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3], c: [4, 5, 6])
iex> Explorer.DataFrame.select(df, ["a", "b"], :drop)
#Explorer.DataFrame<
Polars[3 x 1]
c integer [4, 5, 6]
>
"""
@doc type: :single
@spec select(
df :: DataFrame.t(),
column | columns,
keep_or_drop :: :keep | :drop
) :: DataFrame.t()
def select(df, columns_or_column, keep_or_drop \\ :keep)
def select(df, column, keep_or_drop) when is_column(column) do
select(df, [column], keep_or_drop)
end
def select(df, columns, keep_or_drop) do
columns = to_existing_columns(df, columns)
columns_to_keep =
case keep_or_drop do
:keep -> columns
:drop -> df.names -- columns
end
out_df = %{df | names: columns_to_keep, dtypes: Map.take(df.dtypes, columns_to_keep)}
Shared.apply_impl(df, :select, [out_df])
end
@doc """
Picks rows based on a list or series of values.
## Examples
This function must only be used when you need to select rows based
on external values that are not available to the dataframe. For example,
you can pass a list:
iex> df = Explorer.DataFrame.new(col1: ["a", "b", "c"], col2: [1, 2, 3])
iex> Explorer.DataFrame.mask(df, [false, true, false])
#Explorer.DataFrame<
Polars[1 x 2]
col1 string ["b"]
col2 integer [2]
>
You must avoid using masks when the masks themselves are computed from
other columns. For example, DO NOT do this:
iex> df = Explorer.DataFrame.new(col1: ["a", "b", "c"], col2: [1, 2, 3])
iex> Explorer.DataFrame.mask(df, Explorer.Series.greater(df["col2"], 1))
#Explorer.DataFrame<
Polars[2 x 2]
col1 string ["b", "c"]
col2 integer [2, 3]
>
Instead, do this:
iex> df = Explorer.DataFrame.new(col1: ["a", "b", "c"], col2: [1, 2, 3])
iex> Explorer.DataFrame.filter_with(df, fn df -> Explorer.Series.greater(df["col2"], 1) end)
#Explorer.DataFrame<
Polars[2 x 2]
col1 string ["b", "c"]
col2 integer [2, 3]
>
The `filter_with/2` version is much more efficient because it doesn't need
to create intermediate series representations to apply the mask.
"""
@doc type: :single
@spec mask(df :: DataFrame.t(), mask :: Series.t() | [boolean()]) :: DataFrame.t()
def mask(df, %Series{} = mask) do
s_len = Series.size(mask)
df_len = n_rows(df)
case s_len == df_len do
false ->
raise(
ArgumentError,
"size of the mask (#{s_len}) must match number of rows in the dataframe (#{df_len})"
)
true ->
Shared.apply_impl(df, :mask, [mask])
end
end
def mask(df, mask) when is_list(mask), do: mask |> Series.from_list() |> then(&mask(df, &1))
@spec filter(df :: DataFrame.t(), callback :: function()) :: DataFrame.t()
def filter(df, callback) when is_function(callback) do
IO.warn("filter/2 with a callback is deprecated, please use filter_with/2 instead")
df |> callback.() |> then(&mask(df, &1))
end
@doc """
Picks rows based on a callback function.
This function is efficient because it uses a representation of the
series without pulling them. The only restriction is that
you need to use a function that returns boolean.
But you can also use window functions and aggregations inside comparisons.
## Examples
iex> df = Explorer.DataFrame.new(col1: ["a", "b", "c"], col2: [1, 2, 3])
iex> Explorer.DataFrame.filter_with(df, &Explorer.Series.greater(&1["col2"], 2))
#Explorer.DataFrame<
Polars[1 x 2]
col1 string ["c"]
col2 integer [3]
>
iex> df = Explorer.DataFrame.new(col1: ["a", "b", "c"], col2: [1, 2, 3])
iex> Explorer.DataFrame.filter_with(df, fn df -> Explorer.Series.equal(df["col1"], "b") end)
#Explorer.DataFrame<
Polars[1 x 2]
col1 string ["b"]
col2 integer [2]
>
iex> df = Explorer.DataFrame.new(col1: ["a", "b", "c"], col2: [1, 2, 3])
iex> Explorer.DataFrame.filter_with(df, fn df -> Explorer.Series.cumulative_max(df["col2"]) end)
** (ArgumentError) expecting the function to return a boolean LazySeries, but instead it returned a LazySeries of type :integer
But it's possible to use a boolean operation based on another function:
iex> df = Explorer.DataFrame.new(col1: ["a", "b", "c"], col2: [1, 2, 3])
iex> Explorer.DataFrame.filter_with(df, fn df -> Explorer.Series.equal(Explorer.Series.cumulative_max(df["col2"]), 1) end)
#Explorer.DataFrame<
Polars[1 x 2]
col1 string ["a"]
col2 integer [1]
>
"""
@doc type: :single
@spec filter_with(
df :: DataFrame.t(),
callback :: (Explorer.Backend.LazyFrame.t() -> Series.lazy_t())
) :: DataFrame.t()
def filter_with(df, fun) when is_function(fun, 1) do
ldf = to_opaque_lazy(df)
case fun.(ldf) do
%Series{dtype: :boolean, data: %LazySeries{} = data} ->
Shared.apply_impl(df, :filter_with, [df, data])
%Series{dtype: dtype, data: %LazySeries{}} ->
raise ArgumentError,
"expecting the function to return a boolean LazySeries, " <>
"but instead it returned a LazySeries of type " <>
inspect(dtype)
other ->
raise ArgumentError,
"expecting the function to return a LazySeries, but instead it returned " <>
inspect(other)
end
end
defp to_opaque_lazy(%DataFrame{} = df) do
df
|> Explorer.Backend.LazyFrame.new()
|> Explorer.Backend.DataFrame.new(df.names, df.dtypes)
end
@doc """
Creates and modifies columns.
Columns are added with keyword list or maps. New variables overwrite existing variables of the
same name. Column names are coerced from atoms to strings.
## Examples
You can pass in a list directly as a new column:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.mutate(df, c: [4, 5, 6])
#Explorer.DataFrame<
Polars[3 x 3]
a string ["a", "b", "c"]
b integer [1, 2, 3]
c integer [4, 5, 6]
>
Or you can pass in a series:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> s = Explorer.Series.from_list([4, 5, 6])
iex> Explorer.DataFrame.mutate(df, c: s)
#Explorer.DataFrame<
Polars[3 x 3]
a string ["a", "b", "c"]
b integer [1, 2, 3]
c integer [4, 5, 6]
>
You can overwrite existing columns:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.mutate(df, a: [4, 5, 6])
#Explorer.DataFrame<
Polars[3 x 2]
a integer [4, 5, 6]
b integer [1, 2, 3]
>
Scalar values are repeated to fill the series:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.mutate(df, a: 4)
#Explorer.DataFrame<
Polars[3 x 2]
a integer [4, 4, 4]
b integer [1, 2, 3]
>
Alternatively, all of the above works with a map instead of a keyword list:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.mutate(df, %{"c" => [4, 5, 6]})
#Explorer.DataFrame<
Polars[3 x 3]
a string ["a", "b", "c"]
b integer [1, 2, 3]
c integer [4, 5, 6]
>
"""
@doc type: :single
@spec mutate(df :: DataFrame.t(), columns :: column_pairs(any())) ::
DataFrame.t()
def mutate(df, columns) when is_column_pairs(columns) do
mutations = to_column_pairs(df, columns)
out_df = df_for_mutations(df, mutations)
Shared.apply_impl(df, :mutate, [out_df, mutations])
end
defp df_for_mutations(df, mutations) do
mut_names = Enum.map(mutations, &elem(&1, 0))
new_names = Enum.uniq(df.names ++ mut_names)
mut_dtypes =
for {name, value} <- mutations, into: %{} do
value =
if is_function(value) do
IO.warn("mutate/2 with a callback is deprecated, please use mutate_with/2 instead")
value.(df)
else
value
end
dtype =
case value do
value when is_list(value) ->
Shared.check_types!(value)
%Series{} = value ->
value.dtype
value ->
Shared.check_types!([value])
end
{name, dtype}
end
new_dtypes = Map.merge(df.dtypes, mut_dtypes)
%{df | names: new_names, dtypes: new_dtypes}
end
@doc """
Creates or modifies columns using a callback function.
This function is similar to `mutate/2`, but allows complex operations
to be performed, since it uses a virtual representation of the dataframe.
The only requirement is that a series operation is returned.
New variables overwrite existing variables of the
same name. Column names are coerced from atoms to strings.
## Examples
iex> df = Explorer.DataFrame.new(a: [4, 5, 6], b: [1, 2, 3])
iex> Explorer.DataFrame.mutate_with(df, &[c: Explorer.Series.add(&1["a"], &1["b"])])
#Explorer.DataFrame<
Polars[3 x 3]
a integer [4, 5, 6]
b integer [1, 2, 3]
c integer [5, 7, 9]
>
You can overwrite existing columns:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.mutate_with(df, &[b: Explorer.Series.pow(&1["b"], 2)])
#Explorer.DataFrame<
Polars[3 x 2]
a string ["a", "b", "c"]
b float [1.0, 4.0, 9.0]
>
Scalar values are repeated to fill the series:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.mutate_with(df, &[a: Explorer.Series.max(&1["b"])])
#Explorer.DataFrame<
Polars[3 x 2]
a integer [3, 3, 3]
b integer [1, 2, 3]
>
Alternatively, all of the above works with a map instead of a keyword list:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "c"], b: [1, 2, 3])
iex> Explorer.DataFrame.mutate_with(df, fn df -> %{"c" => Explorer.Series.window_mean(df["b"], 2)} end)
#Explorer.DataFrame<
Polars[3 x 3]
a string ["a", "b", "c"]
b integer [1, 2, 3]
c float [1.0, 1.5, 2.5]
>
Mutations in grouped dataframes are also possible:
iex> df = Explorer.DataFrame.new(id: ["a", "a", "b"], b: [1, 2, 3])
iex> grouped = Explorer.DataFrame.group_by(df, :id)
iex> Explorer.DataFrame.mutate_with(grouped, &[count: Explorer.Series.count(&1["b"])])
#Explorer.DataFrame<
Polars[3 x 3]
Groups: ["id"]
id string ["a", "a", "b"]
b integer [1, 2, 3]
count integer [2, 2, 1]
>
"""
@doc type: :single
@spec mutate_with(
df :: DataFrame.t(),
callback :: (Explorer.Backend.LazyFrame.t() -> column_pairs(Series.lazy_t()))
) :: DataFrame.t()
def mutate_with(%DataFrame{} = df, fun) when is_function(fun) do
ldf = to_opaque_lazy(df)
result = fun.(ldf)
column_pairs =
to_column_pairs(df, result, fn value ->
case value do
%Series{data: %LazySeries{}} ->
value
other ->
raise "expecting a lazy series, but instead got #{inspect(other)}"
end
end)
new_dtypes =
for {column_name, series} <- column_pairs, into: %{} do
{column_name, series.dtype}
end
mut_names = Enum.map(column_pairs, &elem(&1, 0))
new_names = Enum.uniq(df.names ++ mut_names)
df_out = %{df | names: new_names, dtypes: Map.merge(df.dtypes, new_dtypes)}
column_pairs = for {name, %Series{data: lazy_series}} <- column_pairs, do: {name, lazy_series}
Shared.apply_impl(df, :mutate_with, [df_out, column_pairs])
end
@doc """
Arranges/sorts rows by columns.
## Examples
A single column name will sort ascending by that column:
iex> df = Explorer.DataFrame.new(a: ["b", "c", "a"], b: [1, 2, 3])
iex> Explorer.DataFrame.arrange(df, "a")
#Explorer.DataFrame<
Polars[3 x 2]
a string ["a", "b", "c"]
b integer [3, 1, 2]
>
You can also sort descending:
iex> df = Explorer.DataFrame.new(a: ["b", "c", "a"], b: [1, 2, 3])
iex> Explorer.DataFrame.arrange(df, desc: "a")
#Explorer.DataFrame<
Polars[3 x 2]
a string ["c", "b", "a"]
b integer [2, 1, 3]
>
Sorting by more than one column sorts them in the order they are entered:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.arrange(df, asc: "total", desc: "country")
#Explorer.DataFrame<
Polars[1094 x 10]
year integer [2010, 2010, 2011, 2011, 2012, ...]
country string ["NIUE", "TUVALU", "TUVALU", "NIUE", "NIUE", ...]
total integer [1, 2, 2, 2, 2, ...]
solid_fuel integer [0, 0, 0, 0, 0, ...]
liquid_fuel integer [1, 2, 2, 2, 2, ...]
gas_fuel integer [0, 0, 0, 0, 0, ...]
cement integer [0, 0, 0, 0, 0, ...]
gas_flaring integer [0, 0, 0, 0, 0, ...]
per_capita float [0.52, 0.0, 0.0, 1.04, 1.04, ...]
bunker_fuels integer [0, 0, 0, 0, 0, ...]
>
Alternatively you can pass a callback to sort by the given columns:
iex> df = Explorer.DataFrame.new(a: ["b", "c", "a"], b: [1, 2, 3])
iex> Explorer.DataFrame.arrange(df, &String.starts_with?(&1, "a"))
#Explorer.DataFrame<
Polars[3 x 2]
a string ["a", "b", "c"]
b integer [3, 1, 2]
>
Or a callback to sort the columns of a given type:
iex> df = Explorer.DataFrame.new(a: ["b", "c", "a"], b: [1, 2, 3])
iex> Explorer.DataFrame.arrange(df, fn _name, type -> type == :string end)
#Explorer.DataFrame<
Polars[3 x 2]
a string ["a", "b", "c"]
b integer [3, 1, 2]
>
"""
@doc type: :single
@spec arrange(
df :: DataFrame.t(),
column_or_columns :: column() | columns() | [{:asc | :desc, column()}]
) :: DataFrame.t()
def arrange(df, columns_or_column)
def arrange(df, columns) when is_list(columns) do
{dirs, columns} =
Enum.map(columns, fn
{dir, column} when dir in [:asc, :desc] and is_column(column) ->
{dir, column}
column when is_column(column) ->
{:asc, column}
other ->
raise ArgumentError, "not a valid column or arrange instruction: #{inspect(other)}"
end)
|> Enum.unzip()
columns = to_existing_columns(df, columns)
Shared.apply_impl(df, :arrange, [Enum.zip(dirs, columns)])
end
def arrange(df, column) when is_column(column), do: arrange(df, [column])
def arrange(df, columns) do
columns = to_existing_columns(df, columns)
Shared.apply_impl(df, :arrange, [Enum.map(columns, &{:asc, &1})])
end
@doc """
Arranges/sorts rows by columns using a callback function.
## Examples
A single column name will sort ascending by that column:
iex> df = Explorer.DataFrame.new(a: ["b", "c", "a"], b: [1, 2, 3])
iex> Explorer.DataFrame.arrange_with(df, &(&1["a"]))
#Explorer.DataFrame<
Polars[3 x 2]
a string ["a", "b", "c"]
b integer [3, 1, 2]
>
You can also sort descending:
iex> df = Explorer.DataFrame.new(a: ["b", "c", "a"], b: [1, 2, 3])
iex> Explorer.DataFrame.arrange_with(df, &[desc: &1["a"]])
#Explorer.DataFrame<
Polars[3 x 2]
a string ["c", "b", "a"]
b integer [2, 1, 3]
>
Sorting by more than one column sorts them in the order they are entered:
iex> df = Explorer.DataFrame.new(a: [3, 1, 3], b: [2, 1, 3])
iex> Explorer.DataFrame.arrange_with(df, &[desc: &1["a"], asc: &1["b"]])
#Explorer.DataFrame<
Polars[3 x 2]
a integer [3, 3, 1]
b integer [2, 3, 1]
>
"""
@doc type: :single
@spec arrange_with(
df :: DataFrame.t(),
(Explorer.Backend.LazyFrame.t() ->
Series.lazy_t() | [Series.lazy_t()] | [{:asc | :desc, Series.lazy_t()}])
) :: DataFrame.t()
def arrange_with(%DataFrame{} = df, fun) when is_function(fun, 1) do
ldf = to_opaque_lazy(df)
result = fun.(ldf)
dir_and_lazy_series_pairs =
result
|> List.wrap()
|> Enum.map(fn
{dir, %Series{data: %LazySeries{} = lazy_series}} when dir in [:asc, :desc] ->
{dir, lazy_series}
{wrong_dir, %Series{data: %LazySeries{}}} ->
raise "expecting a valid direction, which is :asc or :desc, but got #{inspect(wrong_dir)}."
{_, other} ->
raise "expecting a lazy series, but got #{inspect(other)}."
%Series{data: %LazySeries{} = lazy_series} ->
{:asc, lazy_series}
other ->
raise "not a valid lazy series or arrange instruction: #{inspect(other)}"
end)
Shared.apply_impl(df, :arrange_with, [df, dir_and_lazy_series_pairs])
end
@doc """
Takes distinct rows by a selection of columns.
Distinct is not affected by groups, although groups are kept in the
columns selection if `keep_all` option is false (the default).
## Options
* `keep_all` - If set to `true`, keep all columns. Default is `false`.
## Examples
By default will return unique values of the requested columns:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.distinct(df, ["year", "country"])
#Explorer.DataFrame<
Polars[1094 x 2]
year integer [2010, 2010, 2010, 2010, 2010, ...]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
>
If `keep_all` is set to `true`, then the first value of each column not in the requested
columns will be returned:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.distinct(df, ["year", "country"], keep_all: true)
#Explorer.DataFrame<
Polars[1094 x 10]
year integer [2010, 2010, 2010, 2010, 2010, ...]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
total integer [2308, 1254, 32500, 141, 7924, ...]
solid_fuel integer [627, 117, 332, 0, 0, ...]
liquid_fuel integer [1601, 953, 12381, 141, 3649, ...]
gas_fuel integer [74, 7, 14565, 0, 374, ...]
cement integer [5, 177, 2598, 0, 204, ...]
gas_flaring integer [0, 0, 2623, 0, 3697, ...]
per_capita float [0.08, 0.43, 0.9, 1.68, 0.37, ...]
bunker_fuels integer [9, 7, 663, 0, 321, ...]
>
A callback on the dataframe's names can be passed instead of a list (like `select/3`):
iex> df = Explorer.DataFrame.new(x1: [1, 3, 3], x2: ["a", "c", "c"], y1: [1, 2, 3])
iex> Explorer.DataFrame.distinct(df, &String.starts_with?(&1, "x"))
#Explorer.DataFrame<
Polars[2 x 2]
x1 integer [1, 3]
x2 string ["a", "c"]
>
If the dataframe has groups, then the columns of each group will be added to the distinct columns:
iex> df = Explorer.DataFrame.new(x1: [1, 3, 3], x2: ["a", "c", "c"], y1: [1, 2, 3])
iex> df = Explorer.DataFrame.group_by(df, "x1")
iex> Explorer.DataFrame.distinct(df, ["x2"])
#Explorer.DataFrame<
Polars[2 x 2]
Groups: ["x1"]
x1 integer [1, 3]
x2 string ["a", "c"]
>
"""
@doc type: :single
@spec distinct(df :: DataFrame.t(), columns :: columns(), opts :: Keyword.t()) :: DataFrame.t()
def distinct(df, columns \\ 0..-1//1, opts \\ [])
def distinct(df, columns, opts) do
opts = Keyword.validate!(opts, keep_all: false)
columns = to_existing_columns(df, columns)
if columns != [] do
out_df =
if opts[:keep_all] do
df
else
groups = df.groups
keep = if groups == [], do: columns, else: Enum.uniq(groups ++ columns)
%{df | names: keep, dtypes: Map.take(df.dtypes, keep)}
end
Shared.apply_impl(df, :distinct, [out_df, columns, opts[:keep_all]])
else
df
end
end
@doc """
Drop nil values.
Optionally accepts a subset of columns.
## Examples
To drop nils on all columns:
iex> df = Explorer.DataFrame.new(a: [1, 2, nil], b: [1, nil, 3])
iex> Explorer.DataFrame.drop_nil(df)
#Explorer.DataFrame<
Polars[1 x 2]
a integer [1]
b integer [1]
>
To select some columns:
iex> df = Explorer.DataFrame.new(a: [1, 2, nil], b: [1, nil, 3], c: [nil, 5, 6])
iex> Explorer.DataFrame.drop_nil(df, [:a, :c])
#Explorer.DataFrame<
Polars[1 x 3]
a integer [2]
b integer [nil]
c integer [5]
>
To select some columns by range:
iex> df = Explorer.DataFrame.new(a: [1, 2, nil], b: [1, nil, 3], c: [nil, 5, 6])
iex> Explorer.DataFrame.drop_nil(df, 0..1)
#Explorer.DataFrame<
Polars[1 x 3]
a integer [1]
b integer [1]
c integer [nil]
>
Or to select columns by a callback on the names:
iex> df = Explorer.DataFrame.new(a: [1, 2, nil], b: [1, nil, 3], c: [nil, 5, 6])
iex> Explorer.DataFrame.drop_nil(df, fn name -> name == "a" or name == "b" end)
#Explorer.DataFrame<
Polars[1 x 3]
a integer [1]
b integer [1]
c integer [nil]
>
Or to select columns by a callback on the names and types:
iex> df = Explorer.DataFrame.new(a: [1, 2, nil], b: [1, nil, 3], c: [nil, 5.0, 6.0])
iex> Explorer.DataFrame.drop_nil(df, fn _name, type -> type == :float end)
#Explorer.DataFrame<
Polars[2 x 3]
a integer [2, nil]
b integer [nil, 3]
c float [5.0, 6.0]
>
"""
@doc type: :single
@spec drop_nil(df :: DataFrame.t(), column() | columns()) ::
DataFrame.t()
def drop_nil(df, columns_or_column \\ 0..-1)
def drop_nil(df, column) when is_column(column), do: drop_nil(df, [column])
def drop_nil(df, columns) do
columns = to_existing_columns(df, columns)
Shared.apply_impl(df, :drop_nil, [columns])
end
@doc """
Renames columns.
To apply a function to a subset of columns, see `rename_with/3`.
## Examples
You can pass in a list of new names:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "a"], b: [1, 3, 1])
iex> Explorer.DataFrame.rename(df, ["c", "d"])
#Explorer.DataFrame<
Polars[3 x 2]
c string ["a", "b", "a"]
d integer [1, 3, 1]
>
Or you can rename individual columns using keyword args:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "a"], b: [1, 3, 1])
iex> Explorer.DataFrame.rename(df, a: "first")
#Explorer.DataFrame<
Polars[3 x 2]
first string ["a", "b", "a"]
b integer [1, 3, 1]
>
Or you can rename individual columns using a map:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "a"], b: [1, 3, 1])
iex> Explorer.DataFrame.rename(df, %{"a" => "first"})
#Explorer.DataFrame<
Polars[3 x 2]
first string ["a", "b", "a"]
b integer [1, 3, 1]
>
"""
@doc type: :single
@spec rename(
df :: DataFrame.t(),
names :: column_names() | column_pairs(column_name())
) ::
DataFrame.t()
def rename(df, [name | _] = names) when is_column_name(name) do
check_new_names_length!(df, names)
rename(df, Enum.zip(df.names, names))
end
def rename(df, names) when is_column_pairs(names) do
case to_column_pairs(df, names, &to_column_name(&1)) do
[] ->
df
pairs ->
pairs_map = Map.new(pairs)
old_dtypes = df.dtypes
for {name, _} <- pairs do
maybe_raise_column_not_found(df, name)
end
new_dtypes =
for name <- df.names do
{Map.get(pairs_map, name, name), Map.fetch!(old_dtypes, name)}
end
new_names = Enum.map(new_dtypes, &elem(&1, 0))
new_groups =
for group <- df.groups do
Map.get(pairs_map, group, group)
end
out_df = %{df | names: new_names, dtypes: Map.new(new_dtypes), groups: new_groups}
Shared.apply_impl(df, :rename, [out_df, pairs])
end
end
defp check_new_names_length!(df, names) do
width = n_columns(df)
n_new_names = length(names)
if width != n_new_names,
do:
raise(
ArgumentError,
"list of new names must match the number of columns in the dataframe; found " <>
"#{n_new_names} new name(s), but the supplied dataframe has #{width} column(s)"
)
end
@doc """
Renames columns with a function.
## Examples
If no columns are specified, it will apply the function to all column names:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.rename_with(df, &String.upcase/1)
#Explorer.DataFrame<
Polars[1094 x 10]
YEAR integer [2010, 2010, 2010, 2010, 2010, ...]
COUNTRY string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
TOTAL integer [2308, 1254, 32500, 141, 7924, ...]
SOLID_FUEL integer [627, 117, 332, 0, 0, ...]
LIQUID_FUEL integer [1601, 953, 12381, 141, 3649, ...]
GAS_FUEL integer [74, 7, 14565, 0, 374, ...]
CEMENT integer [5, 177, 2598, 0, 204, ...]
GAS_FLARING integer [0, 0, 2623, 0, 3697, ...]
PER_CAPITA float [0.08, 0.43, 0.9, 1.68, 0.37, ...]
BUNKER_FUELS integer [9, 7, 663, 0, 321, ...]
>
A callback can be used to filter the column names that will be renamed, similarly to `select/3`:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.rename_with(df, &String.ends_with?(&1, "_fuel"), &String.trim_trailing(&1, "_fuel"))
#Explorer.DataFrame<
Polars[1094 x 10]
year integer [2010, 2010, 2010, 2010, 2010, ...]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
total integer [2308, 1254, 32500, 141, 7924, ...]
solid integer [627, 117, 332, 0, 0, ...]
liquid integer [1601, 953, 12381, 141, 3649, ...]
gas integer [74, 7, 14565, 0, 374, ...]
cement integer [5, 177, 2598, 0, 204, ...]
gas_flaring integer [0, 0, 2623, 0, 3697, ...]
per_capita float [0.08, 0.43, 0.9, 1.68, 0.37, ...]
bunker_fuels integer [9, 7, 663, 0, 321, ...]
>
Or you can just pass in the list of column names you'd like to apply the function to:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.rename_with(df, ["total", "cement"], &String.upcase/1)
#Explorer.DataFrame<
Polars[1094 x 10]
year integer [2010, 2010, 2010, 2010, 2010, ...]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
TOTAL integer [2308, 1254, 32500, 141, 7924, ...]
solid_fuel integer [627, 117, 332, 0, 0, ...]
liquid_fuel integer [1601, 953, 12381, 141, 3649, ...]
gas_fuel integer [74, 7, 14565, 0, 374, ...]
CEMENT integer [5, 177, 2598, 0, 204, ...]
gas_flaring integer [0, 0, 2623, 0, 3697, ...]
per_capita float [0.08, 0.43, 0.9, 1.68, 0.37, ...]
bunker_fuels integer [9, 7, 663, 0, 321, ...]
>
"""
@doc type: :single
@spec rename_with(df :: DataFrame.t(), columns :: columns(), callback :: function()) ::
DataFrame.t()
def rename_with(df, columns \\ 0..-1//1, callback)
def rename_with(df, 0..-1//1, callback) when is_function(callback) do
df.names
|> Enum.map(callback)
|> then(&rename(df, &1))
end
def rename_with(df, columns, callback) when is_function(callback) do
columns = to_existing_columns(df, columns)
renames = for name <- df.names, name in columns, do: {name, callback.(name)}
rename(df, renames)
end
@doc """
Turns a set of columns to dummy variables.
## Examples
To mark a single column as dummy:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "a", "c"], b: ["b", "a", "b", "d"])
iex> Explorer.DataFrame.dummies(df, "a")
#Explorer.DataFrame<
Polars[4 x 3]
a_a integer [1, 0, 1, 0]
a_b integer [0, 1, 0, 0]
a_c integer [0, 0, 0, 1]
>
Or multiple columns:
iex> df = Explorer.DataFrame.new(a: ["a", "b", "a", "c"], b: ["b", "a", "b", "d"])
iex> Explorer.DataFrame.dummies(df, ["a", "b"])
#Explorer.DataFrame<
Polars[4 x 6]
a_a integer [1, 0, 1, 0]
a_b integer [0, 1, 0, 0]
a_c integer [0, 0, 0, 1]
b_a integer [0, 1, 0, 0]
b_b integer [1, 0, 1, 0]
b_d integer [0, 0, 0, 1]
>
Or all string columns:
iex> df = Explorer.DataFrame.new(num: [1, 2, 3, 4], b: ["b", "a", "b", "d"])
iex> Explorer.DataFrame.dummies(df, fn _name, type -> type == :string end)
#Explorer.DataFrame<
Polars[4 x 3]
b_a integer [0, 1, 0, 0]
b_b integer [1, 0, 1, 0]
b_d integer [0, 0, 0, 1]
>
"""
@doc type: :single
@spec dummies(df :: DataFrame.t(), column() | columns()) ::
DataFrame.t()
def dummies(df, columns_or_column)
def dummies(df, column) when is_column(column), do: dummies(df, [column])
def dummies(df, columns),
do: Shared.apply_impl(df, :dummies, [to_existing_columns(df, columns)])
@doc """
Extracts a single column as a series.
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.pull(df, "total")
#Explorer.Series<
integer[1094]
[2308, 1254, 32500, 141, 7924, 41, 143, 51246, 1150, 684, 106589, 18408, 8366, 451, 7981, 16345, 403, 17192, 30222, 147, 1388, 166, 133, 5802, 1278, 114468, 47, 2237, 12030, 535, 58, 1367, 145806, 152, 152, 72, 141, 19703, 2393248, 20773, 44, 540, 19, 2064, 1900, 5501, 10465, 2102, 30428, 18122, ...]
>
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.pull(df, 2)
#Explorer.Series<
integer[1094]
[2308, 1254, 32500, 141, 7924, 41, 143, 51246, 1150, 684, 106589, 18408, 8366, 451, 7981, 16345, 403, 17192, 30222, 147, 1388, 166, 133, 5802, 1278, 114468, 47, 2237, 12030, 535, 58, 1367, 145806, 152, 152, 72, 141, 19703, 2393248, 20773, 44, 540, 19, 2064, 1900, 5501, 10465, 2102, 30428, 18122, ...]
>
"""
@doc type: :single
@spec pull(df :: DataFrame.t(), column :: column()) :: Series.t()
def pull(df, column) when is_column(column) do
[column] = to_existing_columns(df, [column])
Shared.apply_impl(df, :pull, [column])
end
@doc """
Subset a continuous set of rows.
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.slice(df, 1, 2)
#Explorer.DataFrame<
Polars[2 x 10]
year integer [2010, 2010]
country string ["ALBANIA", "ALGERIA"]
total integer [1254, 32500]
solid_fuel integer [117, 332]
liquid_fuel integer [953, 12381]
gas_fuel integer [7, 14565]
cement integer [177, 2598]
gas_flaring integer [0, 2623]
per_capita float [0.43, 0.9]
bunker_fuels integer [7, 663]
>
Negative offsets count from the end of the series:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.slice(df, -10, 2)
#Explorer.DataFrame<
Polars[2 x 10]
year integer [2014, 2014]
country string ["UNITED STATES OF AMERICA", "URUGUAY"]
total integer [1432855, 1840]
solid_fuel integer [450047, 2]
liquid_fuel integer [576531, 1700]
gas_fuel integer [390719, 25]
cement integer [11314, 112]
gas_flaring integer [4244, 0]
per_capita float [4.43, 0.54]
bunker_fuels integer [30722, 251]
>
If the length would run past the end of the dataframe, the result may be shorter than the length:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.slice(df, -10, 20)
#Explorer.DataFrame<
Polars[10 x 10]
year integer [2014, 2014, 2014, 2014, 2014, ...]
country string ["UNITED STATES OF AMERICA", "URUGUAY", "UZBEKISTAN", "VANUATU", "VENEZUELA", ...]
total integer [1432855, 1840, 28692, 42, 50510, ...]
solid_fuel integer [450047, 2, 1677, 0, 204, ...]
liquid_fuel integer [576531, 1700, 2086, 42, 28445, ...]
gas_fuel integer [390719, 25, 23929, 0, 12731, ...]
cement integer [11314, 112, 1000, 0, 1088, ...]
gas_flaring integer [4244, 0, 0, 0, 8042, ...]
per_capita float [4.43, 0.54, 0.97, 0.16, 1.65, ...]
bunker_fuels integer [30722, 251, 0, 10, 1256, ...]
>
"""
@doc type: :rows
def slice(df, offset, length), do: Shared.apply_impl(df, :slice, [offset, length])
@doc """
Subset rows with a list of indices or a range.
## Examples
iex> df = Explorer.DataFrame.new(a: [1, 2, 3], b: ["a", "b", "c"])
iex> Explorer.DataFrame.slice(df, [0, 2])
#Explorer.DataFrame<
Polars[2 x 2]
a integer [1, 3]
b string ["a", "c"]
>
iex> df = Explorer.DataFrame.new(a: [1, 2, 3], b: ["a", "b", "c"])
iex> Explorer.DataFrame.slice(df, 1..2)
#Explorer.DataFrame<
Polars[2 x 2]
a integer [2, 3]
b string ["b", "c"]
>
"""
@doc type: :rows
def slice(df, row_indices) when is_list(row_indices) do
n_rows = n_rows(df)
Enum.each(row_indices, fn idx ->
if idx > n_rows or idx < -n_rows,
do:
raise(
ArgumentError,
"requested row index (#{idx}) out of bounds (-#{n_rows}:#{n_rows})"
)
end)
Shared.apply_impl(df, :slice, [row_indices])
end
def slice(df, %Range{} = range) do
slice(df, Enum.slice(0..(n_rows(df) - 1)//1, range))
end
@doc """
Sample rows from a dataframe.
If given an integer as the second argument, it will return N samples. If given a float, it will
return that proportion of the series.
Can sample with or without replacement.
## Options
* `replacement` - If set to `true`, each sample will be independent and therefore values may repeat.
Required to be `true` for `n` greater then the number of rows in the dataframe or `frac` > 1.0. (default: `false`)
* `seed` - An integer to be used as a random seed. If nil, a random value between 1 and 1e12 will be used. (default: nil)
## Examples
You can sample N rows:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.sample(df, 3, seed: 100)
#Explorer.DataFrame<
Polars[3 x 10]
year integer [2012, 2012, 2013]
country string ["ZIMBABWE", "NICARAGUA", "NIGER"]
total integer [2125, 1260, 529]
solid_fuel integer [917, 0, 93]
liquid_fuel integer [1006, 1176, 432]
gas_fuel integer [0, 0, 0]
cement integer [201, 84, 4]
gas_flaring integer [0, 0, 0]
per_capita float [0.15, 0.21, 0.03]
bunker_fuels integer [9, 18, 19]
>
Or you can sample a proportion of rows:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.sample(df, 0.03, seed: 100)
#Explorer.DataFrame<
Polars[33 x 10]
year integer [2013, 2012, 2013, 2012, 2010, ...]
country string ["BAHAMAS", "POLAND", "SLOVAKIA", "MOZAMBIQUE", "OMAN", ...]
total integer [764, 81792, 9024, 851, 12931, ...]
solid_fuel integer [1, 53724, 3657, 11, 0, ...]
liquid_fuel integer [763, 17353, 2090, 632, 2331, ...]
gas_fuel integer [0, 8544, 2847, 47, 9309, ...]
cement integer [0, 2165, 424, 161, 612, ...]
gas_flaring integer [0, 6, 7, 0, 679, ...]
per_capita float [2.02, 2.12, 1.67, 0.03, 4.39, ...]
bunker_fuels integer [167, 573, 34, 56, 1342, ...]
>
"""
@doc type: :rows
@spec sample(df :: DataFrame.t(), n_or_frac :: number(), opts :: Keyword.t()) :: DataFrame.t()
def sample(df, n_or_frac, opts \\ [])
def sample(df, n, opts) when is_integer(n) do
opts = Keyword.validate!(opts, replacement: false, seed: Enum.random(1..1_000_000_000_000))
n_rows = n_rows(df)
case {n > n_rows, opts[:replacement]} do
{true, false} ->
raise ArgumentError,
"in order to sample more rows than are in the dataframe (#{n_rows}), sampling " <>
"`replacement` must be true"
_ ->
:ok
end
Shared.apply_impl(df, :sample, [n, opts[:replacement], opts[:seed]])
end
def sample(df, frac, opts) when is_float(frac) do
n_rows = n_rows(df)
n = round(frac * n_rows)
sample(df, n, opts)
end
@doc """
Pivot data from wide to long.
`Explorer.DataFrame.pivot_longer/3` "lengthens" data, increasing the number of rows and
decreasing the number of columns. The inverse transformation is
`Explorer.DataFrame.pivot_wider/4`.
The second argument (`columns_to_pivot`) can be either an array of column names to pivot
or a filter callback on the dataframe's names. These columns must always have the same
data type.
## Options
* `keep` - Columns that are not in the list of pivot and should be kept in the dataframe.
May be a filter callback on the dataframe's column names.
Defaults to all columns except the ones to pivot.
* `drop` - Columns that are not in the list of pivot and should be dropped from the dataframe.
May be a filter callback on the dataframe's column names. This list of columns is going to be
subtracted from the list of `keep`.
Defaults to an empty list.
* `names_to` - A string specifying the name of the column to create from the data stored
in the column names of the dataframe. Defaults to `"variable"`.
* `values_to` - A string specifying the name of the column to create from the data stored
in series element values. Defaults to `"value"`.
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.pivot_longer(df, &String.ends_with?(&1, "fuel"))
#Explorer.DataFrame<
Polars[3282 x 9]
year integer [2010, 2010, 2010, 2010, 2010, ...]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
total integer [2308, 1254, 32500, 141, 7924, ...]
cement integer [5, 177, 2598, 0, 204, ...]
gas_flaring integer [0, 0, 2623, 0, 3697, ...]
per_capita float [0.08, 0.43, 0.9, 1.68, 0.37, ...]
bunker_fuels integer [9, 7, 663, 0, 321, ...]
variable string ["solid_fuel", "solid_fuel", "solid_fuel", "solid_fuel", "solid_fuel", ...]
value integer [627, 117, 332, 0, 0, ...]
>
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.pivot_longer(df, &String.ends_with?(&1, "fuel"), keep: ["year", "country"])
#Explorer.DataFrame<
Polars[3282 x 4]
year integer [2010, 2010, 2010, 2010, 2010, ...]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
variable string ["solid_fuel", "solid_fuel", "solid_fuel", "solid_fuel", "solid_fuel", ...]
value integer [627, 117, 332, 0, 0, ...]
>
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.pivot_longer(df, ["total"], keep: ["year", "country"], drop: ["country"])
#Explorer.DataFrame<
Polars[1094 x 3]
year integer [2010, 2010, 2010, 2010, 2010, ...]
variable string ["total", "total", "total", "total", "total", ...]
value integer [2308, 1254, 32500, 141, 7924, ...]
>
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.pivot_longer(df, ["total"], keep: [], names_to: "my_var", values_to: "my_value")
#Explorer.DataFrame<
Polars[1094 x 2]
my_var string ["total", "total", "total", "total", "total", ...]
my_value integer [2308, 1254, 32500, 141, 7924, ...]
>
"""
@doc type: :single
@spec pivot_longer(
df :: DataFrame.t(),
columns_to_pivot :: columns(),
opts :: Keyword.t()
) :: DataFrame.t()
def pivot_longer(df, columns_to_pivot, opts \\ [])
def pivot_longer(df, columns_to_pivot, opts) do
opts =
Keyword.validate!(opts,
keep: 0..-1//1,
drop: [],
names_to: "variable",
values_to: "value"
)
names_to = to_column_name(opts[:names_to])
values_to = to_column_name(opts[:values_to])
columns_to_pivot = to_existing_columns(df, columns_to_pivot)
dtypes = df.dtypes
columns_to_keep =
case opts[:keep] do
keep when is_list(keep) ->
Enum.each(keep, fn column ->
if column in columns_to_pivot do
raise ArgumentError,
"columns to keep must not include columns to pivot, but found #{inspect(column)} in both"
end
end)
to_existing_columns(df, keep)
keep ->
to_existing_columns(df, keep)
end
columns_to_keep =
(columns_to_keep -- columns_to_pivot) -- to_existing_columns(df, opts[:drop])
values_dtype =
dtypes
|> Map.take(columns_to_pivot)
|> Map.values()
|> Enum.uniq()
|> case do
[dtype] ->
dtype
[_ | _] = dtypes ->
raise ArgumentError,
"columns to pivot must include columns with the same dtype, but found multiple dtypes: #{inspect(dtypes)}"
end
new_dtypes =
dtypes
|> Map.take(columns_to_keep)
|> Map.put(names_to, :string)
|> Map.put(values_to, values_dtype)
out_df = %{df | names: columns_to_keep ++ [names_to, values_to], dtypes: new_dtypes}
args = [out_df, columns_to_pivot, columns_to_keep, names_to, values_to]
Shared.apply_impl(df, :pivot_longer, args)
end
@doc """
Pivot data from long to wide.
`Explorer.DataFrame.pivot_wider/4` "widens" data, increasing the number of columns and
decreasing the number of rows. The inverse transformation is
`Explorer.DataFrame.pivot_longer/3`.
Due to a restriction upstream, `values_from` must be a numeric type.
## Options
* `id_columns` - A set of columns that uniquely identifies each observation.
Defaults to all columns in data except for the columns specified in `names_from` and `values_from`.
Typically used when you have redundant variables, i.e. variables whose values are perfectly correlated
with existing variables. May accept a filter callback, a list or a range of column names.
Default value is `0..-1`. If an empty list is passed, or a range that results in a empty list of
column names, it raises an error.
ID columns cannot be of the float type and attempting so will raise an error.
If you need to use float columns as IDs, you must carefully consider rounding
or truncating the column and converting it to integer, as long as doing so
preserves the properties of the column.
* `names_prefix` - String added to the start of every variable name.
This is particularly useful if `names_from` is a numeric vector and you want to create syntactic variable names.
## Examples
iex> df = Explorer.DataFrame.new(id: [1, 1], variable: ["a", "b"], value: [1, 2])
iex> Explorer.DataFrame.pivot_wider(df, "variable", "value")
#Explorer.DataFrame<
Polars[1 x 3]
id integer [1]
a integer [1]
b integer [2]
>
"""
@doc type: :single
@spec pivot_wider(
df :: DataFrame.t(),
names_from :: column(),
values_from :: column(),
opts ::
Keyword.t()
) :: DataFrame.t()
def pivot_wider(df, names_from, values_from, opts \\ []) do
opts = Keyword.validate!(opts, id_columns: 0..-1, names_prefix: "")
[values_from, names_from] = to_existing_columns(df, [values_from, names_from])
dtypes = df.dtypes
unless dtypes[values_from] in [:integer, :float, :date, :datetime] do
raise ArgumentError,
"the values_from column must be numeric, but found #{dtypes[values_from]}"
end
id_columns = to_existing_columns(df, opts[:id_columns]) -- [names_from, values_from]
if id_columns == [] do
raise ArgumentError,
"id_columns must select at least one existing column, but #{inspect(opts[:id_columns])} selects none."
end
for column_name <- id_columns do
if df.dtypes[column_name] == :float do
raise ArgumentError,
"id_columns cannot have columns of the type float, but #{inspect(column_name)} column is float."
end
end
Shared.apply_impl(df, :pivot_wider, [id_columns, names_from, values_from, opts[:names_prefix]])
end
# Two table verbs
@valid_join_types [:inner, :left, :right, :outer, :cross]
@doc """
Join two tables.
## Join types
* `inner` - Returns all rows from `left` where there are matching values in `right`, and all columns from `left` and `right`.
* `left` - Returns all rows from `left` and all columns from `left` and `right`. Rows in `left` with no match in `right` will have `nil` values in the new columns.
* `right` - Returns all rows from `right` and all columns from `left` and `right`. Rows in `right` with no match in `left` will have `nil` values in the new columns.
* `outer` - Returns all rows and all columns from both `left` and `right`. Where there are not matching values, returns `nil` for the one missing.
* `cross` - Also known as a cartesian join. Returns all combinations of `left` and `right`. Can be very computationally expensive.
## Options
* `on` - The columns to join on. Defaults to overlapping columns. Does not apply to cross join.
* `how` - One of the join types (as an atom) described above. Defaults to `:inner`.
## Examples
Inner join:
iex> left = Explorer.DataFrame.new(a: [1, 2, 3], b: ["a", "b", "c"])
iex> right = Explorer.DataFrame.new(a: [1, 2, 2], c: ["d", "e", "f"])
iex> Explorer.DataFrame.join(left, right)
#Explorer.DataFrame<
Polars[3 x 3]
a integer [1, 2, 2]
b string ["a", "b", "b"]
c string ["d", "e", "f"]
>
Left join:
iex> left = Explorer.DataFrame.new(a: [1, 2, 3], b: ["a", "b", "c"])
iex> right = Explorer.DataFrame.new(a: [1, 2, 2], c: ["d", "e", "f"])
iex> Explorer.DataFrame.join(left, right, how: :left)
#Explorer.DataFrame<
Polars[4 x 3]
a integer [1, 2, 2, 3]
b string ["a", "b", "b", "c"]
c string ["d", "e", "f", nil]
>
Right join:
iex> left = Explorer.DataFrame.new(a: [1, 2, 3], b: ["a", "b", "c"])
iex> right = Explorer.DataFrame.new(a: [1, 2, 4], c: ["d", "e", "f"])
iex> Explorer.DataFrame.join(left, right, how: :right)
#Explorer.DataFrame<
Polars[3 x 3]
a integer [1, 2, 4]
c string ["d", "e", "f"]
b string ["a", "b", nil]
>
Outer join:
iex> left = Explorer.DataFrame.new(a: [1, 2, 3], b: ["a", "b", "c"])
iex> right = Explorer.DataFrame.new(a: [1, 2, 4], c: ["d", "e", "f"])
iex> Explorer.DataFrame.join(left, right, how: :outer)
#Explorer.DataFrame<
Polars[4 x 3]
a integer [1, 2, 4, 3]
b string ["a", "b", nil, "c"]
c string ["d", "e", "f", nil]
>
Cross join:
iex> left = Explorer.DataFrame.new(a: [1, 2, 3], b: ["a", "b", "c"])
iex> right = Explorer.DataFrame.new(a: [1, 2, 4], c: ["d", "e", "f"])
iex> Explorer.DataFrame.join(left, right, how: :cross)
#Explorer.DataFrame<
Polars[9 x 4]
a integer [1, 1, 1, 2, 2, ...]
b string ["a", "a", "a", "b", "b", ...]
a_right integer [1, 2, 4, 1, 2, ...]
c string ["d", "e", "f", "d", "e", ...]
>
Inner join with different names:
iex> left = Explorer.DataFrame.new(a: [1, 2, 3], b: ["a", "b", "c"])
iex> right = Explorer.DataFrame.new(d: [1, 2, 2], c: ["d", "e", "f"])
iex> Explorer.DataFrame.join(left, right, on: [{"a", "d"}])
#Explorer.DataFrame<
Polars[3 x 3]
a integer [1, 2, 2]
b string ["a", "b", "b"]
c string ["d", "e", "f"]
>
"""
@doc type: :multi
@spec join(left :: DataFrame.t(), right :: DataFrame.t(), opts :: Keyword.t()) :: DataFrame.t()
def join(%DataFrame{} = left, %DataFrame{} = right, opts \\ []) do
left_columns = left.names
right_columns = right.names
opts =
Keyword.validate!(opts,
on: find_overlapping_columns(left_columns, right_columns),
how: :inner
)
unless opts[:how] in @valid_join_types do
raise ArgumentError,
"join type is not valid: #{inspect(opts[:how])}. " <>
"Valid options are: #{Enum.map_join(@valid_join_types, ", ", &inspect/1)}"
end
{on, how} =
case {opts[:on], opts[:how]} do
{[], :cross} ->
{[], :cross}
{[], _} ->
raise(ArgumentError, "could not find any overlapping columns")
{[_ | _] = on, how} ->
normalized_on =
Enum.map(on, fn
{l_name, r_name} ->
[l_column] = to_existing_columns(left, [l_name])
[r_column] = to_existing_columns(right, [r_name])
{l_column, r_column}
name ->
[l_column] = to_existing_columns(left, [name])
[r_column] = to_existing_columns(right, [name])
# This is an edge case for when an index is passed as column selection
if l_column != r_column do
raise ArgumentError,
"the column given to option `:on` is not the same for both dataframes"
end
{l_column, r_column}
end)
{normalized_on, how}
end
out_df = out_df_for_join(how, left, right, on)
Shared.apply_impl(left, :join, [right, out_df, on, how])
end
defp find_overlapping_columns(left_columns, right_columns) do
left_columns = MapSet.new(left_columns)
right_columns = MapSet.new(right_columns)
left_columns |> MapSet.intersection(right_columns) |> MapSet.to_list()
end
defp out_df_for_join(:right, left, right, on) do
{left_on, _right_on} = Enum.unzip(on)
pairs = dtypes_pairs_for_common_join(right, left, left_on, "_left")
{new_names, _} = Enum.unzip(pairs)
%{right | names: new_names, dtypes: Map.new(pairs)}
end
defp out_df_for_join(how, left, right, on) do
{_left_on, right_on} = Enum.unzip(on)
right_on = if how == :cross, do: [], else: right_on
pairs = dtypes_pairs_for_common_join(left, right, right_on)
{new_names, _} = Enum.unzip(pairs)
%{left | names: new_names, dtypes: Map.new(pairs)}
end
defp dtypes_pairs_for_common_join(left, right, right_on, suffix \\ "_right") do
Enum.map(left.names, fn name -> {name, left.dtypes[name]} end) ++
Enum.map(right.names -- right_on, fn right_name ->
name =
if right_name in left.names do
right_name <> suffix
else
right_name
end
{name, right.dtypes[right_name]}
end)
end
@doc """
Combine two or more dataframes column-wise.
Dataframes must have the same number of rows.
## Examples
iex> df1 = Explorer.DataFrame.new(x: [1, 2, 3], y: ["a", "b", "c"])
iex> df2 = Explorer.DataFrame.new(z: [4, 5, 6], a: ["d", "e", "f"])
iex> Explorer.DataFrame.concat_columns([df1, df2])
#Explorer.DataFrame<
Polars[3 x 4]
x integer [1, 2, 3]
y string ["a", "b", "c"]
z integer [4, 5, 6]
a string ["d", "e", "f"]
>
Conflicting names are suffixed with the index of the dataframe in the array:
iex> df1 = Explorer.DataFrame.new(x: [1, 2, 3], y: ["a", "b", "c"])
iex> df2 = Explorer.DataFrame.new(x: [4, 5, 6], a: ["d", "e", "f"])
iex> Explorer.DataFrame.concat_columns([df1, df2])
#Explorer.DataFrame<
Polars[3 x 4]
x integer [1, 2, 3]
y string ["a", "b", "c"]
x_1 integer [4, 5, 6]
a string ["d", "e", "f"]
>
"""
@doc type: :multi
@spec concat_columns([DataFrame.t()]) :: DataFrame.t()
def concat_columns([%DataFrame{} = head | tail] = dfs) do
n_rows = n_rows(head)
if Enum.all?(tail, &(n_rows(&1) == n_rows)) do
Shared.apply_impl(dfs, :concat_columns)
else
raise ArgumentError, "all dataframes must have the same number of rows"
end
end
@doc """
Combine two dataframes column-wise.
`concat_columns(df1, df2)` is equivalent to `concat_columns([df1, df2])`.
"""
@doc type: :multi
@spec concat_columns(DataFrame.t(), DataFrame.t()) :: DataFrame.t()
def concat_columns(%DataFrame{} = df1, %DataFrame{} = df2), do: concat_columns([df1, df2])
def concat_columns(%DataFrame{} = df, [%DataFrame{} | _] = dfs), do: concat_columns([df | dfs])
@doc """
Combine two or more dataframes row-wise (stack).
Column names and dtypes must match. The only exception is for numeric
columns that can be mixed together, and casted automatically to float columns.
## Examples
iex> df1 = Explorer.DataFrame.new(x: [1, 2, 3], y: ["a", "b", "c"])
iex> df2 = Explorer.DataFrame.new(x: [4, 5, 6], y: ["d", "e", "f"])
iex> Explorer.DataFrame.concat_rows([df1, df2])
#Explorer.DataFrame<
Polars[6 x 2]
x integer [1, 2, 3, 4, 5, ...]
y string ["a", "b", "c", "d", "e", ...]
>
iex> df1 = Explorer.DataFrame.new(x: [1, 2, 3], y: ["a", "b", "c"])
iex> df2 = Explorer.DataFrame.new(x: [4.2, 5.3, 6.4], y: ["d", "e", "f"])
iex> Explorer.DataFrame.concat_rows([df1, df2])
#Explorer.DataFrame<
Polars[6 x 2]
x float [1.0, 2.0, 3.0, 4.2, 5.3, ...]
y string ["a", "b", "c", "d", "e", ...]
>
"""
@doc type: :multi
@spec concat_rows([DataFrame.t()]) :: DataFrame.t()
def concat_rows([%DataFrame{} | _t] = dfs) do
changed_types = compute_changed_types_concat_rows(dfs)
if Enum.empty?(changed_types) do
Shared.apply_impl(dfs, :concat_rows)
else
dfs
|> cast_numeric_columns_to_float(changed_types)
|> Shared.apply_impl(:concat_rows)
end
end
defp compute_changed_types_concat_rows([head | tail]) do
types = head.dtypes
Enum.reduce(tail, %{}, fn df, changed_types ->
if n_columns(df) != map_size(types) do
raise ArgumentError,
"dataframes must have the same columns"
end
Enum.reduce(df.dtypes, changed_types, fn {name, type}, changed_types ->
cond do
not Map.has_key?(types, name) ->
raise ArgumentError,
"dataframes must have the same columns"
types[name] == type ->
changed_types
types_are_numeric_compatible?(types, name, type) ->
Map.put(changed_types, name, :float)
true ->
raise ArgumentError,
"columns and dtypes must be identical for all dataframes"
end
end)
end)
end
defp types_are_numeric_compatible?(types, name, type) do
numeric_types = [:float, :integer]
types[name] != type and types[name] in numeric_types and type in numeric_types
end
defp cast_numeric_columns_to_float(dfs, changed_types) do
for df <- dfs do
columns =
for {name, :integer} <- df.dtypes,
changed_types[name] == :float,
do: name
if Enum.empty?(columns) do
df
else
changes = for column <- columns, into: %{}, do: {column, Series.cast(df[column], :float)}
mutate(df, changes)
end
end
end
@doc """
Combine two dataframes row-wise.
`concat_rows(df1, df2)` is equivalent to `concat_rows([df1, df2])`.
"""
@doc type: :multi
@spec concat_rows(DataFrame.t(), DataFrame.t()) :: DataFrame.t()
def concat_rows(%DataFrame{} = df1, %DataFrame{} = df2), do: concat_rows([df1, df2])
def concat_rows(%DataFrame{} = df, [%DataFrame{} | _] = dfs), do: concat_rows([df | dfs])
# Groups
@doc """
Group the dataframe by one or more variables.
When the dataframe has grouping variables, operations are performed per group.
`Explorer.DataFrame.ungroup/2` removes grouping.
## Examples
You can group by a single variable:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.group_by(df, "country")
#Explorer.DataFrame<
Polars[1094 x 10]
Groups: ["country"]
year integer [2010, 2010, 2010, 2010, 2010, ...]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
total integer [2308, 1254, 32500, 141, 7924, ...]
solid_fuel integer [627, 117, 332, 0, 0, ...]
liquid_fuel integer [1601, 953, 12381, 141, 3649, ...]
gas_fuel integer [74, 7, 14565, 0, 374, ...]
cement integer [5, 177, 2598, 0, 204, ...]
gas_flaring integer [0, 0, 2623, 0, 3697, ...]
per_capita float [0.08, 0.43, 0.9, 1.68, 0.37, ...]
bunker_fuels integer [9, 7, 663, 0, 321, ...]
>
Or you can group by multiple:
iex> df = Explorer.Datasets.fossil_fuels()
iex> Explorer.DataFrame.group_by(df, ["country", "year"])
#Explorer.DataFrame<
Polars[1094 x 10]
Groups: ["country", "year"]
year integer [2010, 2010, 2010, 2010, 2010, ...]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
total integer [2308, 1254, 32500, 141, 7924, ...]
solid_fuel integer [627, 117, 332, 0, 0, ...]
liquid_fuel integer [1601, 953, 12381, 141, 3649, ...]
gas_fuel integer [74, 7, 14565, 0, 374, ...]
cement integer [5, 177, 2598, 0, 204, ...]
gas_flaring integer [0, 0, 2623, 0, 3697, ...]
per_capita float [0.08, 0.43, 0.9, 1.68, 0.37, ...]
bunker_fuels integer [9, 7, 663, 0, 321, ...]
>
"""
@doc type: :single
@spec group_by(df :: DataFrame.t(), groups_or_group :: column_names() | column_name()) ::
DataFrame.t()
def group_by(df, groups) when is_list(groups) do
groups = to_existing_columns(df, groups)
all_groups = Enum.uniq(df.groups ++ groups)
%{df | groups: all_groups}
end
def group_by(df, group) when is_column_name(group), do: group_by(df, [group])
@doc """
Removes grouping variables.
Accepts a list of group names. If groups is not specified, then all groups are
removed.
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> df = Explorer.DataFrame.group_by(df, ["country", "year"])
iex> Explorer.DataFrame.ungroup(df, ["country"])
#Explorer.DataFrame<
Polars[1094 x 10]
Groups: ["year"]
year integer [2010, 2010, 2010, 2010, 2010, ...]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
total integer [2308, 1254, 32500, 141, 7924, ...]
solid_fuel integer [627, 117, 332, 0, 0, ...]
liquid_fuel integer [1601, 953, 12381, 141, 3649, ...]
gas_fuel integer [74, 7, 14565, 0, 374, ...]
cement integer [5, 177, 2598, 0, 204, ...]
gas_flaring integer [0, 0, 2623, 0, 3697, ...]
per_capita float [0.08, 0.43, 0.9, 1.68, 0.37, ...]
bunker_fuels integer [9, 7, 663, 0, 321, ...]
>
iex> df = Explorer.Datasets.fossil_fuels()
iex> df = Explorer.DataFrame.group_by(df, ["country", "year"])
iex> Explorer.DataFrame.ungroup(df)
#Explorer.DataFrame<
Polars[1094 x 10]
year integer [2010, 2010, 2010, 2010, 2010, ...]
country string ["AFGHANISTAN", "ALBANIA", "ALGERIA", "ANDORRA", "ANGOLA", ...]
total integer [2308, 1254, 32500, 141, 7924, ...]
solid_fuel integer [627, 117, 332, 0, 0, ...]
liquid_fuel integer [1601, 953, 12381, 141, 3649, ...]
gas_fuel integer [74, 7, 14565, 0, 374, ...]
cement integer [5, 177, 2598, 0, 204, ...]
gas_flaring integer [0, 0, 2623, 0, 3697, ...]
per_capita float [0.08, 0.43, 0.9, 1.68, 0.37, ...]
bunker_fuels integer [9, 7, 663, 0, 321, ...]
>
"""
@doc type: :single
@spec ungroup(df :: DataFrame.t(), groups_or_group :: column_names() | column_name() | :all) ::
DataFrame.t()
def ungroup(df, groups \\ :all)
def ungroup(df, :all), do: %{df | groups: []}
def ungroup(df, groups) when is_list(groups) do
current_groups = groups(df)
groups = to_existing_columns(df, groups)
Enum.each(groups, fn group ->
if group not in current_groups,
do:
raise(
ArgumentError,
"could not find #{group} in current groups (#{current_groups})"
)
end)
%{df | groups: current_groups -- groups}
end
def ungroup(df, group) when is_column_name(group), do: ungroup(df, [group])
@supported_aggs ~w[min max sum mean median first last count n_distinct]a
@doc """
Summarise each group to a single row.
Implicitly ungroups.
## Supported operations
The following aggregations may be performed:
* `:min` - Take the minimum value within the group. See `Explorer.Series.min/1`.
* `:max` - Take the maximum value within the group. See `Explorer.Series.max/1`.
* `:sum` - Take the sum of the series within the group. See `Explorer.Series.sum/1`.
* `:mean` - Take the mean of the series within the group. See `Explorer.Series.mean/1`.
* `:median` - Take the median of the series within the group. See `Explorer.Series.median/1`.
* `:first` - Take the first value within the group. See `Explorer.Series.first/1`.
* `:last` - Take the last value within the group. See `Explorer.Series.last/1`.
* `:count` - Count the number of rows per group.
* `:n_distinct` - Count the number of unique rows per group.
## Examples
iex> df = Explorer.Datasets.fossil_fuels()
iex> df |> Explorer.DataFrame.group_by("year") |> Explorer.DataFrame.summarise(total: [:max, :min], country: [:n_distinct])
#Explorer.DataFrame<
Polars[5 x 4]
year integer [2010, 2011, 2012, 2013, 2014]
total_max integer [2393248, 2654360, 2734817, 2797384, 2806634]
total_min integer [1, 2, 2, 2, 3]
country_n_distinct integer [217, 217, 220, 220, 220]
>
"""
@doc type: :single
@spec summarise(df :: DataFrame.t(), columns :: Keyword.t() | map()) :: DataFrame.t()
def summarise(%DataFrame{groups: []}, _),
do:
raise(
ArgumentError,
"dataframe must be grouped in order to perform summarisation"
)
def summarise(df, columns) when is_column_pairs(columns) do
column_pairs =
to_column_pairs(df, columns, fn values ->
case values -- @supported_aggs do
[] ->
values
unsupported ->
raise ArgumentError, "found unsupported aggregations #{inspect(unsupported)}"
end
end)
new_dtypes = names_with_dtypes_for_summarise(df, column_pairs)
new_names = for {name, _} <- new_dtypes, do: name
df_out = %{df | names: new_names, dtypes: Map.new(new_dtypes), groups: []}
Shared.apply_impl(df, :summarise, [df_out, Map.new(column_pairs)])
end
defp names_with_dtypes_for_summarise(df, column_pairs) do
groups = for group <- df.groups, do: {group, df.dtypes[group]}
agg_pairs =
for {column_name, aggregations} <- column_pairs, agg <- aggregations do
name = "#{column_name}_#{agg}"
dtype =
case agg do
:median ->
:float
:mean ->
:float
agg when agg in [:count, :n_distinct] ->
:integer
_other ->
df.dtypes[column_name]
end
{name, dtype}
end
groups ++ agg_pairs
end
@doc """
Summarise each group to a single row using a callback function.
Implicitly ungroups.
The main difference between `summarise/2` and `summarise_with/2` is that the later
accepts a function that can be used to perform complex operations.
This is efficient because it doesn't need
to create intermediate series representations to summarise.
## Supported operations
The function callback should be in the form of `[name_of_col: "operation"]`,
where `"operation"` is one of the `Explorer.Series` functions. It's required
that at least one of the following functions is used for summarisation:
* `Explorer.Series.min/1` - Take the minimum value within the group.
* `Explorer.Series.max/1` - Take the maximum value within the group.
* `Explorer.Series.sum/1` - Take the sum of the series within the group.
* `Explorer.Series.mean/1` - Take the mean of the series within the group.
* `Explorer.Series.median/1` - Take the median of the series within the group.
* `Explorer.Series.first/1` - Take the first value within the group.
* `Explorer.Series.last/1` - Take the last value within the group.
* `Explorer.Series.count/1` - Count the number of rows per group.
* `Explorer.Series.n_distinct/1` - Count the number of unique rows per group.
## Examples
iex> alias Explorer.{DataFrame, Series}
iex> df = Explorer.Datasets.fossil_fuels() |> DataFrame.group_by("year")
iex> DataFrame.summarise_with(df, &[total_max: Series.max(&1["total"]), countries: Series.n_distinct(&1["country"])])
#Explorer.DataFrame<
Polars[5 x 3]
year integer [2010, 2011, 2012, 2013, 2014]
total_max integer [2393248, 2654360, 2734817, 2797384, 2806634]
countries integer [217, 217, 220, 220, 220]
>
"""
@doc type: :single
@spec summarise_with(
df :: DataFrame.t(),
callback :: (Explorer.Backend.LazyFrame.t() -> column_pairs(Series.lazy_t()))
) :: DataFrame.t()
def summarise_with(%DataFrame{groups: []}, _),
do:
raise(
ArgumentError,
"dataframe must be grouped in order to perform summarisation"
)
def summarise_with(%DataFrame{} = df, fun) when is_function(fun, 1) do
ldf = to_opaque_lazy(df)
result = fun.(ldf)
column_pairs =
to_column_pairs(df, result, fn value ->
case value do
%Series{data: %LazySeries{aggregation: true}} ->
value
%Series{data: %LazySeries{op: op}} ->
raise "expecting summarise with an aggregation operation inside. But instead got #{inspect(op)}."
other ->
raise "expecting a lazy series, but instead got #{inspect(other)}"
end
end)
new_dtypes = names_with_dtypes_for_column_pairs(df, column_pairs)
new_names = for {name, _} <- new_dtypes, do: name
df_out = %{df | names: new_names, dtypes: Map.new(new_dtypes), groups: []}
column_pairs = for {name, %Series{data: lazy_series}} <- column_pairs, do: {name, lazy_series}
Shared.apply_impl(df, :summarise_with, [df_out, column_pairs])
end
defp names_with_dtypes_for_column_pairs(df, column_pairs) do
groups = for group <- df.groups, do: {group, df.dtypes[group]}
names_with_dtypes =
for {column_name, series} <- column_pairs do
{column_name, series.dtype}
end
groups ++ names_with_dtypes
end
@doc """
Display the DataFrame in a tabular fashion.
## Examples
df = Explorer.Datasets.iris()
Explorer.DataFrame.table(df)
Explorer.DataFrame.table(df, limit: 1)
Explorer.DataFrame.table(df, limit: :infinity)
"""
@doc type: :single
@spec table(df :: DataFrame.t(), opts :: Keyword.t()) :: :ok
def table(df, opts \\ []) do
{rows, columns} = shape(df)
headers = df.names
df =
case opts[:limit] do
:infinity -> df
nrow when is_integer(nrow) and nrow >= 0 -> slice(df, 0, nrow)
_ -> slice(df, 0, @default_sample_nrows)
end
types = Enum.map(df.names, &"\n<#{Atom.to_string(df.dtypes[&1])}>")
values =
headers
|> Enum.map(&Series.to_list(df[&1]))
|> Enum.zip_with(& &1)
name_type = Enum.zip_with(headers, types, fn x, y -> x <> y end)
TableRex.Table.new()
|> TableRex.Table.put_title("Explorer DataFrame: [rows: #{rows}, columns: #{columns}]")
|> TableRex.Table.put_header(name_type)
|> TableRex.Table.put_header_meta(0..columns, align: :center)
|> TableRex.Table.add_rows(values)
|> TableRex.Table.render!(
header_separator_symbol: "=",
horizontal_style: :all
)
|> IO.puts()
end
# Helpers
defp backend_from_options!(opts) do
backend = Explorer.Shared.backend_from_options!(opts) || Explorer.Backend.get()
:"#{backend}.DataFrame"
end
defp maybe_raise_column_not_found(df, name) do
unless Map.has_key?(df.dtypes, name) do
raise ArgumentError,
List.to_string(
["could not find column name \"#{name}\""] ++ did_you_mean(name, df.names)
)
end
end
@threshold 0.77
@max_suggestions 5
defp did_you_mean(missing_key, available_keys) do
suggestions =
for key <- available_keys,
distance = String.jaro_distance(missing_key, key),
distance >= @threshold,
do: {distance, key}
case suggestions do
[] -> []
suggestions -> [". Did you mean:\n\n" | format_suggestions(suggestions)]
end
end
defp format_suggestions(suggestions) do
suggestions
|> Enum.sort(&(elem(&1, 0) >= elem(&2, 0)))
|> Enum.take(@max_suggestions)
|> Enum.sort(&(elem(&1, 1) <= elem(&2, 1)))
|> Enum.map(fn {_, key} -> [" * ", inspect(key), ?\n] end)
end
defimpl Inspect do
import Inspect.Algebra
def inspect(df, opts) do
force_unfit(
concat([
color("#Explorer.DataFrame<", :map, opts),
nest(
concat([line(), Shared.apply_impl(df, :inspect, [opts])]),
2
),
line(),
color(">", :map, opts)
])
)
end
end
end
defimpl Table.Reader, for: Explorer.DataFrame do
def init(df) do
columns = Explorer.DataFrame.names(df)
data =
Enum.map(columns, fn column ->
df
|> Explorer.DataFrame.pull(column)
|> Explorer.Series.to_enum()
end)
{:columns, %{columns: columns, count: Explorer.DataFrame.n_rows(df)}, data}
end
end