Compare commits
4 commits
master
...
rnewman/ba
Author | SHA1 | Date | |
---|---|---|---|
|
e94337c683 | ||
|
c7ea94b4c9 | ||
|
65e7252b56 | ||
|
29ccbee911 |
13 changed files with 1275 additions and 679 deletions
|
@ -499,7 +499,7 @@ fn read_ident_map(conn: &rusqlite::Connection) -> Result<IdentMap> {
|
||||||
fn read_attribute_map(conn: &rusqlite::Connection) -> Result<AttributeMap> {
|
fn read_attribute_map(conn: &rusqlite::Connection) -> Result<AttributeMap> {
|
||||||
let entid_triples = read_materialized_view(conn, "schema")?;
|
let entid_triples = read_materialized_view(conn, "schema")?;
|
||||||
let mut attribute_map = AttributeMap::default();
|
let mut attribute_map = AttributeMap::default();
|
||||||
metadata::update_attribute_map_from_entid_triples(&mut attribute_map, entid_triples)?;
|
metadata::update_attribute_map_from_entid_triples(&mut attribute_map, entid_triples, ::std::iter::empty())?;
|
||||||
Ok(attribute_map)
|
Ok(attribute_map)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1637,7 +1637,7 @@ mod tests {
|
||||||
// Cannot retract a characteristic of an installed attribute.
|
// Cannot retract a characteristic of an installed attribute.
|
||||||
assert_transact!(conn,
|
assert_transact!(conn,
|
||||||
"[[:db/retract 100 :db/cardinality :db.cardinality/many]]",
|
"[[:db/retract 100 :db/cardinality :db.cardinality/many]]",
|
||||||
Err("not yet implemented: Retracting metadata attribute assertions not yet implemented: retracted [e a] pairs [[100 8]]"));
|
Err("bad schema assertion: Retracting 8 for 100 not permitted."));
|
||||||
|
|
||||||
// Trying to install an attribute without a :db/ident is allowed.
|
// Trying to install an attribute without a :db/ident is allowed.
|
||||||
assert_transact!(conn, "[[:db/add 101 :db/valueType :db.type/long]
|
assert_transact!(conn, "[[:db/add 101 :db/valueType :db.type/long]
|
||||||
|
@ -1823,7 +1823,7 @@ mod tests {
|
||||||
|
|
||||||
assert_transact!(conn,
|
assert_transact!(conn,
|
||||||
"[[:db/retract 111 :db/fulltext true]]",
|
"[[:db/retract 111 :db/fulltext true]]",
|
||||||
Err("not yet implemented: Retracting metadata attribute assertions not yet implemented: retracted [e a] pairs [[111 12]]"));
|
Err("bad schema assertion: Retracting 12 for 111 not permitted."));
|
||||||
|
|
||||||
assert_transact!(conn,
|
assert_transact!(conn,
|
||||||
"[[:db/add 222 :db/fulltext true]]",
|
"[[:db/add 222 :db/fulltext true]]",
|
||||||
|
|
|
@ -27,8 +27,6 @@
|
||||||
use std::collections::{BTreeMap, BTreeSet};
|
use std::collections::{BTreeMap, BTreeSet};
|
||||||
use std::collections::btree_map::Entry;
|
use std::collections::btree_map::Entry;
|
||||||
|
|
||||||
use itertools::Itertools; // For join().
|
|
||||||
|
|
||||||
use add_retract_alter_set::{
|
use add_retract_alter_set::{
|
||||||
AddRetractAlterSet,
|
AddRetractAlterSet,
|
||||||
};
|
};
|
||||||
|
@ -104,14 +102,66 @@ impl MetadataReport {
|
||||||
/// contain install and alter markers.
|
/// contain install and alter markers.
|
||||||
///
|
///
|
||||||
/// Returns a report summarizing the mutations that were applied.
|
/// Returns a report summarizing the mutations that were applied.
|
||||||
pub fn update_attribute_map_from_entid_triples<U>(attribute_map: &mut AttributeMap, assertions: U) -> Result<MetadataReport>
|
pub fn update_attribute_map_from_entid_triples<A, R>(attribute_map: &mut AttributeMap, assertions: A, retractions: R) -> Result<MetadataReport>
|
||||||
where U: IntoIterator<Item=(Entid, Entid, TypedValue)> {
|
where A: IntoIterator<Item=(Entid, Entid, TypedValue)>,
|
||||||
|
R: IntoIterator<Item=(Entid, Entid, TypedValue)> {
|
||||||
|
|
||||||
|
fn attribute_builder_to_modify(attribute_id: Entid, existing: &AttributeMap) -> AttributeBuilder {
|
||||||
|
existing.get(&attribute_id)
|
||||||
|
.map(AttributeBuilder::to_modify_attribute)
|
||||||
|
.unwrap_or_else(AttributeBuilder::default)
|
||||||
|
}
|
||||||
|
|
||||||
// Group mutations by impacted entid.
|
// Group mutations by impacted entid.
|
||||||
let mut builders: BTreeMap<Entid, AttributeBuilder> = BTreeMap::new();
|
let mut builders: BTreeMap<Entid, AttributeBuilder> = BTreeMap::new();
|
||||||
|
|
||||||
|
// For retractions, we start with an attribute builder that's pre-populated with the existing
|
||||||
|
// attribute values. That allows us to check existing values and unset them.
|
||||||
|
for (entid, attr, ref value) in retractions.into_iter() {
|
||||||
|
let builder = builders.entry(entid).or_insert_with(|| attribute_builder_to_modify(entid, attribute_map));
|
||||||
|
match attr {
|
||||||
|
// You can only retract :db/unique, :db/doc, :db/isComponent; all others
|
||||||
|
// must be altered instead of retracted, or are not allowed to change.
|
||||||
|
entids::DB_DOC => {
|
||||||
|
// Nothing to do here; we don't keep docstrings inside `Attribute`s.
|
||||||
|
},
|
||||||
|
entids::DB_IS_COMPONENT => {
|
||||||
|
match value {
|
||||||
|
&TypedValue::Boolean(v) if builder.component == Some(v) => {
|
||||||
|
builder.component(false);
|
||||||
|
},
|
||||||
|
v => {
|
||||||
|
bail!(ErrorKind::BadSchemaAssertion(format!("Attempted to retract :db/isComponent with the wrong value {:?}.", v)));
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
entids::DB_UNIQUE => {
|
||||||
|
match *value {
|
||||||
|
TypedValue::Ref(u) => {
|
||||||
|
match u {
|
||||||
|
entids::DB_UNIQUE_VALUE if builder.unique == Some(Some(attribute::Unique::Value)) => {
|
||||||
|
builder.non_unique();
|
||||||
|
},
|
||||||
|
entids::DB_UNIQUE_IDENTITY if builder.unique == Some(Some(attribute::Unique::Identity)) => {
|
||||||
|
builder.non_unique();
|
||||||
|
},
|
||||||
|
v => {
|
||||||
|
bail!(ErrorKind::BadSchemaAssertion(format!("Attempted to retract :db/unique with the wrong value {}.", v)));
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
_ => bail!(ErrorKind::BadSchemaAssertion(format!("Expected [:db/retract _ :db/unique :db.unique/_] but got [:db/retract {} :db/unique {:?}]", entid, value)))
|
||||||
|
}
|
||||||
|
},
|
||||||
|
_ => {
|
||||||
|
bail!(ErrorKind::BadSchemaAssertion(format!("Retracting {} for {} not permitted.", attr, entid)));
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
for (entid, attr, ref value) in assertions.into_iter() {
|
for (entid, attr, ref value) in assertions.into_iter() {
|
||||||
let builder = builders.entry(entid).or_insert(AttributeBuilder::default());
|
// For assertions, we can start with an empty attribute builder.
|
||||||
|
let builder = builders.entry(entid).or_insert_with(Default::default);
|
||||||
|
|
||||||
// TODO: improve error messages throughout.
|
// TODO: improve error messages throughout.
|
||||||
match attr {
|
match attr {
|
||||||
|
@ -146,11 +196,6 @@ pub fn update_attribute_map_from_entid_triples<U>(attribute_map: &mut AttributeM
|
||||||
|
|
||||||
entids::DB_UNIQUE => {
|
entids::DB_UNIQUE => {
|
||||||
match *value {
|
match *value {
|
||||||
// TODO: accept nil in some form.
|
|
||||||
// TypedValue::Nil => {
|
|
||||||
// builder.unique_value(false);
|
|
||||||
// builder.unique_identity(false);
|
|
||||||
// },
|
|
||||||
TypedValue::Ref(entids::DB_UNIQUE_VALUE) => { builder.unique(attribute::Unique::Value); },
|
TypedValue::Ref(entids::DB_UNIQUE_VALUE) => { builder.unique(attribute::Unique::Value); },
|
||||||
TypedValue::Ref(entids::DB_UNIQUE_IDENTITY) => { builder.unique(attribute::Unique::Identity); },
|
TypedValue::Ref(entids::DB_UNIQUE_IDENTITY) => { builder.unique(attribute::Unique::Identity); },
|
||||||
_ => bail!(ErrorKind::BadSchemaAssertion(format!("Expected [... :db/unique :db.unique/value|:db.unique/identity] but got [... :db/unique {:?}]", value)))
|
_ => bail!(ErrorKind::BadSchemaAssertion(format!("Expected [... :db/unique :db.unique/value|:db.unique/identity] but got [... :db/unique {:?}]", value)))
|
||||||
|
@ -257,17 +302,14 @@ pub fn update_schema_from_entid_quadruples<U>(schema: &mut Schema, assertions: U
|
||||||
attribute_set.witness((e, a), typed_value, added);
|
attribute_set.witness((e, a), typed_value, added);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Datomic does not allow to retract attributes or idents. For now, Mentat follows suit.
|
|
||||||
if !attribute_set.retracted.is_empty() {
|
|
||||||
bail!(ErrorKind::NotYetImplemented(format!("Retracting metadata attribute assertions not yet implemented: retracted [e a] pairs [{}]",
|
|
||||||
attribute_set.retracted.keys().map(|&(e, a)| format!("[{} {}]", e, a)).join(", "))));
|
|
||||||
}
|
|
||||||
|
|
||||||
// Collect triples.
|
// Collect triples.
|
||||||
|
let retracted_triples = attribute_set.retracted.into_iter().map(|((e, a), typed_value)| (e, a, typed_value));
|
||||||
let asserted_triples = attribute_set.asserted.into_iter().map(|((e, a), typed_value)| (e, a, typed_value));
|
let asserted_triples = attribute_set.asserted.into_iter().map(|((e, a), typed_value)| (e, a, typed_value));
|
||||||
let altered_triples = attribute_set.altered.into_iter().map(|((e, a), (_old_value, new_value))| (e, a, new_value));
|
let altered_triples = attribute_set.altered.into_iter().map(|((e, a), (_old_value, new_value))| (e, a, new_value));
|
||||||
|
|
||||||
let report = update_attribute_map_from_entid_triples(&mut schema.attribute_map, asserted_triples.chain(altered_triples))?;
|
let report = update_attribute_map_from_entid_triples(&mut schema.attribute_map,
|
||||||
|
asserted_triples.chain(altered_triples),
|
||||||
|
retracted_triples)?;
|
||||||
|
|
||||||
let mut idents_altered: BTreeMap<Entid, IdentAlteration> = BTreeMap::new();
|
let mut idents_altered: BTreeMap<Entid, IdentAlteration> = BTreeMap::new();
|
||||||
|
|
||||||
|
|
|
@ -73,13 +73,13 @@ fn validate_attribute_map(entid_map: &EntidMap, attribute_map: &AttributeMap) ->
|
||||||
#[derive(Clone,Debug,Default,Eq,Hash,Ord,PartialOrd,PartialEq)]
|
#[derive(Clone,Debug,Default,Eq,Hash,Ord,PartialOrd,PartialEq)]
|
||||||
pub struct AttributeBuilder {
|
pub struct AttributeBuilder {
|
||||||
helpful: bool,
|
helpful: bool,
|
||||||
value_type: Option<ValueType>,
|
pub value_type: Option<ValueType>,
|
||||||
multival: Option<bool>,
|
pub multival: Option<bool>,
|
||||||
unique: Option<Option<attribute::Unique>>,
|
pub unique: Option<Option<attribute::Unique>>,
|
||||||
index: Option<bool>,
|
pub index: Option<bool>,
|
||||||
fulltext: Option<bool>,
|
pub fulltext: Option<bool>,
|
||||||
component: Option<bool>,
|
pub component: Option<bool>,
|
||||||
no_history: Option<bool>,
|
pub no_history: Option<bool>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl AttributeBuilder {
|
impl AttributeBuilder {
|
||||||
|
@ -92,6 +92,16 @@ impl AttributeBuilder {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Make a new AttributeBuilder from an existing Attribute. This is important to allow
|
||||||
|
/// retraction. Only attributes that we allow to change are duplicated here.
|
||||||
|
pub fn to_modify_attribute(attribute: &Attribute) -> Self {
|
||||||
|
let mut ab = AttributeBuilder::default();
|
||||||
|
ab.multival = Some(attribute.multival);
|
||||||
|
ab.unique = Some(attribute.unique);
|
||||||
|
ab.component = Some(attribute.component);
|
||||||
|
ab
|
||||||
|
}
|
||||||
|
|
||||||
pub fn value_type<'a>(&'a mut self, value_type: ValueType) -> &'a mut Self {
|
pub fn value_type<'a>(&'a mut self, value_type: ValueType) -> &'a mut Self {
|
||||||
self.value_type = Some(value_type);
|
self.value_type = Some(value_type);
|
||||||
self
|
self
|
||||||
|
@ -102,6 +112,11 @@ impl AttributeBuilder {
|
||||||
self
|
self
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn non_unique<'a>(&'a mut self) -> &'a mut Self {
|
||||||
|
self.unique = Some(None);
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
pub fn unique<'a>(&'a mut self, unique: attribute::Unique) -> &'a mut Self {
|
pub fn unique<'a>(&'a mut self, unique: attribute::Unique) -> &'a mut Self {
|
||||||
if self.helpful && unique == attribute::Unique::Identity {
|
if self.helpful && unique == attribute::Unique::Identity {
|
||||||
self.index = Some(true);
|
self.index = Some(true);
|
||||||
|
@ -185,12 +200,19 @@ impl AttributeBuilder {
|
||||||
mutations.push(AttributeAlteration::Cardinality);
|
mutations.push(AttributeAlteration::Cardinality);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if let Some(ref unique) = self.unique {
|
if let Some(ref unique) = self.unique {
|
||||||
if *unique != attribute.unique {
|
if *unique != attribute.unique {
|
||||||
attribute.unique = unique.clone();
|
attribute.unique = unique.clone();
|
||||||
mutations.push(AttributeAlteration::Unique);
|
mutations.push(AttributeAlteration::Unique);
|
||||||
}
|
}
|
||||||
|
} else {
|
||||||
|
if attribute.unique != None {
|
||||||
|
attribute.unique = None;
|
||||||
|
mutations.push(AttributeAlteration::Unique);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if let Some(index) = self.index {
|
if let Some(index) = self.index {
|
||||||
if index != attribute.index {
|
if index != attribute.index {
|
||||||
attribute.index = index;
|
attribute.index = index;
|
||||||
|
@ -255,7 +277,10 @@ impl SchemaBuilding for Schema {
|
||||||
}).collect();
|
}).collect();
|
||||||
|
|
||||||
let mut schema = Schema::from_ident_map_and_attribute_map(ident_map, AttributeMap::default())?;
|
let mut schema = Schema::from_ident_map_and_attribute_map(ident_map, AttributeMap::default())?;
|
||||||
let metadata_report = metadata::update_attribute_map_from_entid_triples(&mut schema.attribute_map, entid_assertions?)?;
|
let metadata_report = metadata::update_attribute_map_from_entid_triples(&mut schema.attribute_map,
|
||||||
|
entid_assertions?,
|
||||||
|
// No retractions.
|
||||||
|
::std::iter::empty())?;
|
||||||
|
|
||||||
// Rebuild the component attributes list if necessary.
|
// Rebuild the component attributes list if necessary.
|
||||||
if metadata_report.attributes_did_change() {
|
if metadata_report.attributes_did_change() {
|
||||||
|
|
216
query-projector/src/aggregates.rs
Normal file
216
query-projector/src/aggregates.rs
Normal file
|
@ -0,0 +1,216 @@
|
||||||
|
// Copyright 2018 Mozilla
|
||||||
|
//
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use
|
||||||
|
// this file except in compliance with the License. You may obtain a copy of the
|
||||||
|
// License at http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
// Unless required by applicable law or agreed to in writing, software distributed
|
||||||
|
// under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
|
||||||
|
// CONDITIONS OF ANY KIND, either express or implied. See the License for the
|
||||||
|
// specific language governing permissions and limitations under the License.
|
||||||
|
|
||||||
|
use mentat_core::{
|
||||||
|
ValueType,
|
||||||
|
ValueTypeSet,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_query::{
|
||||||
|
Aggregate,
|
||||||
|
QueryFunction,
|
||||||
|
Variable,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_query_algebrizer::{
|
||||||
|
ColumnName,
|
||||||
|
ConjoiningClauses,
|
||||||
|
VariableColumn,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_query_sql::{
|
||||||
|
ColumnOrExpression,
|
||||||
|
Expression,
|
||||||
|
Name,
|
||||||
|
ProjectedColumn,
|
||||||
|
};
|
||||||
|
|
||||||
|
use errors::{
|
||||||
|
ErrorKind,
|
||||||
|
Result,
|
||||||
|
};
|
||||||
|
|
||||||
|
#[derive(Clone, Copy, Debug, Eq, PartialEq)]
|
||||||
|
pub enum SimpleAggregationOp {
|
||||||
|
Avg,
|
||||||
|
Count,
|
||||||
|
Max,
|
||||||
|
Min,
|
||||||
|
Sum,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl SimpleAggregationOp {
|
||||||
|
pub(crate) fn to_sql(&self) -> &'static str {
|
||||||
|
use self::SimpleAggregationOp::*;
|
||||||
|
match self {
|
||||||
|
&Avg => "avg",
|
||||||
|
&Count => "count",
|
||||||
|
&Max => "max",
|
||||||
|
&Min => "min",
|
||||||
|
&Sum => "sum",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn for_function(function: &QueryFunction) -> Option<SimpleAggregationOp> {
|
||||||
|
match function.0.plain_name() {
|
||||||
|
"avg" => Some(SimpleAggregationOp::Avg),
|
||||||
|
"count" => Some(SimpleAggregationOp::Count),
|
||||||
|
"max" => Some(SimpleAggregationOp::Max),
|
||||||
|
"min" => Some(SimpleAggregationOp::Min),
|
||||||
|
"sum" => Some(SimpleAggregationOp::Sum),
|
||||||
|
_ => None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// With knowledge of the types to which a variable might be bound,
|
||||||
|
/// return a `Result` to determine whether this aggregation is suitable.
|
||||||
|
/// For example, it's valid to take the `Avg` of `{Double, Long}`, invalid
|
||||||
|
/// to take `Sum` of `{Instant}`, valid to take (lexicographic) `Max` of `{String}`,
|
||||||
|
/// but invalid to take `Max` of `{Uuid, String}`.
|
||||||
|
///
|
||||||
|
/// The returned type is the type of the result of the aggregation.
|
||||||
|
pub(crate) fn is_applicable_to_types(&self, possibilities: ValueTypeSet) -> Result<ValueType> {
|
||||||
|
use self::SimpleAggregationOp::*;
|
||||||
|
if possibilities.is_empty() {
|
||||||
|
bail!(ErrorKind::CannotProjectImpossibleBinding(*self))
|
||||||
|
}
|
||||||
|
|
||||||
|
match self {
|
||||||
|
// One can always count results.
|
||||||
|
&Count => Ok(ValueType::Long),
|
||||||
|
|
||||||
|
// Only numeric types can be averaged or summed.
|
||||||
|
&Avg => {
|
||||||
|
if possibilities.is_only_numeric() {
|
||||||
|
// The mean of a set of numeric values will always, for our purposes, be a double.
|
||||||
|
Ok(ValueType::Double)
|
||||||
|
} else {
|
||||||
|
bail!(ErrorKind::CannotApplyAggregateOperationToTypes(*self, possibilities))
|
||||||
|
}
|
||||||
|
},
|
||||||
|
&Sum => {
|
||||||
|
if possibilities.is_only_numeric() {
|
||||||
|
if possibilities.contains(ValueType::Double) {
|
||||||
|
Ok(ValueType::Double)
|
||||||
|
} else {
|
||||||
|
// TODO: BigInt.
|
||||||
|
Ok(ValueType::Long)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
bail!(ErrorKind::CannotApplyAggregateOperationToTypes(*self, possibilities))
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
&Max | &Min => {
|
||||||
|
if possibilities.is_unit() {
|
||||||
|
use ValueType::*;
|
||||||
|
let the_type = possibilities.exemplar().expect("a type");
|
||||||
|
match the_type {
|
||||||
|
// These types are numerically ordered.
|
||||||
|
Double | Long | Instant => Ok(the_type),
|
||||||
|
|
||||||
|
// Boolean: false < true.
|
||||||
|
Boolean => Ok(the_type),
|
||||||
|
|
||||||
|
// String: lexicographic order.
|
||||||
|
String => Ok(the_type),
|
||||||
|
|
||||||
|
// These types are unordered.
|
||||||
|
Keyword | Ref | Uuid => {
|
||||||
|
bail!(ErrorKind::CannotApplyAggregateOperationToTypes(*self, possibilities))
|
||||||
|
},
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// It cannot be empty -- we checked.
|
||||||
|
// The only types that are valid to compare cross-type are numbers.
|
||||||
|
if possibilities.is_only_numeric() {
|
||||||
|
// Note that if the max/min is a Long, it will be returned as a Double!
|
||||||
|
if possibilities.contains(ValueType::Double) {
|
||||||
|
Ok(ValueType::Double)
|
||||||
|
} else {
|
||||||
|
// TODO: BigInt.
|
||||||
|
Ok(ValueType::Long)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
bail!(ErrorKind::CannotApplyAggregateOperationToTypes(*self, possibilities))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub(crate) struct SimpleAggregate {
|
||||||
|
pub op: SimpleAggregationOp,
|
||||||
|
pub var: Variable,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl SimpleAggregate {
|
||||||
|
pub(crate) fn column_name(&self) -> Name {
|
||||||
|
format!("({} {})", self.op.to_sql(), self.var.name())
|
||||||
|
}
|
||||||
|
|
||||||
|
pub(crate) fn use_static_value(&self) -> bool {
|
||||||
|
use self::SimpleAggregationOp::*;
|
||||||
|
match self.op {
|
||||||
|
Avg | Max | Min => true,
|
||||||
|
Count | Sum => false,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub(crate) trait SimpleAggregation {
|
||||||
|
fn to_simple(&self) -> Option<SimpleAggregate>;
|
||||||
|
}
|
||||||
|
|
||||||
|
impl SimpleAggregation for Aggregate {
|
||||||
|
fn to_simple(&self) -> Option<SimpleAggregate> {
|
||||||
|
if self.args.len() != 1 {
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
self.args[0]
|
||||||
|
.as_variable()
|
||||||
|
.and_then(|v| SimpleAggregationOp::for_function(&self.func)
|
||||||
|
.map(|op| SimpleAggregate { op, var: v.clone(), }))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Returns two values:
|
||||||
|
/// - The `ColumnOrExpression` to use in the query. This will always refer to other
|
||||||
|
/// variables by name; never to a datoms column.
|
||||||
|
/// - The known type of that value.
|
||||||
|
pub(crate) fn projected_column_for_simple_aggregate(simple: &SimpleAggregate, cc: &ConjoiningClauses) -> Result<(ProjectedColumn, ValueType)> {
|
||||||
|
let known_types = cc.known_type_set(&simple.var);
|
||||||
|
let return_type = simple.op.is_applicable_to_types(known_types)?;
|
||||||
|
let projected_column_or_expression =
|
||||||
|
if let Some(value) = cc.bound_value(&simple.var) {
|
||||||
|
// Oh, we already know the value!
|
||||||
|
if simple.use_static_value() {
|
||||||
|
// We can statically compute the aggregate result for some operators -- not count or
|
||||||
|
// sum, but avg/max/min are OK.
|
||||||
|
ColumnOrExpression::Value(value)
|
||||||
|
} else {
|
||||||
|
let expression = Expression::Unary {
|
||||||
|
sql_op: simple.op.to_sql(),
|
||||||
|
arg: ColumnOrExpression::Value(value),
|
||||||
|
};
|
||||||
|
ColumnOrExpression::Expression(Box::new(expression), return_type)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// The common case: the values are bound during execution.
|
||||||
|
let name = VariableColumn::Variable(simple.var.clone()).column_name();
|
||||||
|
let expression = Expression::Unary {
|
||||||
|
sql_op: simple.op.to_sql(),
|
||||||
|
arg: ColumnOrExpression::ExistingColumn(name),
|
||||||
|
};
|
||||||
|
ColumnOrExpression::Expression(Box::new(expression), return_type)
|
||||||
|
};
|
||||||
|
Ok((ProjectedColumn(projected_column_or_expression, simple.column_name()), return_type))
|
||||||
|
}
|
73
query-projector/src/errors.rs
Normal file
73
query-projector/src/errors.rs
Normal file
|
@ -0,0 +1,73 @@
|
||||||
|
// Copyright 2018 Mozilla
|
||||||
|
//
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use
|
||||||
|
// this file except in compliance with the License. You may obtain a copy of the
|
||||||
|
// License at http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
// Unless required by applicable law or agreed to in writing, software distributed
|
||||||
|
// under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
|
||||||
|
// CONDITIONS OF ANY KIND, either express or implied. See the License for the
|
||||||
|
// specific language governing permissions and limitations under the License.
|
||||||
|
|
||||||
|
use rusqlite;
|
||||||
|
|
||||||
|
use mentat_core::{
|
||||||
|
ValueTypeSet,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_db;
|
||||||
|
|
||||||
|
use mentat_query::{
|
||||||
|
PlainSymbol,
|
||||||
|
};
|
||||||
|
|
||||||
|
use aggregates::{
|
||||||
|
SimpleAggregationOp,
|
||||||
|
};
|
||||||
|
|
||||||
|
error_chain! {
|
||||||
|
types {
|
||||||
|
Error, ErrorKind, ResultExt, Result;
|
||||||
|
}
|
||||||
|
|
||||||
|
errors {
|
||||||
|
/// We're just not done yet. Message that the feature is recognized but not yet
|
||||||
|
/// implemented.
|
||||||
|
NotYetImplemented(t: String) {
|
||||||
|
description("not yet implemented")
|
||||||
|
display("not yet implemented: {}", t)
|
||||||
|
}
|
||||||
|
CannotProjectImpossibleBinding(op: SimpleAggregationOp) {
|
||||||
|
description("no possible types for variable in projection list")
|
||||||
|
display("no possible types for value provided to {:?}", op)
|
||||||
|
}
|
||||||
|
CannotApplyAggregateOperationToTypes(op: SimpleAggregationOp, types: ValueTypeSet) {
|
||||||
|
description("cannot apply projection operation to types")
|
||||||
|
display("cannot apply projection operation {:?} to types {:?}", op, types)
|
||||||
|
}
|
||||||
|
UnboundVariable(var: PlainSymbol) {
|
||||||
|
description("cannot project unbound variable")
|
||||||
|
display("cannot project unbound variable {:?}", var)
|
||||||
|
}
|
||||||
|
NoTypeAvailableForVariable(var: PlainSymbol) {
|
||||||
|
description("cannot find type for variable")
|
||||||
|
display("cannot find type for variable {:?}", var)
|
||||||
|
}
|
||||||
|
UnexpectedResultsType(actual: &'static str, expected: &'static str) {
|
||||||
|
description("unexpected query results type")
|
||||||
|
display("expected {}, got {}", expected, actual)
|
||||||
|
}
|
||||||
|
AmbiguousAggregates(min_max_count: usize, corresponding_count: usize) {
|
||||||
|
description("ambiguous aggregates")
|
||||||
|
display("min/max expressions: {} (max 1), corresponding: {}", min_max_count, corresponding_count)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
foreign_links {
|
||||||
|
Rusqlite(rusqlite::Error);
|
||||||
|
}
|
||||||
|
|
||||||
|
links {
|
||||||
|
DbError(mentat_db::Error, mentat_db::ErrorKind);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
|
@ -28,21 +28,14 @@ use std::iter;
|
||||||
|
|
||||||
use std::rc::Rc;
|
use std::rc::Rc;
|
||||||
|
|
||||||
use indexmap::{
|
|
||||||
IndexSet,
|
|
||||||
};
|
|
||||||
|
|
||||||
use rusqlite::{
|
use rusqlite::{
|
||||||
Row,
|
Row,
|
||||||
Rows,
|
Rows,
|
||||||
};
|
};
|
||||||
|
|
||||||
use mentat_core::{
|
use mentat_core::{
|
||||||
SQLValueType,
|
|
||||||
SQLValueTypeSet,
|
|
||||||
TypedValue,
|
TypedValue,
|
||||||
ValueType,
|
ValueType,
|
||||||
ValueTypeSet,
|
|
||||||
ValueTypeTag,
|
ValueTypeTag,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -55,79 +48,44 @@ use mentat_db::{
|
||||||
};
|
};
|
||||||
|
|
||||||
use mentat_query::{
|
use mentat_query::{
|
||||||
Aggregate,
|
|
||||||
Element,
|
Element,
|
||||||
FindSpec,
|
FindSpec,
|
||||||
Limit,
|
Limit,
|
||||||
PlainSymbol,
|
|
||||||
QueryFunction,
|
|
||||||
Variable,
|
Variable,
|
||||||
};
|
};
|
||||||
|
|
||||||
use mentat_query_algebrizer::{
|
use mentat_query_algebrizer::{
|
||||||
AlgebraicQuery,
|
AlgebraicQuery,
|
||||||
ColumnName,
|
|
||||||
ConjoiningClauses,
|
|
||||||
QualifiedAlias,
|
|
||||||
VariableBindings,
|
VariableBindings,
|
||||||
VariableColumn,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
use mentat_query_sql::{
|
use mentat_query_sql::{
|
||||||
ColumnOrExpression,
|
|
||||||
Expression,
|
|
||||||
GroupBy,
|
GroupBy,
|
||||||
Name,
|
|
||||||
Projection,
|
Projection,
|
||||||
ProjectedColumn,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
error_chain! {
|
mod aggregates;
|
||||||
types {
|
mod project;
|
||||||
Error, ErrorKind, ResultExt, Result;
|
mod pull;
|
||||||
}
|
pub mod errors;
|
||||||
|
|
||||||
errors {
|
pub use aggregates::{
|
||||||
/// We're just not done yet. Message that the feature is recognized but not yet
|
SimpleAggregationOp,
|
||||||
/// implemented.
|
};
|
||||||
NotYetImplemented(t: String) {
|
|
||||||
description("not yet implemented")
|
|
||||||
display("not yet implemented: {}", t)
|
|
||||||
}
|
|
||||||
CannotProjectImpossibleBinding(op: SimpleAggregationOp) {
|
|
||||||
description("no possible types for variable in projection list")
|
|
||||||
display("no possible types for value provided to {:?}", op)
|
|
||||||
}
|
|
||||||
CannotApplyAggregateOperationToTypes(op: SimpleAggregationOp, types: ValueTypeSet) {
|
|
||||||
description("cannot apply projection operation to types")
|
|
||||||
display("cannot apply projection operation {:?} to types {:?}", op, types)
|
|
||||||
}
|
|
||||||
UnboundVariable(var: PlainSymbol) {
|
|
||||||
description("cannot project unbound variable")
|
|
||||||
display("cannot project unbound variable {:?}", var)
|
|
||||||
}
|
|
||||||
NoTypeAvailableForVariable(var: PlainSymbol) {
|
|
||||||
description("cannot find type for variable")
|
|
||||||
display("cannot find type for variable {:?}", var)
|
|
||||||
}
|
|
||||||
UnexpectedResultsType(actual: &'static str, expected: &'static str) {
|
|
||||||
description("unexpected query results type")
|
|
||||||
display("expected {}, got {}", expected, actual)
|
|
||||||
}
|
|
||||||
AmbiguousAggregates(min_max_count: usize, corresponding_count: usize) {
|
|
||||||
description("ambiguous aggregates")
|
|
||||||
display("min/max expressions: {} (max 1), corresponding: {}", min_max_count, corresponding_count)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
foreign_links {
|
use project::{
|
||||||
Rusqlite(rusqlite::Error);
|
ProjectedElements,
|
||||||
}
|
project_elements,
|
||||||
|
};
|
||||||
|
|
||||||
links {
|
pub use project::{
|
||||||
DbError(mentat_db::Error, mentat_db::ErrorKind);
|
projected_column_for_var,
|
||||||
}
|
};
|
||||||
}
|
|
||||||
|
use errors::{
|
||||||
|
ErrorKind,
|
||||||
|
Result,
|
||||||
|
};
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq)]
|
#[derive(Debug, PartialEq, Eq)]
|
||||||
pub struct QueryOutput {
|
pub struct QueryOutput {
|
||||||
|
@ -353,525 +311,6 @@ impl TypedIndex {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn cc_column(cc: &ConjoiningClauses, var: &Variable) -> Result<QualifiedAlias> {
|
|
||||||
cc.column_bindings
|
|
||||||
.get(var)
|
|
||||||
.and_then(|cols| cols.get(0).cloned())
|
|
||||||
.ok_or_else(|| ErrorKind::UnboundVariable(var.name()).into())
|
|
||||||
}
|
|
||||||
|
|
||||||
fn candidate_column(cc: &ConjoiningClauses, var: &Variable) -> Result<(ColumnOrExpression, Name)> {
|
|
||||||
// Every variable should be bound by the top-level CC to at least
|
|
||||||
// one column in the query. If that constraint is violated it's a
|
|
||||||
// bug in our code, so it's appropriate to panic here.
|
|
||||||
cc_column(cc, var)
|
|
||||||
.map(|qa| {
|
|
||||||
let name = VariableColumn::Variable(var.clone()).column_name();
|
|
||||||
(ColumnOrExpression::Column(qa), name)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
fn candidate_type_column(cc: &ConjoiningClauses, var: &Variable) -> Result<(ColumnOrExpression, Name)> {
|
|
||||||
cc.extracted_types
|
|
||||||
.get(var)
|
|
||||||
.cloned()
|
|
||||||
.map(|alias| {
|
|
||||||
let type_name = VariableColumn::VariableTypeTag(var.clone()).column_name();
|
|
||||||
(ColumnOrExpression::Column(alias), type_name)
|
|
||||||
})
|
|
||||||
.ok_or_else(|| ErrorKind::UnboundVariable(var.name()).into())
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Return the projected column -- that is, a value or SQL column and an associated name -- for a
|
|
||||||
/// given variable. Also return the type.
|
|
||||||
/// Callers are expected to determine whether to project a type tag as an additional SQL column.
|
|
||||||
pub fn projected_column_for_var(var: &Variable, cc: &ConjoiningClauses) -> Result<(ProjectedColumn, ValueTypeSet)> {
|
|
||||||
if let Some(value) = cc.bound_value(&var) {
|
|
||||||
// If we already know the value, then our lives are easy.
|
|
||||||
let tag = value.value_type();
|
|
||||||
let name = VariableColumn::Variable(var.clone()).column_name();
|
|
||||||
Ok((ProjectedColumn(ColumnOrExpression::Value(value.clone()), name), ValueTypeSet::of_one(tag)))
|
|
||||||
} else {
|
|
||||||
// If we don't, then the CC *must* have bound the variable.
|
|
||||||
let (column, name) = candidate_column(cc, var)?;
|
|
||||||
Ok((ProjectedColumn(column, name), cc.known_type_set(var)))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Returns two values:
|
|
||||||
/// - The `ColumnOrExpression` to use in the query. This will always refer to other
|
|
||||||
/// variables by name; never to a datoms column.
|
|
||||||
/// - The known type of that value.
|
|
||||||
fn projected_column_for_simple_aggregate(simple: &SimpleAggregate, cc: &ConjoiningClauses) -> Result<(ProjectedColumn, ValueType)> {
|
|
||||||
let known_types = cc.known_type_set(&simple.var);
|
|
||||||
let return_type = simple.op.is_applicable_to_types(known_types)?;
|
|
||||||
let projected_column_or_expression =
|
|
||||||
if let Some(value) = cc.bound_value(&simple.var) {
|
|
||||||
// Oh, we already know the value!
|
|
||||||
if simple.use_static_value() {
|
|
||||||
// We can statically compute the aggregate result for some operators -- not count or
|
|
||||||
// sum, but avg/max/min are OK.
|
|
||||||
ColumnOrExpression::Value(value)
|
|
||||||
} else {
|
|
||||||
let expression = Expression::Unary {
|
|
||||||
sql_op: simple.op.to_sql(),
|
|
||||||
arg: ColumnOrExpression::Value(value),
|
|
||||||
};
|
|
||||||
ColumnOrExpression::Expression(Box::new(expression), return_type)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
// The common case: the values are bound during execution.
|
|
||||||
let name = VariableColumn::Variable(simple.var.clone()).column_name();
|
|
||||||
let expression = Expression::Unary {
|
|
||||||
sql_op: simple.op.to_sql(),
|
|
||||||
arg: ColumnOrExpression::ExistingColumn(name),
|
|
||||||
};
|
|
||||||
ColumnOrExpression::Expression(Box::new(expression), return_type)
|
|
||||||
};
|
|
||||||
Ok((ProjectedColumn(projected_column_or_expression, simple.column_name()), return_type))
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Clone, Copy, Debug, Eq, PartialEq)]
|
|
||||||
pub enum SimpleAggregationOp {
|
|
||||||
Avg,
|
|
||||||
Count,
|
|
||||||
Max,
|
|
||||||
Min,
|
|
||||||
Sum,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl SimpleAggregationOp {
|
|
||||||
fn to_sql(&self) -> &'static str {
|
|
||||||
use SimpleAggregationOp::*;
|
|
||||||
match self {
|
|
||||||
&Avg => "avg",
|
|
||||||
&Count => "count",
|
|
||||||
&Max => "max",
|
|
||||||
&Min => "min",
|
|
||||||
&Sum => "sum",
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn for_function(function: &QueryFunction) -> Option<SimpleAggregationOp> {
|
|
||||||
match function.0.plain_name() {
|
|
||||||
"avg" => Some(SimpleAggregationOp::Avg),
|
|
||||||
"count" => Some(SimpleAggregationOp::Count),
|
|
||||||
"max" => Some(SimpleAggregationOp::Max),
|
|
||||||
"min" => Some(SimpleAggregationOp::Min),
|
|
||||||
"sum" => Some(SimpleAggregationOp::Sum),
|
|
||||||
_ => None,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// With knowledge of the types to which a variable might be bound,
|
|
||||||
/// return a `Result` to determine whether this aggregation is suitable.
|
|
||||||
/// For example, it's valid to take the `Avg` of `{Double, Long}`, invalid
|
|
||||||
/// to take `Sum` of `{Instant}`, valid to take (lexicographic) `Max` of `{String}`,
|
|
||||||
/// but invalid to take `Max` of `{Uuid, String}`.
|
|
||||||
///
|
|
||||||
/// The returned type is the type of the result of the aggregation.
|
|
||||||
fn is_applicable_to_types(&self, possibilities: ValueTypeSet) -> Result<ValueType> {
|
|
||||||
use SimpleAggregationOp::*;
|
|
||||||
if possibilities.is_empty() {
|
|
||||||
bail!(ErrorKind::CannotProjectImpossibleBinding(*self))
|
|
||||||
}
|
|
||||||
|
|
||||||
match self {
|
|
||||||
// One can always count results.
|
|
||||||
&Count => Ok(ValueType::Long),
|
|
||||||
|
|
||||||
// Only numeric types can be averaged or summed.
|
|
||||||
&Avg => {
|
|
||||||
if possibilities.is_only_numeric() {
|
|
||||||
// The mean of a set of numeric values will always, for our purposes, be a double.
|
|
||||||
Ok(ValueType::Double)
|
|
||||||
} else {
|
|
||||||
bail!(ErrorKind::CannotApplyAggregateOperationToTypes(*self, possibilities))
|
|
||||||
}
|
|
||||||
},
|
|
||||||
&Sum => {
|
|
||||||
if possibilities.is_only_numeric() {
|
|
||||||
if possibilities.contains(ValueType::Double) {
|
|
||||||
Ok(ValueType::Double)
|
|
||||||
} else {
|
|
||||||
// TODO: BigInt.
|
|
||||||
Ok(ValueType::Long)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
bail!(ErrorKind::CannotApplyAggregateOperationToTypes(*self, possibilities))
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
&Max | &Min => {
|
|
||||||
if possibilities.is_unit() {
|
|
||||||
use ValueType::*;
|
|
||||||
let the_type = possibilities.exemplar().expect("a type");
|
|
||||||
match the_type {
|
|
||||||
// These types are numerically ordered.
|
|
||||||
Double | Long | Instant => Ok(the_type),
|
|
||||||
|
|
||||||
// Boolean: false < true.
|
|
||||||
Boolean => Ok(the_type),
|
|
||||||
|
|
||||||
// String: lexicographic order.
|
|
||||||
String => Ok(the_type),
|
|
||||||
|
|
||||||
// These types are unordered.
|
|
||||||
Keyword | Ref | Uuid => {
|
|
||||||
bail!(ErrorKind::CannotApplyAggregateOperationToTypes(*self, possibilities))
|
|
||||||
},
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
// It cannot be empty -- we checked.
|
|
||||||
// The only types that are valid to compare cross-type are numbers.
|
|
||||||
if possibilities.is_only_numeric() {
|
|
||||||
// Note that if the max/min is a Long, it will be returned as a Double!
|
|
||||||
if possibilities.contains(ValueType::Double) {
|
|
||||||
Ok(ValueType::Double)
|
|
||||||
} else {
|
|
||||||
// TODO: BigInt.
|
|
||||||
Ok(ValueType::Long)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
bail!(ErrorKind::CannotApplyAggregateOperationToTypes(*self, possibilities))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
struct SimpleAggregate {
|
|
||||||
op: SimpleAggregationOp,
|
|
||||||
var: Variable,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl SimpleAggregate {
|
|
||||||
fn column_name(&self) -> Name {
|
|
||||||
format!("({} {})", self.op.to_sql(), self.var.name())
|
|
||||||
}
|
|
||||||
|
|
||||||
fn use_static_value(&self) -> bool {
|
|
||||||
use SimpleAggregationOp::*;
|
|
||||||
match self.op {
|
|
||||||
Avg | Max | Min => true,
|
|
||||||
Count | Sum => false,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
trait SimpleAggregation {
|
|
||||||
fn to_simple(&self) -> Option<SimpleAggregate>;
|
|
||||||
}
|
|
||||||
|
|
||||||
impl SimpleAggregation for Aggregate {
|
|
||||||
fn to_simple(&self) -> Option<SimpleAggregate> {
|
|
||||||
if self.args.len() != 1 {
|
|
||||||
return None;
|
|
||||||
}
|
|
||||||
self.args[0]
|
|
||||||
.as_variable()
|
|
||||||
.and_then(|v| SimpleAggregationOp::for_function(&self.func)
|
|
||||||
.map(|op| SimpleAggregate { op, var: v.clone(), }))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// An internal temporary struct to pass between the projection 'walk' and the
|
|
||||||
/// resultant projector.
|
|
||||||
/// Projection accumulates four things:
|
|
||||||
/// - Two SQL projection lists. We need two because aggregate queries are nested
|
|
||||||
/// in order to apply DISTINCT to values prior to aggregation.
|
|
||||||
/// - A collection of templates for the projector to use to extract values.
|
|
||||||
/// - A list of columns to use for grouping. Grouping is a property of the projection!
|
|
||||||
struct ProjectedElements {
|
|
||||||
sql_projection: Projection,
|
|
||||||
pre_aggregate_projection: Option<Projection>,
|
|
||||||
templates: Vec<TypedIndex>,
|
|
||||||
group_by: Vec<GroupBy>,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Walk an iterator of `Element`s, collecting projector templates and columns.
|
|
||||||
///
|
|
||||||
/// Returns a `ProjectedElements`, which combines SQL projections
|
|
||||||
/// and a `Vec` of `TypedIndex` 'keys' to use when looking up values.
|
|
||||||
///
|
|
||||||
/// Callers must ensure that every `Element` is distinct -- a query like
|
|
||||||
///
|
|
||||||
/// ```edn
|
|
||||||
/// [:find ?x ?x :where [?x _ _]]
|
|
||||||
/// ```
|
|
||||||
///
|
|
||||||
/// should fail to parse. See #358.
|
|
||||||
fn project_elements<'a, I: IntoIterator<Item = &'a Element>>(
|
|
||||||
count: usize,
|
|
||||||
elements: I,
|
|
||||||
query: &AlgebraicQuery) -> Result<ProjectedElements> {
|
|
||||||
|
|
||||||
// Give a little padding for type tags.
|
|
||||||
let mut inner_projection = Vec::with_capacity(count + 2);
|
|
||||||
|
|
||||||
// Everything in the outer query will _either_ be an aggregate operation
|
|
||||||
// _or_ a reference to a name projected from the inner.
|
|
||||||
// We'll expand them later.
|
|
||||||
let mut outer_projection: Vec<Either<Name, ProjectedColumn>> = Vec::with_capacity(count + 2);
|
|
||||||
|
|
||||||
let mut i: i32 = 0;
|
|
||||||
let mut min_max_count: usize = 0;
|
|
||||||
let mut corresponding_count: usize = 0;
|
|
||||||
let mut templates = vec![];
|
|
||||||
|
|
||||||
let mut aggregates = false;
|
|
||||||
|
|
||||||
// Any variable that appears intact in the :find clause, not inside an aggregate expression.
|
|
||||||
// "Query variables not in aggregate expressions will group the results and appear intact
|
|
||||||
// in the result."
|
|
||||||
// We use an ordered set here so that we group in the correct order.
|
|
||||||
let mut outer_variables = IndexSet::new();
|
|
||||||
|
|
||||||
// Any variable that we are projecting from the inner query.
|
|
||||||
let mut inner_variables = BTreeSet::new();
|
|
||||||
|
|
||||||
for e in elements {
|
|
||||||
if let &Element::Corresponding(_) = e {
|
|
||||||
corresponding_count += 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
match e {
|
|
||||||
// Each time we come across a variable, we push a SQL column
|
|
||||||
// into the SQL projection, aliased to the name of the variable,
|
|
||||||
// and we push an annotated index into the projector.
|
|
||||||
&Element::Variable(ref var) |
|
|
||||||
&Element::Corresponding(ref var) => {
|
|
||||||
if outer_variables.contains(var) {
|
|
||||||
eprintln!("Warning: duplicate variable {} in query.", var);
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: it's an error to have `[:find ?x (the ?x) …]`.
|
|
||||||
outer_variables.insert(var.clone());
|
|
||||||
inner_variables.insert(var.clone());
|
|
||||||
|
|
||||||
let (projected_column, type_set) = projected_column_for_var(&var, &query.cc)?;
|
|
||||||
outer_projection.push(Either::Left(projected_column.1.clone()));
|
|
||||||
inner_projection.push(projected_column);
|
|
||||||
|
|
||||||
if let Some(tag) = type_set.unique_type_tag() {
|
|
||||||
templates.push(TypedIndex::Known(i, tag));
|
|
||||||
i += 1; // We used one SQL column.
|
|
||||||
} else {
|
|
||||||
templates.push(TypedIndex::Unknown(i, i + 1));
|
|
||||||
i += 2; // We used two SQL columns.
|
|
||||||
|
|
||||||
// Also project the type from the SQL query.
|
|
||||||
let (type_column, type_name) = candidate_type_column(&query.cc, &var)?;
|
|
||||||
inner_projection.push(ProjectedColumn(type_column, type_name.clone()));
|
|
||||||
outer_projection.push(Either::Left(type_name));
|
|
||||||
}
|
|
||||||
},
|
|
||||||
&Element::Aggregate(ref a) => {
|
|
||||||
if let Some(simple) = a.to_simple() {
|
|
||||||
aggregates = true;
|
|
||||||
|
|
||||||
use SimpleAggregationOp::*;
|
|
||||||
match simple.op {
|
|
||||||
Max | Min => {
|
|
||||||
min_max_count += 1;
|
|
||||||
},
|
|
||||||
Avg | Count | Sum => (),
|
|
||||||
}
|
|
||||||
|
|
||||||
// When we encounter a simple aggregate -- one in which the aggregation can be
|
|
||||||
// implemented in SQL, on a single variable -- we just push the SQL aggregation op.
|
|
||||||
// We must ensure the following:
|
|
||||||
// - There's a column for the var.
|
|
||||||
// - The type of the var is known to be restricted to a sensible input set
|
|
||||||
// (not necessarily a single type, but e.g., all vals must be Double or Long).
|
|
||||||
// - The type set must be appropriate for the operation. E.g., `Sum` is not a
|
|
||||||
// meaningful operation on instants.
|
|
||||||
|
|
||||||
let (projected_column, return_type) = projected_column_for_simple_aggregate(&simple, &query.cc)?;
|
|
||||||
outer_projection.push(Either::Right(projected_column));
|
|
||||||
|
|
||||||
if !inner_variables.contains(&simple.var) {
|
|
||||||
inner_variables.insert(simple.var.clone());
|
|
||||||
let (projected_column, _type_set) = projected_column_for_var(&simple.var, &query.cc)?;
|
|
||||||
inner_projection.push(projected_column);
|
|
||||||
if query.cc.known_type_set(&simple.var).unique_type_tag().is_none() {
|
|
||||||
// Also project the type from the SQL query.
|
|
||||||
let (type_column, type_name) = candidate_type_column(&query.cc, &simple.var)?;
|
|
||||||
inner_projection.push(ProjectedColumn(type_column, type_name.clone()));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// We might regret using the type tag here instead of the `ValueType`.
|
|
||||||
templates.push(TypedIndex::Known(i, return_type.value_type_tag()));
|
|
||||||
i += 1;
|
|
||||||
} else {
|
|
||||||
// TODO: complex aggregates.
|
|
||||||
bail!(ErrorKind::NotYetImplemented("complex aggregates".into()));
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
match (min_max_count, corresponding_count) {
|
|
||||||
(0, 0) | (_, 0) => {},
|
|
||||||
(0, _) => {
|
|
||||||
eprintln!("Warning: used `(the ?var)` without `min` or `max`.");
|
|
||||||
},
|
|
||||||
(1, _) => {
|
|
||||||
// This is the success case!
|
|
||||||
},
|
|
||||||
(n, c) => {
|
|
||||||
bail!(ErrorKind::AmbiguousAggregates(n, c));
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
// Anything used in ORDER BY (which we're given in `named_projection`)
|
|
||||||
// needs to be in the SQL column list so we can refer to it by name.
|
|
||||||
//
|
|
||||||
// They don't affect projection.
|
|
||||||
//
|
|
||||||
// If a variable is of a non-fixed type, also project the type tag column, so we don't
|
|
||||||
// accidentally unify across types when considering uniqueness!
|
|
||||||
for var in query.named_projection.iter() {
|
|
||||||
if outer_variables.contains(var) {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
// If it's a fixed value, we need do nothing further.
|
|
||||||
if query.cc.is_value_bound(&var) {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
let already_inner = inner_variables.contains(&var);
|
|
||||||
let (column, name) = candidate_column(&query.cc, &var)?;
|
|
||||||
if !already_inner {
|
|
||||||
inner_projection.push(ProjectedColumn(column, name.clone()));
|
|
||||||
inner_variables.insert(var.clone());
|
|
||||||
}
|
|
||||||
|
|
||||||
outer_projection.push(Either::Left(name));
|
|
||||||
outer_variables.insert(var.clone());
|
|
||||||
|
|
||||||
// We don't care if a column has a single _type_, we care if it has a single type _tag_,
|
|
||||||
// because that's what we'll use if we're projecting. E.g., Long and Double.
|
|
||||||
// Single type implies single type tag, and is cheaper, so we check that first.
|
|
||||||
let types = query.cc.known_type_set(&var);
|
|
||||||
if !types.has_unique_type_tag() {
|
|
||||||
let (type_column, type_name) = candidate_type_column(&query.cc, &var)?;
|
|
||||||
if !already_inner {
|
|
||||||
inner_projection.push(ProjectedColumn(type_column, type_name.clone()));
|
|
||||||
}
|
|
||||||
|
|
||||||
outer_projection.push(Either::Left(type_name));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if !aggregates {
|
|
||||||
// We're done -- we never need to group unless we're aggregating.
|
|
||||||
return Ok(ProjectedElements {
|
|
||||||
sql_projection: Projection::Columns(inner_projection),
|
|
||||||
pre_aggregate_projection: None,
|
|
||||||
templates,
|
|
||||||
group_by: vec![],
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
// OK, on to aggregates.
|
|
||||||
// We need to produce two SQL projection lists: one for an inner query and one for the outer.
|
|
||||||
//
|
|
||||||
// The inner serves these purposes:
|
|
||||||
// - Projecting variables to avoid duplicates being elided. (:with)
|
|
||||||
// - Making bindings available to the outermost query for projection, ordering, and grouping.
|
|
||||||
//
|
|
||||||
// The outer is consumed by the projector.
|
|
||||||
//
|
|
||||||
// We will also be producing:
|
|
||||||
// - A GROUP BY list to group the output of the inner query by non-aggregate variables
|
|
||||||
// so that it can be correctly aggregated.
|
|
||||||
|
|
||||||
// Turn this collection of vars into a collection of columns from the query.
|
|
||||||
// We don't allow grouping on anything but a variable bound in the query.
|
|
||||||
// We group by tag if necessary.
|
|
||||||
let mut group_by = Vec::with_capacity(outer_variables.len() + 2);
|
|
||||||
for var in outer_variables.into_iter() {
|
|
||||||
if query.cc.is_value_bound(&var) {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
// The GROUP BY goes outside, but it needs every variable and type tag to be
|
|
||||||
// projected from inside. Collect in both directions here.
|
|
||||||
let name = VariableColumn::Variable(var.clone()).column_name();
|
|
||||||
group_by.push(GroupBy::ProjectedColumn(name));
|
|
||||||
|
|
||||||
let needs_type_projection = !query.cc.known_type_set(&var).has_unique_type_tag();
|
|
||||||
|
|
||||||
let already_inner = inner_variables.contains(&var);
|
|
||||||
if !already_inner {
|
|
||||||
let (column, name) = candidate_column(&query.cc, &var)?;
|
|
||||||
inner_projection.push(ProjectedColumn(column, name.clone()));
|
|
||||||
}
|
|
||||||
|
|
||||||
if needs_type_projection {
|
|
||||||
let type_name = VariableColumn::VariableTypeTag(var.clone()).column_name();
|
|
||||||
if !already_inner {
|
|
||||||
let type_col = query.cc
|
|
||||||
.extracted_types
|
|
||||||
.get(&var)
|
|
||||||
.cloned()
|
|
||||||
.ok_or_else(|| ErrorKind::NoTypeAvailableForVariable(var.name().clone()))?;
|
|
||||||
inner_projection.push(ProjectedColumn(ColumnOrExpression::Column(type_col), type_name.clone()));
|
|
||||||
}
|
|
||||||
group_by.push(GroupBy::ProjectedColumn(type_name));
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
for var in query.with.iter() {
|
|
||||||
// We never need to project a constant.
|
|
||||||
if query.cc.is_value_bound(&var) {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
// We don't need to add inner projections for :with if they are already there.
|
|
||||||
if !inner_variables.contains(&var) {
|
|
||||||
let (projected_column, type_set) = projected_column_for_var(&var, &query.cc)?;
|
|
||||||
inner_projection.push(projected_column);
|
|
||||||
|
|
||||||
if type_set.unique_type_tag().is_none() {
|
|
||||||
// Also project the type from the SQL query.
|
|
||||||
let (type_column, type_name) = candidate_type_column(&query.cc, &var)?;
|
|
||||||
inner_projection.push(ProjectedColumn(type_column, type_name.clone()));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// At this point we know we have a double-layer projection. Collect the outer.
|
|
||||||
//
|
|
||||||
// If we have an inner and outer layer, the inner layer will name its
|
|
||||||
// variables, and the outer will re-project them.
|
|
||||||
// If we only have one layer, then the outer will do the naming.
|
|
||||||
// (We could try to not use names in the inner query, but then what would we do for
|
|
||||||
// `ground` and known values?)
|
|
||||||
// Walk the projection, switching the outer columns to use the inner names.
|
|
||||||
|
|
||||||
let outer_projection = outer_projection.into_iter().map(|c| {
|
|
||||||
match c {
|
|
||||||
Either::Left(name) => {
|
|
||||||
ProjectedColumn(ColumnOrExpression::ExistingColumn(name.clone()),
|
|
||||||
name)
|
|
||||||
},
|
|
||||||
Either::Right(pc) => pc,
|
|
||||||
}
|
|
||||||
}).collect();
|
|
||||||
|
|
||||||
Ok(ProjectedElements {
|
|
||||||
sql_projection: Projection::Columns(outer_projection),
|
|
||||||
pre_aggregate_projection: Some(Projection::Columns(inner_projection)),
|
|
||||||
templates,
|
|
||||||
group_by,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
pub trait Projector {
|
pub trait Projector {
|
||||||
fn project<'stmt>(&self, rows: Rows<'stmt>) -> Result<QueryOutput>;
|
fn project<'stmt>(&self, rows: Rows<'stmt>) -> Result<QueryOutput>;
|
||||||
fn columns<'s>(&'s self) -> Box<Iterator<Item=&Element> + 's>;
|
fn columns<'s>(&'s self) -> Box<Iterator<Item=&Element> + 's>;
|
||||||
|
|
403
query-projector/src/project.rs
Normal file
403
query-projector/src/project.rs
Normal file
|
@ -0,0 +1,403 @@
|
||||||
|
// Copyright 2018 Mozilla
|
||||||
|
//
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use
|
||||||
|
// this file except in compliance with the License. You may obtain a copy of the
|
||||||
|
// License at http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
// Unless required by applicable law or agreed to in writing, software distributed
|
||||||
|
// under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
|
||||||
|
// CONDITIONS OF ANY KIND, either express or implied. See the License for the
|
||||||
|
// specific language governing permissions and limitations under the License.
|
||||||
|
|
||||||
|
use std::collections::{
|
||||||
|
BTreeSet,
|
||||||
|
};
|
||||||
|
|
||||||
|
use indexmap::{
|
||||||
|
IndexSet,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_core::{
|
||||||
|
SQLValueType,
|
||||||
|
SQLValueTypeSet,
|
||||||
|
ValueTypeSet,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_core::util::{
|
||||||
|
Either,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_query::{
|
||||||
|
Element,
|
||||||
|
Variable,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_query_algebrizer::{
|
||||||
|
AlgebraicQuery,
|
||||||
|
ColumnName,
|
||||||
|
ConjoiningClauses,
|
||||||
|
QualifiedAlias,
|
||||||
|
VariableColumn,
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
|
use mentat_query_sql::{
|
||||||
|
ColumnOrExpression,
|
||||||
|
GroupBy,
|
||||||
|
Name,
|
||||||
|
Projection,
|
||||||
|
ProjectedColumn,
|
||||||
|
};
|
||||||
|
|
||||||
|
use aggregates::{
|
||||||
|
SimpleAggregation,
|
||||||
|
projected_column_for_simple_aggregate,
|
||||||
|
};
|
||||||
|
|
||||||
|
use errors::{
|
||||||
|
ErrorKind,
|
||||||
|
Result,
|
||||||
|
};
|
||||||
|
|
||||||
|
use super::{
|
||||||
|
TypedIndex,
|
||||||
|
};
|
||||||
|
|
||||||
|
/// An internal temporary struct to pass between the projection 'walk' and the
|
||||||
|
/// resultant projector.
|
||||||
|
/// Projection accumulates four things:
|
||||||
|
/// - Two SQL projection lists. We need two because aggregate queries are nested
|
||||||
|
/// in order to apply DISTINCT to values prior to aggregation.
|
||||||
|
/// - A collection of templates for the projector to use to extract values.
|
||||||
|
/// - A list of columns to use for grouping. Grouping is a property of the projection!
|
||||||
|
pub(crate) struct ProjectedElements {
|
||||||
|
pub sql_projection: Projection,
|
||||||
|
pub pre_aggregate_projection: Option<Projection>,
|
||||||
|
pub templates: Vec<TypedIndex>,
|
||||||
|
pub group_by: Vec<GroupBy>,
|
||||||
|
}
|
||||||
|
|
||||||
|
fn candidate_type_column(cc: &ConjoiningClauses, var: &Variable) -> Result<(ColumnOrExpression, Name)> {
|
||||||
|
cc.extracted_types
|
||||||
|
.get(var)
|
||||||
|
.cloned()
|
||||||
|
.map(|alias| {
|
||||||
|
let type_name = VariableColumn::VariableTypeTag(var.clone()).column_name();
|
||||||
|
(ColumnOrExpression::Column(alias), type_name)
|
||||||
|
})
|
||||||
|
.ok_or_else(|| ErrorKind::UnboundVariable(var.name()).into())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn cc_column(cc: &ConjoiningClauses, var: &Variable) -> Result<QualifiedAlias> {
|
||||||
|
cc.column_bindings
|
||||||
|
.get(var)
|
||||||
|
.and_then(|cols| cols.get(0).cloned())
|
||||||
|
.ok_or_else(|| ErrorKind::UnboundVariable(var.name()).into())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn candidate_column(cc: &ConjoiningClauses, var: &Variable) -> Result<(ColumnOrExpression, Name)> {
|
||||||
|
// Every variable should be bound by the top-level CC to at least
|
||||||
|
// one column in the query. If that constraint is violated it's a
|
||||||
|
// bug in our code, so it's appropriate to panic here.
|
||||||
|
cc_column(cc, var)
|
||||||
|
.map(|qa| {
|
||||||
|
let name = VariableColumn::Variable(var.clone()).column_name();
|
||||||
|
(ColumnOrExpression::Column(qa), name)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Return the projected column -- that is, a value or SQL column and an associated name -- for a
|
||||||
|
/// given variable. Also return the type.
|
||||||
|
/// Callers are expected to determine whether to project a type tag as an additional SQL column.
|
||||||
|
pub fn projected_column_for_var(var: &Variable, cc: &ConjoiningClauses) -> Result<(ProjectedColumn, ValueTypeSet)> {
|
||||||
|
if let Some(value) = cc.bound_value(&var) {
|
||||||
|
// If we already know the value, then our lives are easy.
|
||||||
|
let tag = value.value_type();
|
||||||
|
let name = VariableColumn::Variable(var.clone()).column_name();
|
||||||
|
Ok((ProjectedColumn(ColumnOrExpression::Value(value.clone()), name), ValueTypeSet::of_one(tag)))
|
||||||
|
} else {
|
||||||
|
// If we don't, then the CC *must* have bound the variable.
|
||||||
|
let (column, name) = candidate_column(cc, var)?;
|
||||||
|
Ok((ProjectedColumn(column, name), cc.known_type_set(var)))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
/// Walk an iterator of `Element`s, collecting projector templates and columns.
|
||||||
|
///
|
||||||
|
/// Returns a `ProjectedElements`, which combines SQL projections
|
||||||
|
/// and a `Vec` of `TypedIndex` 'keys' to use when looking up values.
|
||||||
|
///
|
||||||
|
/// Callers must ensure that every `Element` is distinct -- a query like
|
||||||
|
///
|
||||||
|
/// ```edn
|
||||||
|
/// [:find ?x ?x :where [?x _ _]]
|
||||||
|
/// ```
|
||||||
|
///
|
||||||
|
/// should fail to parse. See #358.
|
||||||
|
pub(crate) fn project_elements<'a, I: IntoIterator<Item = &'a Element>>(
|
||||||
|
count: usize,
|
||||||
|
elements: I,
|
||||||
|
query: &AlgebraicQuery) -> Result<ProjectedElements> {
|
||||||
|
|
||||||
|
// Give a little padding for type tags.
|
||||||
|
let mut inner_projection = Vec::with_capacity(count + 2);
|
||||||
|
|
||||||
|
// Everything in the outer query will _either_ be an aggregate operation
|
||||||
|
// _or_ a reference to a name projected from the inner.
|
||||||
|
// We'll expand them later.
|
||||||
|
let mut outer_projection: Vec<Either<Name, ProjectedColumn>> = Vec::with_capacity(count + 2);
|
||||||
|
|
||||||
|
let mut i: i32 = 0;
|
||||||
|
let mut min_max_count: usize = 0;
|
||||||
|
let mut corresponding_count: usize = 0;
|
||||||
|
let mut templates = vec![];
|
||||||
|
|
||||||
|
let mut aggregates = false;
|
||||||
|
|
||||||
|
// Any variable that appears intact in the :find clause, not inside an aggregate expression.
|
||||||
|
// "Query variables not in aggregate expressions will group the results and appear intact
|
||||||
|
// in the result."
|
||||||
|
// We use an ordered set here so that we group in the correct order.
|
||||||
|
let mut outer_variables = IndexSet::new();
|
||||||
|
|
||||||
|
// Any variable that we are projecting from the inner query.
|
||||||
|
let mut inner_variables = BTreeSet::new();
|
||||||
|
|
||||||
|
for e in elements {
|
||||||
|
if let &Element::Corresponding(_) = e {
|
||||||
|
corresponding_count += 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
match e {
|
||||||
|
// Each time we come across a variable, we push a SQL column
|
||||||
|
// into the SQL projection, aliased to the name of the variable,
|
||||||
|
// and we push an annotated index into the projector.
|
||||||
|
&Element::Variable(ref var) |
|
||||||
|
&Element::Corresponding(ref var) => {
|
||||||
|
if outer_variables.contains(var) {
|
||||||
|
eprintln!("Warning: duplicate variable {} in query.", var);
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO: it's an error to have `[:find ?x (the ?x) …]`.
|
||||||
|
outer_variables.insert(var.clone());
|
||||||
|
inner_variables.insert(var.clone());
|
||||||
|
|
||||||
|
let (projected_column, type_set) = projected_column_for_var(&var, &query.cc)?;
|
||||||
|
outer_projection.push(Either::Left(projected_column.1.clone()));
|
||||||
|
inner_projection.push(projected_column);
|
||||||
|
|
||||||
|
if let Some(tag) = type_set.unique_type_tag() {
|
||||||
|
templates.push(TypedIndex::Known(i, tag));
|
||||||
|
i += 1; // We used one SQL column.
|
||||||
|
} else {
|
||||||
|
templates.push(TypedIndex::Unknown(i, i + 1));
|
||||||
|
i += 2; // We used two SQL columns.
|
||||||
|
|
||||||
|
// Also project the type from the SQL query.
|
||||||
|
let (type_column, type_name) = candidate_type_column(&query.cc, &var)?;
|
||||||
|
inner_projection.push(ProjectedColumn(type_column, type_name.clone()));
|
||||||
|
outer_projection.push(Either::Left(type_name));
|
||||||
|
}
|
||||||
|
},
|
||||||
|
&Element::Aggregate(ref a) => {
|
||||||
|
if let Some(simple) = a.to_simple() {
|
||||||
|
aggregates = true;
|
||||||
|
|
||||||
|
use aggregates::SimpleAggregationOp::*;
|
||||||
|
match simple.op {
|
||||||
|
Max | Min => {
|
||||||
|
min_max_count += 1;
|
||||||
|
},
|
||||||
|
Avg | Count | Sum => (),
|
||||||
|
}
|
||||||
|
|
||||||
|
// When we encounter a simple aggregate -- one in which the aggregation can be
|
||||||
|
// implemented in SQL, on a single variable -- we just push the SQL aggregation op.
|
||||||
|
// We must ensure the following:
|
||||||
|
// - There's a column for the var.
|
||||||
|
// - The type of the var is known to be restricted to a sensible input set
|
||||||
|
// (not necessarily a single type, but e.g., all vals must be Double or Long).
|
||||||
|
// - The type set must be appropriate for the operation. E.g., `Sum` is not a
|
||||||
|
// meaningful operation on instants.
|
||||||
|
|
||||||
|
let (projected_column, return_type) = projected_column_for_simple_aggregate(&simple, &query.cc)?;
|
||||||
|
outer_projection.push(Either::Right(projected_column));
|
||||||
|
|
||||||
|
if !inner_variables.contains(&simple.var) {
|
||||||
|
inner_variables.insert(simple.var.clone());
|
||||||
|
let (projected_column, _type_set) = projected_column_for_var(&simple.var, &query.cc)?;
|
||||||
|
inner_projection.push(projected_column);
|
||||||
|
if query.cc.known_type_set(&simple.var).unique_type_tag().is_none() {
|
||||||
|
// Also project the type from the SQL query.
|
||||||
|
let (type_column, type_name) = candidate_type_column(&query.cc, &simple.var)?;
|
||||||
|
inner_projection.push(ProjectedColumn(type_column, type_name.clone()));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// We might regret using the type tag here instead of the `ValueType`.
|
||||||
|
templates.push(TypedIndex::Known(i, return_type.value_type_tag()));
|
||||||
|
i += 1;
|
||||||
|
} else {
|
||||||
|
// TODO: complex aggregates.
|
||||||
|
bail!(ErrorKind::NotYetImplemented("complex aggregates".into()));
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
match (min_max_count, corresponding_count) {
|
||||||
|
(0, 0) | (_, 0) => {},
|
||||||
|
(0, _) => {
|
||||||
|
eprintln!("Warning: used `(the ?var)` without `min` or `max`.");
|
||||||
|
},
|
||||||
|
(1, _) => {
|
||||||
|
// This is the success case!
|
||||||
|
},
|
||||||
|
(n, c) => {
|
||||||
|
bail!(ErrorKind::AmbiguousAggregates(n, c));
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
// Anything used in ORDER BY (which we're given in `named_projection`)
|
||||||
|
// needs to be in the SQL column list so we can refer to it by name.
|
||||||
|
//
|
||||||
|
// They don't affect projection.
|
||||||
|
//
|
||||||
|
// If a variable is of a non-fixed type, also project the type tag column, so we don't
|
||||||
|
// accidentally unify across types when considering uniqueness!
|
||||||
|
for var in query.named_projection.iter() {
|
||||||
|
if outer_variables.contains(var) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// If it's a fixed value, we need do nothing further.
|
||||||
|
if query.cc.is_value_bound(&var) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
let already_inner = inner_variables.contains(&var);
|
||||||
|
let (column, name) = candidate_column(&query.cc, &var)?;
|
||||||
|
if !already_inner {
|
||||||
|
inner_projection.push(ProjectedColumn(column, name.clone()));
|
||||||
|
inner_variables.insert(var.clone());
|
||||||
|
}
|
||||||
|
|
||||||
|
outer_projection.push(Either::Left(name));
|
||||||
|
outer_variables.insert(var.clone());
|
||||||
|
|
||||||
|
// We don't care if a column has a single _type_, we care if it has a single type _tag_,
|
||||||
|
// because that's what we'll use if we're projecting. E.g., Long and Double.
|
||||||
|
// Single type implies single type tag, and is cheaper, so we check that first.
|
||||||
|
let types = query.cc.known_type_set(&var);
|
||||||
|
if !types.has_unique_type_tag() {
|
||||||
|
let (type_column, type_name) = candidate_type_column(&query.cc, &var)?;
|
||||||
|
if !already_inner {
|
||||||
|
inner_projection.push(ProjectedColumn(type_column, type_name.clone()));
|
||||||
|
}
|
||||||
|
|
||||||
|
outer_projection.push(Either::Left(type_name));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if !aggregates {
|
||||||
|
// We're done -- we never need to group unless we're aggregating.
|
||||||
|
return Ok(ProjectedElements {
|
||||||
|
sql_projection: Projection::Columns(inner_projection),
|
||||||
|
pre_aggregate_projection: None,
|
||||||
|
templates,
|
||||||
|
group_by: vec![],
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// OK, on to aggregates.
|
||||||
|
// We need to produce two SQL projection lists: one for an inner query and one for the outer.
|
||||||
|
//
|
||||||
|
// The inner serves these purposes:
|
||||||
|
// - Projecting variables to avoid duplicates being elided. (:with)
|
||||||
|
// - Making bindings available to the outermost query for projection, ordering, and grouping.
|
||||||
|
//
|
||||||
|
// The outer is consumed by the projector.
|
||||||
|
//
|
||||||
|
// We will also be producing:
|
||||||
|
// - A GROUP BY list to group the output of the inner query by non-aggregate variables
|
||||||
|
// so that it can be correctly aggregated.
|
||||||
|
|
||||||
|
// Turn this collection of vars into a collection of columns from the query.
|
||||||
|
// We don't allow grouping on anything but a variable bound in the query.
|
||||||
|
// We group by tag if necessary.
|
||||||
|
let mut group_by = Vec::with_capacity(outer_variables.len() + 2);
|
||||||
|
for var in outer_variables.into_iter() {
|
||||||
|
if query.cc.is_value_bound(&var) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// The GROUP BY goes outside, but it needs every variable and type tag to be
|
||||||
|
// projected from inside. Collect in both directions here.
|
||||||
|
let name = VariableColumn::Variable(var.clone()).column_name();
|
||||||
|
group_by.push(GroupBy::ProjectedColumn(name));
|
||||||
|
|
||||||
|
let needs_type_projection = !query.cc.known_type_set(&var).has_unique_type_tag();
|
||||||
|
|
||||||
|
let already_inner = inner_variables.contains(&var);
|
||||||
|
if !already_inner {
|
||||||
|
let (column, name) = candidate_column(&query.cc, &var)?;
|
||||||
|
inner_projection.push(ProjectedColumn(column, name.clone()));
|
||||||
|
}
|
||||||
|
|
||||||
|
if needs_type_projection {
|
||||||
|
let type_name = VariableColumn::VariableTypeTag(var.clone()).column_name();
|
||||||
|
if !already_inner {
|
||||||
|
let type_col = query.cc
|
||||||
|
.extracted_types
|
||||||
|
.get(&var)
|
||||||
|
.cloned()
|
||||||
|
.ok_or_else(|| ErrorKind::NoTypeAvailableForVariable(var.name().clone()))?;
|
||||||
|
inner_projection.push(ProjectedColumn(ColumnOrExpression::Column(type_col), type_name.clone()));
|
||||||
|
}
|
||||||
|
group_by.push(GroupBy::ProjectedColumn(type_name));
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
for var in query.with.iter() {
|
||||||
|
// We never need to project a constant.
|
||||||
|
if query.cc.is_value_bound(&var) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// We don't need to add inner projections for :with if they are already there.
|
||||||
|
if !inner_variables.contains(&var) {
|
||||||
|
let (projected_column, type_set) = projected_column_for_var(&var, &query.cc)?;
|
||||||
|
inner_projection.push(projected_column);
|
||||||
|
|
||||||
|
if type_set.unique_type_tag().is_none() {
|
||||||
|
// Also project the type from the SQL query.
|
||||||
|
let (type_column, type_name) = candidate_type_column(&query.cc, &var)?;
|
||||||
|
inner_projection.push(ProjectedColumn(type_column, type_name.clone()));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// At this point we know we have a double-layer projection. Collect the outer.
|
||||||
|
//
|
||||||
|
// If we have an inner and outer layer, the inner layer will name its
|
||||||
|
// variables, and the outer will re-project them.
|
||||||
|
// If we only have one layer, then the outer will do the naming.
|
||||||
|
// (We could try to not use names in the inner query, but then what would we do for
|
||||||
|
// `ground` and known values?)
|
||||||
|
// Walk the projection, switching the outer columns to use the inner names.
|
||||||
|
|
||||||
|
let outer_projection = outer_projection.into_iter().map(|c| {
|
||||||
|
match c {
|
||||||
|
Either::Left(name) => {
|
||||||
|
ProjectedColumn(ColumnOrExpression::ExistingColumn(name.clone()),
|
||||||
|
name)
|
||||||
|
},
|
||||||
|
Either::Right(pc) => pc,
|
||||||
|
}
|
||||||
|
}).collect();
|
||||||
|
|
||||||
|
Ok(ProjectedElements {
|
||||||
|
sql_projection: Projection::Columns(outer_projection),
|
||||||
|
pre_aggregate_projection: Some(Projection::Columns(inner_projection)),
|
||||||
|
templates,
|
||||||
|
group_by,
|
||||||
|
})
|
||||||
|
}
|
87
query-projector/src/pull.rs
Normal file
87
query-projector/src/pull.rs
Normal file
|
@ -0,0 +1,87 @@
|
||||||
|
// Copyright 2018 Mozilla
|
||||||
|
//
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use
|
||||||
|
// this file except in compliance with the License. You may obtain a copy of the
|
||||||
|
// License at http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
// Unless required by applicable law or agreed to in writing, software distributed
|
||||||
|
// under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
|
||||||
|
// CONDITIONS OF ANY KIND, either express or implied. See the License for the
|
||||||
|
// specific language governing permissions and limitations under the License.
|
||||||
|
|
||||||
|
use std::collections::{
|
||||||
|
BTreeSet,
|
||||||
|
};
|
||||||
|
|
||||||
|
use indexmap::{
|
||||||
|
IndexMap,
|
||||||
|
IndexSet,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_core::{
|
||||||
|
Entid,
|
||||||
|
SQLValueType,
|
||||||
|
SQLValueTypeSet,
|
||||||
|
TypedValue,
|
||||||
|
ValueType,
|
||||||
|
ValueTypeSet,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_core::util::{
|
||||||
|
Either,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_query::{
|
||||||
|
Element,
|
||||||
|
Variable,
|
||||||
|
};
|
||||||
|
|
||||||
|
use mentat_query_algebrizer::{
|
||||||
|
AlgebraicQuery,
|
||||||
|
ColumnName,
|
||||||
|
ConjoiningClauses,
|
||||||
|
QualifiedAlias,
|
||||||
|
VariableColumn,
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
|
use mentat_query_sql::{
|
||||||
|
ColumnOrExpression,
|
||||||
|
GroupBy,
|
||||||
|
Name,
|
||||||
|
Projection,
|
||||||
|
ProjectedColumn,
|
||||||
|
};
|
||||||
|
|
||||||
|
use aggregates::{
|
||||||
|
SimpleAggregation,
|
||||||
|
projected_column_for_simple_aggregate,
|
||||||
|
};
|
||||||
|
|
||||||
|
use errors::{
|
||||||
|
ErrorKind,
|
||||||
|
Result,
|
||||||
|
};
|
||||||
|
|
||||||
|
use super::{
|
||||||
|
TypedIndex,
|
||||||
|
};
|
||||||
|
|
||||||
|
/// A pull expression expands a binding into a structure. The returned structure
|
||||||
|
/// associates attributes named in the input or retrieved from the store with values.
|
||||||
|
/// This association is a `StructuredMap`.
|
||||||
|
struct StructuredMap {
|
||||||
|
attrs: IndexMap<Entid, StructuredValue>,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// The values stored in a `StructuredMap` can be:
|
||||||
|
/// * Vecs of structured values, for multi-valued component attributes or nested expressions.
|
||||||
|
/// * Vecs of typed values, for multi-valued simple attributes. Unlike Datomic, Mentat can express
|
||||||
|
/// an entity without a `{:db/id 12345678}` map.
|
||||||
|
/// * Single structured values, for single-valued component attributes or nested expressions.
|
||||||
|
/// * Single typed values, for simple attributes.
|
||||||
|
enum StructuredValue {
|
||||||
|
Value(TypedValue),
|
||||||
|
Values(Vec<TypedValue>),
|
||||||
|
Structure(StructuredMap),
|
||||||
|
Structures(Vec<StructuredMap>),
|
||||||
|
}
|
|
@ -38,6 +38,6 @@ error_chain! {
|
||||||
}
|
}
|
||||||
|
|
||||||
links {
|
links {
|
||||||
ProjectorError(mentat_query_projector::Error, mentat_query_projector::ErrorKind);
|
ProjectorError(mentat_query_projector::errors::Error, mentat_query_projector::errors::ErrorKind);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -47,7 +47,7 @@ error_chain! {
|
||||||
DbError(mentat_db::Error, mentat_db::ErrorKind);
|
DbError(mentat_db::Error, mentat_db::ErrorKind);
|
||||||
QueryError(mentat_query_algebrizer::Error, mentat_query_algebrizer::ErrorKind); // Let's not leak the term 'algebrizer'.
|
QueryError(mentat_query_algebrizer::Error, mentat_query_algebrizer::ErrorKind); // Let's not leak the term 'algebrizer'.
|
||||||
QueryParseError(mentat_query_parser::Error, mentat_query_parser::ErrorKind);
|
QueryParseError(mentat_query_parser::Error, mentat_query_parser::ErrorKind);
|
||||||
ProjectorError(mentat_query_projector::Error, mentat_query_projector::ErrorKind);
|
ProjectorError(mentat_query_projector::errors::Error, mentat_query_projector::errors::ErrorKind);
|
||||||
TranslatorError(mentat_query_translator::Error, mentat_query_translator::ErrorKind);
|
TranslatorError(mentat_query_translator::Error, mentat_query_translator::ErrorKind);
|
||||||
SqlError(mentat_sql::Error, mentat_sql::ErrorKind);
|
SqlError(mentat_sql::Error, mentat_sql::ErrorKind);
|
||||||
TxParseError(mentat_tx_parser::Error, mentat_tx_parser::ErrorKind);
|
TxParseError(mentat_tx_parser::Error, mentat_tx_parser::ErrorKind);
|
||||||
|
|
|
@ -85,7 +85,9 @@ use std::collections::BTreeMap;
|
||||||
|
|
||||||
pub use mentat_core::attribute;
|
pub use mentat_core::attribute;
|
||||||
use mentat_core::attribute::Unique;
|
use mentat_core::attribute::Unique;
|
||||||
use mentat_core::KnownEntid;
|
use mentat_core::{
|
||||||
|
KnownEntid,
|
||||||
|
};
|
||||||
|
|
||||||
use ::{
|
use ::{
|
||||||
CORE_SCHEMA_VERSION,
|
CORE_SCHEMA_VERSION,
|
||||||
|
@ -126,7 +128,7 @@ pub type Datom = (Entid, Entid, TypedValue);
|
||||||
/// its version number, we need to know the attributes that the application cares about -- it's
|
/// its version number, we need to know the attributes that the application cares about -- it's
|
||||||
/// not enough to know the name and version. Indeed, we even care about the details of each attribute,
|
/// not enough to know the name and version. Indeed, we even care about the details of each attribute,
|
||||||
/// because that's how we'll detect errors.
|
/// because that's how we'll detect errors.
|
||||||
#[derive(Debug)]
|
#[derive(Clone, Debug)]
|
||||||
pub struct Definition {
|
pub struct Definition {
|
||||||
pub name: NamespacedKeyword,
|
pub name: NamespacedKeyword,
|
||||||
pub version: Version,
|
pub version: Version,
|
||||||
|
@ -243,7 +245,7 @@ impl<T> HasCoreSchema for T where T: HasSchema {
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Definition {
|
impl Definition {
|
||||||
fn description_for_attributes<'s, T, R>(&'s self, attributes: &[R], via: &T) -> Result<Terms>
|
fn description_for_attributes<'s, T, R>(&'s self, attributes: &[R], via: &T, diff: Option<BTreeMap<NamespacedKeyword, Attribute>>) -> Result<Terms>
|
||||||
where T: HasCoreSchema,
|
where T: HasCoreSchema,
|
||||||
R: ::std::borrow::Borrow<(NamespacedKeyword, Attribute)> {
|
R: ::std::borrow::Borrow<(NamespacedKeyword, Attribute)> {
|
||||||
|
|
||||||
|
@ -279,13 +281,10 @@ impl Definition {
|
||||||
// Describe each of its attributes.
|
// Describe each of its attributes.
|
||||||
// This is a lot like Schema::to_edn_value; at some point we should tidy this up.
|
// This is a lot like Schema::to_edn_value; at some point we should tidy this up.
|
||||||
for ref r in attributes.iter() {
|
for ref r in attributes.iter() {
|
||||||
let &(ref name, ref attr) = r.borrow();
|
let &(ref kw, ref attr) = r.borrow();
|
||||||
|
|
||||||
// Note that we allow tempid resolution to find an existing entity, if it
|
let tempid = builder.named_tempid(kw.to_string());
|
||||||
// exists. We don't yet support upgrades, which will involve producing
|
let name: TypedValue = kw.clone().into();
|
||||||
// alteration statements.
|
|
||||||
let tempid = builder.named_tempid(name.to_string());
|
|
||||||
let name: TypedValue = name.clone().into();
|
|
||||||
builder.add(tempid.clone(), a_ident, name)?;
|
builder.add(tempid.clone(), a_ident, name)?;
|
||||||
builder.add(schema.clone(), a_attr, tempid.clone())?;
|
builder.add(schema.clone(), a_attr, tempid.clone())?;
|
||||||
|
|
||||||
|
@ -299,18 +298,12 @@ impl Definition {
|
||||||
};
|
};
|
||||||
builder.add(tempid.clone(), a_cardinality, c)?;
|
builder.add(tempid.clone(), a_cardinality, c)?;
|
||||||
|
|
||||||
if attr.index {
|
// These are all unconditional because we use attribute descriptions to _alter_, not
|
||||||
builder.add(tempid.clone(), a_index, TypedValue::Boolean(true))?;
|
// just to _add_, and so absence is distinct from negation!
|
||||||
}
|
builder.add(tempid.clone(), a_index, TypedValue::Boolean(attr.index))?;
|
||||||
if attr.fulltext {
|
builder.add(tempid.clone(), a_fulltext, TypedValue::Boolean(attr.fulltext))?;
|
||||||
builder.add(tempid.clone(), a_fulltext, TypedValue::Boolean(true))?;
|
builder.add(tempid.clone(), a_is_component, TypedValue::Boolean(attr.component))?;
|
||||||
}
|
builder.add(tempid.clone(), a_no_history, TypedValue::Boolean(attr.no_history))?;
|
||||||
if attr.component {
|
|
||||||
builder.add(tempid.clone(), a_is_component, TypedValue::Boolean(true))?;
|
|
||||||
}
|
|
||||||
if attr.no_history {
|
|
||||||
builder.add(tempid.clone(), a_no_history, TypedValue::Boolean(true))?;
|
|
||||||
}
|
|
||||||
|
|
||||||
if let Some(u) = attr.unique {
|
if let Some(u) = attr.unique {
|
||||||
let uu = match u {
|
let uu = match u {
|
||||||
|
@ -318,15 +311,49 @@ impl Definition {
|
||||||
Unique::Value => v_unique_value,
|
Unique::Value => v_unique_value,
|
||||||
};
|
};
|
||||||
builder.add(tempid.clone(), a_unique, uu)?;
|
builder.add(tempid.clone(), a_unique, uu)?;
|
||||||
|
} else {
|
||||||
|
let existing_unique =
|
||||||
|
if let Some(ref diff) = diff {
|
||||||
|
diff.get(kw).and_then(|a| a.unique)
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
};
|
||||||
|
match existing_unique {
|
||||||
|
None => {
|
||||||
|
// Nothing to do.
|
||||||
|
},
|
||||||
|
Some(Unique::Identity) => {
|
||||||
|
builder.retract(tempid.clone(), a_unique, v_unique_identity.clone())?;
|
||||||
|
},
|
||||||
|
Some(Unique::Value) => {
|
||||||
|
builder.retract(tempid.clone(), a_unique, v_unique_value.clone())?;
|
||||||
|
},
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
builder.build()
|
builder.build()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Return a sequence of terms that describes this vocabulary definition and its attributes.
|
||||||
|
fn description_diff<T>(&self, via: &T, from: &Vocabulary) -> Result<Terms> where T: HasSchema {
|
||||||
|
let relevant = self.attributes.iter()
|
||||||
|
.filter_map(|(ref keyword, _)|
|
||||||
|
// Look up the keyword to see if it's currently in use.
|
||||||
|
via.get_entid(keyword)
|
||||||
|
|
||||||
|
// If so, map it to the existing attribute.
|
||||||
|
.and_then(|e| from.find(e).cloned())
|
||||||
|
|
||||||
|
// Collect enough that we can do lookups.
|
||||||
|
.map(|e| (keyword.clone(), e)))
|
||||||
|
.collect();
|
||||||
|
self.description_for_attributes(self.attributes.as_slice(), via, Some(relevant))
|
||||||
|
}
|
||||||
|
|
||||||
/// Return a sequence of terms that describes this vocabulary definition and its attributes.
|
/// Return a sequence of terms that describes this vocabulary definition and its attributes.
|
||||||
fn description<T>(&self, via: &T) -> Result<Terms> where T: HasSchema {
|
fn description<T>(&self, via: &T) -> Result<Terms> where T: HasSchema {
|
||||||
self.description_for_attributes(self.attributes.as_slice(), via)
|
self.description_for_attributes(self.attributes.as_slice(), via, None)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -361,46 +388,8 @@ pub trait HasVocabularies {
|
||||||
fn read_vocabulary_named(&self, name: &NamespacedKeyword) -> Result<Option<Vocabulary>>;
|
fn read_vocabulary_named(&self, name: &NamespacedKeyword) -> Result<Option<Vocabulary>>;
|
||||||
}
|
}
|
||||||
|
|
||||||
pub trait VersionedStore {
|
pub trait VersionedStore: HasVocabularies + HasSchema {
|
||||||
/// Check whether the vocabulary described by the provided metadata is present in the store.
|
/// Check whether the vocabulary described by the provided metadata is present in the store.
|
||||||
fn check_vocabulary<'definition>(&self, definition: &'definition Definition) -> Result<VocabularyCheck<'definition>>;
|
|
||||||
|
|
||||||
/// Check whether the provided vocabulary is present in the store. If it isn't, make it so.
|
|
||||||
fn ensure_vocabulary(&mut self, definition: &Definition) -> Result<VocabularyOutcome>;
|
|
||||||
|
|
||||||
/// Make sure that our expectations of the core vocabulary -- basic types and attributes -- are met.
|
|
||||||
fn verify_core_schema(&self) -> Result<()>;
|
|
||||||
}
|
|
||||||
|
|
||||||
trait VocabularyMechanics {
|
|
||||||
fn install_vocabulary(&mut self, definition: &Definition) -> Result<VocabularyOutcome>;
|
|
||||||
fn install_attributes_for<'definition>(&mut self, definition: &'definition Definition, attributes: Vec<&'definition (NamespacedKeyword, Attribute)>) -> Result<VocabularyOutcome>;
|
|
||||||
fn upgrade_vocabulary(&mut self, definition: &Definition, from_version: Vocabulary) -> Result<VocabularyOutcome>;
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Vocabulary {
|
|
||||||
// TODO: don't do linear search!
|
|
||||||
fn find<T>(&self, entid: T) -> Option<&Attribute> where T: Into<Entid> {
|
|
||||||
let to_find = entid.into();
|
|
||||||
self.attributes.iter().find(|&&(e, _)| e == to_find).map(|&(_, ref a)| a)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<'a, 'c> VersionedStore for InProgress<'a, 'c> {
|
|
||||||
fn verify_core_schema(&self) -> Result<()> {
|
|
||||||
if let Some(core) = self.read_vocabulary_named(&DB_SCHEMA_CORE)? {
|
|
||||||
if core.version != CORE_SCHEMA_VERSION {
|
|
||||||
bail!(ErrorKind::UnexpectedCoreSchema(Some(core.version)));
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: check things other than the version.
|
|
||||||
} else {
|
|
||||||
// This would be seriously messed up.
|
|
||||||
bail!(ErrorKind::UnexpectedCoreSchema(None));
|
|
||||||
}
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
fn check_vocabulary<'definition>(&self, definition: &'definition Definition) -> Result<VocabularyCheck<'definition>> {
|
fn check_vocabulary<'definition>(&self, definition: &'definition Definition) -> Result<VocabularyCheck<'definition>> {
|
||||||
if let Some(vocabulary) = self.read_vocabulary_named(&definition.name)? {
|
if let Some(vocabulary) = self.read_vocabulary_named(&definition.name)? {
|
||||||
// The name is present.
|
// The name is present.
|
||||||
|
@ -449,6 +438,49 @@ impl<'a, 'c> VersionedStore for InProgress<'a, 'c> {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check whether the provided vocabulary is present in the store. If it isn't, make it so.
|
||||||
|
fn ensure_vocabulary(&mut self, definition: &Definition) -> Result<VocabularyOutcome>;
|
||||||
|
|
||||||
|
/// Check whether the provided vocabularies are present in the store at the correct
|
||||||
|
/// version and with all defined attributes. If any are not, invoke the `pre`
|
||||||
|
/// function on the provided `VocabularyProvider`, install or upgrade the necessary vocabularies,
|
||||||
|
/// then invoke `post`. Returns `Ok` if all of these steps succeed.
|
||||||
|
///
|
||||||
|
/// Use this function instead of calling `ensure_vocabulary` if you need to have pre/post
|
||||||
|
/// functions invoked when vocabulary changes are necessary.
|
||||||
|
fn ensure_vocabularies(&mut self, vocabularies: &VocabularyProvider) -> Result<BTreeMap<NamespacedKeyword, VocabularyOutcome>>;
|
||||||
|
|
||||||
|
/// Make sure that our expectations of the core vocabulary -- basic types and attributes -- are met.
|
||||||
|
fn verify_core_schema(&self) -> Result<()> {
|
||||||
|
if let Some(core) = self.read_vocabulary_named(&DB_SCHEMA_CORE)? {
|
||||||
|
if core.version != CORE_SCHEMA_VERSION {
|
||||||
|
bail!(ErrorKind::UnexpectedCoreSchema(Some(core.version)));
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO: check things other than the version.
|
||||||
|
} else {
|
||||||
|
// This would be seriously messed up.
|
||||||
|
bail!(ErrorKind::UnexpectedCoreSchema(None));
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
trait VocabularyMechanics {
|
||||||
|
fn install_vocabulary(&mut self, definition: &Definition) -> Result<VocabularyOutcome>;
|
||||||
|
fn install_attributes_for<'definition>(&mut self, definition: &'definition Definition, attributes: Vec<&'definition (NamespacedKeyword, Attribute)>) -> Result<VocabularyOutcome>;
|
||||||
|
fn upgrade_vocabulary(&mut self, definition: &Definition, from_version: Vocabulary) -> Result<VocabularyOutcome>;
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Vocabulary {
|
||||||
|
// TODO: don't do linear search!
|
||||||
|
fn find<T>(&self, entid: T) -> Option<&Attribute> where T: Into<Entid> {
|
||||||
|
let to_find = entid.into();
|
||||||
|
self.attributes.iter().find(|&&(e, _)| e == to_find).map(|&(_, ref a)| a)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a, 'c> VersionedStore for InProgress<'a, 'c> {
|
||||||
fn ensure_vocabulary(&mut self, definition: &Definition) -> Result<VocabularyOutcome> {
|
fn ensure_vocabulary(&mut self, definition: &Definition) -> Result<VocabularyOutcome> {
|
||||||
match self.check_vocabulary(definition)? {
|
match self.check_vocabulary(definition)? {
|
||||||
VocabularyCheck::Present => Ok(VocabularyOutcome::Existed),
|
VocabularyCheck::Present => Ok(VocabularyOutcome::Existed),
|
||||||
|
@ -458,6 +490,59 @@ impl<'a, 'c> VersionedStore for InProgress<'a, 'c> {
|
||||||
VocabularyCheck::PresentButTooNew { newer_version } => Err(ErrorKind::ExistingVocabularyTooNew(definition.name.to_string(), newer_version.version, definition.version).into()),
|
VocabularyCheck::PresentButTooNew { newer_version } => Err(ErrorKind::ExistingVocabularyTooNew(definition.name.to_string(), newer_version.version, definition.version).into()),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn ensure_vocabularies(&mut self, vocabularies: &VocabularyProvider) -> Result<BTreeMap<NamespacedKeyword, VocabularyOutcome>> {
|
||||||
|
let mut install = Vec::new();
|
||||||
|
let mut update = Vec::new();
|
||||||
|
let mut missing = Vec::new();
|
||||||
|
let mut out = BTreeMap::new();
|
||||||
|
|
||||||
|
for definition in vocabularies.definitions.iter() {
|
||||||
|
match self.check_vocabulary(definition)? {
|
||||||
|
VocabularyCheck::Present => {
|
||||||
|
out.insert(definition.name.clone(), VocabularyOutcome::Existed);
|
||||||
|
},
|
||||||
|
VocabularyCheck::NotPresent => {
|
||||||
|
install.push(definition);
|
||||||
|
},
|
||||||
|
VocabularyCheck::PresentButNeedsUpdate { older_version } => {
|
||||||
|
update.push((definition, older_version));
|
||||||
|
},
|
||||||
|
VocabularyCheck::PresentButMissingAttributes { attributes } => {
|
||||||
|
missing.push((definition, attributes));
|
||||||
|
},
|
||||||
|
VocabularyCheck::PresentButTooNew { newer_version } => {
|
||||||
|
bail!(ErrorKind::ExistingVocabularyTooNew(definition.name.to_string(), newer_version.version, definition.version));
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if install.is_empty() && update.is_empty() && missing.is_empty() {
|
||||||
|
return Ok(out);
|
||||||
|
}
|
||||||
|
|
||||||
|
// If any work needs to be done, run pre/post.
|
||||||
|
(vocabularies.pre)(self)?;
|
||||||
|
|
||||||
|
for d in install {
|
||||||
|
out.insert(d.name.clone(), self.install_vocabulary(d)?);
|
||||||
|
}
|
||||||
|
for (d, v) in update {
|
||||||
|
out.insert(d.name.clone(), self.upgrade_vocabulary(d, v)?);
|
||||||
|
}
|
||||||
|
for (d, a) in missing {
|
||||||
|
out.insert(d.name.clone(), self.install_attributes_for(d, a)?);
|
||||||
|
}
|
||||||
|
|
||||||
|
(vocabularies.post)(self)?;
|
||||||
|
Ok(out)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct VocabularyProvider {
|
||||||
|
pub pre: fn(&mut InProgress) -> Result<()>,
|
||||||
|
pub post: fn(&mut InProgress) -> Result<()>,
|
||||||
|
pub definitions: Vec<Definition>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<'a, 'c> VocabularyMechanics for InProgress<'a, 'c> {
|
impl<'a, 'c> VocabularyMechanics for InProgress<'a, 'c> {
|
||||||
|
@ -469,17 +554,23 @@ impl<'a, 'c> VocabularyMechanics for InProgress<'a, 'c> {
|
||||||
}
|
}
|
||||||
|
|
||||||
fn install_attributes_for<'definition>(&mut self, definition: &'definition Definition, attributes: Vec<&'definition (NamespacedKeyword, Attribute)>) -> Result<VocabularyOutcome> {
|
fn install_attributes_for<'definition>(&mut self, definition: &'definition Definition, attributes: Vec<&'definition (NamespacedKeyword, Attribute)>) -> Result<VocabularyOutcome> {
|
||||||
let (terms, tempids) = definition.description_for_attributes(&attributes, self)?;
|
let (terms, tempids) = definition.description_for_attributes(&attributes, self, None)?;
|
||||||
self.transact_terms(terms, tempids)?;
|
self.transact_terms(terms, tempids)?;
|
||||||
Ok(VocabularyOutcome::InstalledMissingAttributes)
|
Ok(VocabularyOutcome::InstalledMissingAttributes)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Turn the declarative parts of the vocabulary into alterations. Run the 'pre' steps.
|
/// Turn the declarative parts of the vocabulary into alterations. Run the 'pre' steps.
|
||||||
/// Transact the changes. Run the 'post' steps. Return the result and the new `InProgress`!
|
/// Transact the changes. Run the 'post' steps. Return the result and the new `InProgress`!
|
||||||
fn upgrade_vocabulary(&mut self, _definition: &Definition, _from_version: Vocabulary) -> Result<VocabularyOutcome> {
|
fn upgrade_vocabulary(&mut self, definition: &Definition, from_version: Vocabulary) -> Result<VocabularyOutcome> {
|
||||||
unimplemented!();
|
// It's sufficient for us to generate the datom form of each attribute and transact that.
|
||||||
// TODO
|
// We trust that the vocabulary will implement a 'pre' function that cleans up data for any
|
||||||
// Ok(VocabularyOutcome::Installed)
|
// failable conversion (e.g., cardinality-many to cardinality-one).
|
||||||
|
|
||||||
|
// TODO: don't do work for attributes that are unchanged. Here we rely on the transactor
|
||||||
|
// to elide duplicate datoms.
|
||||||
|
let (terms, tempids) = definition.description_diff(self, &from_version)?;
|
||||||
|
self.transact_terms(terms, tempids)?;
|
||||||
|
Ok(VocabularyOutcome::Upgraded)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -584,7 +584,7 @@ fn test_aggregates_type_handling() {
|
||||||
Error(
|
Error(
|
||||||
ErrorKind::TranslatorError(
|
ErrorKind::TranslatorError(
|
||||||
::mentat_query_translator::ErrorKind::ProjectorError(
|
::mentat_query_translator::ErrorKind::ProjectorError(
|
||||||
::mentat_query_projector::ErrorKind::CannotApplyAggregateOperationToTypes(
|
::mentat_query_projector::errors::ErrorKind::CannotApplyAggregateOperationToTypes(
|
||||||
SimpleAggregationOp::Sum,
|
SimpleAggregationOp::Sum,
|
||||||
types
|
types
|
||||||
),
|
),
|
||||||
|
@ -605,7 +605,7 @@ fn test_aggregates_type_handling() {
|
||||||
Error(
|
Error(
|
||||||
ErrorKind::TranslatorError(
|
ErrorKind::TranslatorError(
|
||||||
::mentat_query_translator::ErrorKind::ProjectorError(
|
::mentat_query_translator::ErrorKind::ProjectorError(
|
||||||
::mentat_query_projector::ErrorKind::CannotApplyAggregateOperationToTypes(
|
::mentat_query_projector::errors::ErrorKind::CannotApplyAggregateOperationToTypes(
|
||||||
SimpleAggregationOp::Sum,
|
SimpleAggregationOp::Sum,
|
||||||
types
|
types
|
||||||
),
|
),
|
||||||
|
@ -1173,7 +1173,7 @@ fn test_aggregation_implicit_grouping() {
|
||||||
Error(
|
Error(
|
||||||
ErrorKind::TranslatorError(
|
ErrorKind::TranslatorError(
|
||||||
::mentat_query_translator::ErrorKind::ProjectorError(
|
::mentat_query_translator::ErrorKind::ProjectorError(
|
||||||
::mentat_query_projector::ErrorKind::AmbiguousAggregates(mmc, cc)
|
::mentat_query_projector::errors::ErrorKind::AmbiguousAggregates(mmc, cc)
|
||||||
)
|
)
|
||||||
), _)) => {
|
), _)) => {
|
||||||
assert_eq!(mmc, 2);
|
assert_eq!(mmc, 2);
|
||||||
|
|
|
@ -23,6 +23,7 @@ use mentat::vocabulary::{
|
||||||
VersionedStore,
|
VersionedStore,
|
||||||
VocabularyCheck,
|
VocabularyCheck,
|
||||||
VocabularyOutcome,
|
VocabularyOutcome,
|
||||||
|
VocabularyProvider,
|
||||||
};
|
};
|
||||||
|
|
||||||
use mentat::query::IntoResult;
|
use mentat::query::IntoResult;
|
||||||
|
@ -38,6 +39,7 @@ use mentat::{
|
||||||
Conn,
|
Conn,
|
||||||
NamespacedKeyword,
|
NamespacedKeyword,
|
||||||
Queryable,
|
Queryable,
|
||||||
|
Store,
|
||||||
TypedValue,
|
TypedValue,
|
||||||
ValueType,
|
ValueType,
|
||||||
};
|
};
|
||||||
|
@ -291,4 +293,222 @@ fn test_add_vocab() {
|
||||||
_ => panic!(),
|
_ => panic!(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Some alterations -- cardinality/one to cardinality/many, unique to weaker unique or
|
||||||
|
// no unique, unindexed to indexed -- can be applied automatically, so long as you
|
||||||
|
// bump the version number.
|
||||||
|
|
||||||
|
let multival_bar = vocabulary::AttributeBuilder::helpful()
|
||||||
|
.value_type(ValueType::Instant)
|
||||||
|
.multival(true)
|
||||||
|
.index(true)
|
||||||
|
.build();
|
||||||
|
let multival_bar_and_baz = vec![
|
||||||
|
(kw!(:foo/bar), multival_bar),
|
||||||
|
(kw!(:foo/baz), baz.clone()),
|
||||||
|
];
|
||||||
|
|
||||||
|
let altered_vocabulary = vocabulary::Definition {
|
||||||
|
name: kw!(:org.mozilla/foo),
|
||||||
|
version: 2,
|
||||||
|
attributes: multival_bar_and_baz,
|
||||||
|
};
|
||||||
|
|
||||||
|
// foo/bar starts single-valued.
|
||||||
|
assert_eq!(false, conn.current_schema().attribute_for_ident(&kw!(:foo/bar)).expect("attribute").0.multival);
|
||||||
|
|
||||||
|
// Scoped borrow of `conn`.
|
||||||
|
{
|
||||||
|
let mut in_progress = conn.begin_transaction(&mut sqlite).expect("begun successfully");
|
||||||
|
assert_eq!(in_progress.ensure_vocabulary(&altered_vocabulary).expect("success"),
|
||||||
|
VocabularyOutcome::Upgraded);
|
||||||
|
in_progress.commit().expect("commit succeeded");
|
||||||
|
}
|
||||||
|
|
||||||
|
// Now it's multi-valued.
|
||||||
|
assert_eq!(true, conn.current_schema().attribute_for_ident(&kw!(:foo/bar)).expect("attribute").0.multival);
|
||||||
|
}
|
||||||
|
|
||||||
|
// This is a real-world-style test that evolves a schema with data changes.
|
||||||
|
// We start with a basic vocabulary in three parts:
|
||||||
|
//
|
||||||
|
// Part 1 describes foods by name.
|
||||||
|
// Part 2 describes movies by title.
|
||||||
|
// Part 3 describes people: their names and heights, and their likes.
|
||||||
|
//
|
||||||
|
// We simulate three common migrations:
|
||||||
|
// - We made a trivial modeling error: movie names should not be unique.
|
||||||
|
// - We made a less trivial modeling error, one that can fail: food names should be unique so that
|
||||||
|
// we can more easily refer to them during writes.
|
||||||
|
// In order for this migration to succeed, we need to merge duplicates, then alter the schema --
|
||||||
|
// which we will do by introducing a new property in the same vocabulary, deprecating the old one
|
||||||
|
// -- then transact the transformed data.
|
||||||
|
// - We need to normalize some non-unique data: we recorded heights in inches when they should be
|
||||||
|
// in centimeters.
|
||||||
|
// - We need to normalize some unique data: food names should all be lowercase. Again, that can fail
|
||||||
|
// because of a uniqueness constraint. (We might know that it can't fail thanks to application
|
||||||
|
// restrictions, in which case we can treat this as we did the height alteration.)
|
||||||
|
// - We made a more significant modeling error: we used 'like' to identify both movies and foods,
|
||||||
|
// and we have decided that food preferences and movie preferences should be different attributes.
|
||||||
|
// We wish to split these up and deprecate the old attribute. In order to do so we need to retract
|
||||||
|
// all of the datoms that use the old attribute, transact new attributes _in both movies and foods_,
|
||||||
|
// then re-assert the data.
|
||||||
|
#[test]
|
||||||
|
fn test_upgrade_with_functions() {
|
||||||
|
let mut store = Store::open("").expect("open");
|
||||||
|
|
||||||
|
let food_v1 = vocabulary::Definition {
|
||||||
|
name: kw!(:org.mozilla/food),
|
||||||
|
version: 1,
|
||||||
|
attributes: vec![
|
||||||
|
(kw!(:food/name),
|
||||||
|
vocabulary::AttributeBuilder::helpful()
|
||||||
|
.value_type(ValueType::String)
|
||||||
|
.multival(false)
|
||||||
|
.build()),
|
||||||
|
],
|
||||||
|
};
|
||||||
|
|
||||||
|
let movies_v1 = vocabulary::Definition {
|
||||||
|
name: kw!(:org.mozilla/movies),
|
||||||
|
version: 1,
|
||||||
|
attributes: vec![
|
||||||
|
(kw!(:movie/year),
|
||||||
|
vocabulary::AttributeBuilder::helpful()
|
||||||
|
.value_type(ValueType::Long) // No need for Instant here.
|
||||||
|
.multival(false)
|
||||||
|
.build()),
|
||||||
|
(kw!(:movie/title),
|
||||||
|
vocabulary::AttributeBuilder::helpful()
|
||||||
|
.value_type(ValueType::String)
|
||||||
|
.multival(false)
|
||||||
|
.unique(vocabulary::attribute::Unique::Identity)
|
||||||
|
.index(true)
|
||||||
|
.build()),
|
||||||
|
],
|
||||||
|
};
|
||||||
|
|
||||||
|
let people_v1 = vocabulary::Definition {
|
||||||
|
name: kw!(:org.mozilla/people),
|
||||||
|
version: 1,
|
||||||
|
attributes: vec![
|
||||||
|
(kw!(:person/name),
|
||||||
|
vocabulary::AttributeBuilder::helpful()
|
||||||
|
.value_type(ValueType::String)
|
||||||
|
.multival(false)
|
||||||
|
.unique(vocabulary::attribute::Unique::Identity)
|
||||||
|
.index(true)
|
||||||
|
.build()),
|
||||||
|
(kw!(:person/height),
|
||||||
|
vocabulary::AttributeBuilder::helpful()
|
||||||
|
.value_type(ValueType::Long)
|
||||||
|
.multival(false)
|
||||||
|
.build()),
|
||||||
|
(kw!(:person/likes),
|
||||||
|
vocabulary::AttributeBuilder::helpful()
|
||||||
|
.value_type(ValueType::Ref)
|
||||||
|
.multival(true)
|
||||||
|
.build()),
|
||||||
|
],
|
||||||
|
};
|
||||||
|
|
||||||
|
// Apply v1 of each.
|
||||||
|
let v1_provider = VocabularyProvider {
|
||||||
|
pre: |_ip| Ok(()),
|
||||||
|
definitions: vec![
|
||||||
|
food_v1.clone(),
|
||||||
|
movies_v1.clone(),
|
||||||
|
people_v1.clone(),
|
||||||
|
],
|
||||||
|
post: |_ip| Ok(()),
|
||||||
|
};
|
||||||
|
|
||||||
|
// Mutable borrow of store.
|
||||||
|
{
|
||||||
|
let mut in_progress = store.begin_transaction().expect("began");
|
||||||
|
|
||||||
|
in_progress.ensure_vocabularies(&v1_provider).expect("success");
|
||||||
|
|
||||||
|
// Also add some data. We do this in one transaction 'cos -- thanks to the modeling errors
|
||||||
|
// we are about to fix! -- it's a little awkward to make references to entities without
|
||||||
|
// unique attributes.
|
||||||
|
in_progress.transact(r#"[
|
||||||
|
{:movie/title "John Wick"
|
||||||
|
:movie/year 2014
|
||||||
|
:db/id "mjw"}
|
||||||
|
{:movie/title "Terminator 2: Judgment Day"
|
||||||
|
:movie/year 1991
|
||||||
|
:db/id "mt2"}
|
||||||
|
{:movie/title "Dune"
|
||||||
|
:db/id "md"
|
||||||
|
:movie/year 1984}
|
||||||
|
{:movie/title "Upstream Color"
|
||||||
|
:movie/year 2013
|
||||||
|
:db/id "muc"}
|
||||||
|
{:movie/title "Primer"
|
||||||
|
:db/id "mp"
|
||||||
|
:movie/year 2004}
|
||||||
|
|
||||||
|
;; No year: not yet released.
|
||||||
|
{:movie/title "The Modern Ocean"
|
||||||
|
:db/id "mtmo"}
|
||||||
|
|
||||||
|
{:food/name "Carrots" :db/id "fc"}
|
||||||
|
{:food/name "Weird blue worms" :db/id "fwbw"}
|
||||||
|
{:food/name "Spice" :db/id "fS"}
|
||||||
|
{:food/name "spice" :db/id "fs"}
|
||||||
|
|
||||||
|
;; Sam likes action movies, carrots, and lowercase spice.
|
||||||
|
{:person/name "Sam"
|
||||||
|
:person/height 64
|
||||||
|
:person/likes ["mjw", "mt2", "fc", "fs"]}
|
||||||
|
|
||||||
|
;; Beth likes thoughtful and weird movies, weird blue worms, and Spice.
|
||||||
|
{:person/name "Beth"
|
||||||
|
:person/height 68
|
||||||
|
:person/likes ["muc", "mp", "md", "fwbw", "fS"]}
|
||||||
|
|
||||||
|
]"#).expect("transacted");
|
||||||
|
|
||||||
|
in_progress.commit().expect("commit succeeded");
|
||||||
|
}
|
||||||
|
|
||||||
|
// Mutable borrow of store.
|
||||||
|
{
|
||||||
|
|
||||||
|
// Crap, there are several movies named Dune. We need to de-uniqify that attribute.
|
||||||
|
let movies_v2 = vocabulary::Definition {
|
||||||
|
name: kw!(:org.mozilla/movies),
|
||||||
|
version: 2,
|
||||||
|
attributes: vec![
|
||||||
|
(kw!(:movie/title),
|
||||||
|
vocabulary::AttributeBuilder::helpful()
|
||||||
|
.value_type(ValueType::String)
|
||||||
|
.multival(false)
|
||||||
|
.non_unique()
|
||||||
|
.index(true)
|
||||||
|
.build()),
|
||||||
|
],
|
||||||
|
};
|
||||||
|
let mut in_progress = store.begin_transaction().expect("began");
|
||||||
|
in_progress.ensure_vocabulary(&movies_v2).expect("success");
|
||||||
|
|
||||||
|
// We can now add another Dune movie: Denis Villeneuve's 2019 version.
|
||||||
|
// (Let's just pretend that it's been released, here in 2018!)
|
||||||
|
in_progress.transact(r#"[
|
||||||
|
{:movie/title "Dune"
|
||||||
|
:movie/year 2019}
|
||||||
|
]"#).expect("transact succeeded");
|
||||||
|
|
||||||
|
// And we can query both.
|
||||||
|
let years =
|
||||||
|
in_progress.q_once(r#"[:find [?year ...]
|
||||||
|
:where [?movie :movie/title "Dune"]
|
||||||
|
[?movie :movie/year ?year]
|
||||||
|
:order (asc ?year)]"#, None)
|
||||||
|
.into_coll_result()
|
||||||
|
.expect("coll");
|
||||||
|
assert_eq!(years, vec![TypedValue::Long(1984), TypedValue::Long(2019)]);
|
||||||
|
in_progress.commit().expect("commit succeeded");
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue