Formatting.
This commit is contained in:
parent
1618405c83
commit
953b47036c
65 changed files with 1385 additions and 1342 deletions
|
@ -1,6 +1,7 @@
|
|||
#!/bin/bash
|
||||
|
||||
for f in $(find src -name \*.java); do
|
||||
java -jar ./lib/google-java-format-1.3-all-deps.jar --replace $f
|
||||
for f in $(find ./src -name \*.java); do
|
||||
echo Formatting $f
|
||||
java -jar ./lib/google-java-format-1.3-all-deps.jar --replace $f
|
||||
done
|
||||
|
||||
|
|
|
@ -1,33 +1,33 @@
|
|||
package com.datastax.driver.core.querybuilder;
|
||||
|
||||
import com.datastax.driver.core.CodecRegistry;
|
||||
|
||||
import java.util.List;
|
||||
|
||||
public class IsNotNullClause extends Clause {
|
||||
|
||||
final String name;
|
||||
final String name;
|
||||
|
||||
public IsNotNullClause(String name) {
|
||||
this.name = name;
|
||||
}
|
||||
public IsNotNullClause(String name) {
|
||||
this.name = name;
|
||||
}
|
||||
|
||||
@Override
|
||||
String name() {
|
||||
return name;
|
||||
}
|
||||
@Override
|
||||
String name() {
|
||||
return name;
|
||||
}
|
||||
|
||||
@Override Object firstValue() {
|
||||
return null;
|
||||
}
|
||||
@Override
|
||||
Object firstValue() {
|
||||
return null;
|
||||
}
|
||||
|
||||
@Override
|
||||
void appendTo(StringBuilder sb, List<Object> variables, CodecRegistry codecRegistry) {
|
||||
Utils.appendName(name, sb).append(" IS NOT NULL");
|
||||
}
|
||||
|
||||
@Override boolean containsBindMarker() {
|
||||
return false;
|
||||
}
|
||||
@Override
|
||||
void appendTo(StringBuilder sb, List<Object> variables, CodecRegistry codecRegistry) {
|
||||
Utils.appendName(name, sb).append(" IS NOT NULL");
|
||||
}
|
||||
|
||||
@Override
|
||||
boolean containsBindMarker() {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -5,43 +5,44 @@ import com.datastax.driver.core.querybuilder.Select;
|
|||
|
||||
public class CreateMaterializedView extends Create {
|
||||
|
||||
private String viewName;
|
||||
private Select.Where selection;
|
||||
private String primaryKey;
|
||||
private String viewName;
|
||||
private Select.Where selection;
|
||||
private String primaryKey;
|
||||
|
||||
public CreateMaterializedView(String keyspaceName, String viewName, Select.Where selection, String primaryKey) {
|
||||
super(keyspaceName, viewName);
|
||||
this.viewName = viewName;
|
||||
this.selection = selection;
|
||||
this.primaryKey = primaryKey;
|
||||
public CreateMaterializedView(
|
||||
String keyspaceName, String viewName, Select.Where selection, String primaryKey) {
|
||||
super(keyspaceName, viewName);
|
||||
this.viewName = viewName;
|
||||
this.selection = selection;
|
||||
this.primaryKey = primaryKey;
|
||||
}
|
||||
|
||||
public String getQueryString(CodecRegistry codecRegistry) {
|
||||
return buildInternal();
|
||||
}
|
||||
|
||||
public String buildInternal() {
|
||||
StringBuilder createStatement =
|
||||
new StringBuilder(STATEMENT_START).append("CREATE MATERIALIZED VIEW");
|
||||
if (ifNotExists) {
|
||||
createStatement.append(" IF NOT EXISTS");
|
||||
}
|
||||
|
||||
public String getQueryString(CodecRegistry codecRegistry) {
|
||||
return buildInternal();
|
||||
createStatement.append(" ");
|
||||
if (keyspaceName.isPresent()) {
|
||||
createStatement.append(keyspaceName.get()).append(".");
|
||||
}
|
||||
createStatement.append(viewName);
|
||||
createStatement.append(" AS ");
|
||||
createStatement.append(selection.getQueryString());
|
||||
createStatement.setLength(createStatement.length() - 1);
|
||||
createStatement.append(" ");
|
||||
createStatement.append(primaryKey);
|
||||
createStatement.append(";");
|
||||
|
||||
public String buildInternal() {
|
||||
StringBuilder createStatement = new StringBuilder(STATEMENT_START).append("CREATE MATERIALIZED VIEW");
|
||||
if (ifNotExists) {
|
||||
createStatement.append(" IF NOT EXISTS");
|
||||
}
|
||||
createStatement.append(" ");
|
||||
if (keyspaceName.isPresent()) {
|
||||
createStatement.append(keyspaceName.get()).append(".");
|
||||
}
|
||||
createStatement.append(viewName);
|
||||
createStatement.append(" AS ");
|
||||
createStatement.append(selection.getQueryString());
|
||||
createStatement.setLength(createStatement.length() - 1);
|
||||
createStatement.append(" ");
|
||||
createStatement.append(primaryKey);
|
||||
createStatement.append(";");
|
||||
return createStatement.toString();
|
||||
}
|
||||
|
||||
return createStatement.toString();
|
||||
|
||||
}
|
||||
|
||||
public String toString() {
|
||||
return buildInternal();
|
||||
}
|
||||
public String toString() {
|
||||
return buildInternal();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -4,46 +4,50 @@ import com.google.common.base.Optional;
|
|||
|
||||
public class DropMaterializedView extends Drop {
|
||||
|
||||
enum DroppedItem {TABLE, TYPE, INDEX, MATERIALIZED_VIEW}
|
||||
enum DroppedItem {
|
||||
TABLE,
|
||||
TYPE,
|
||||
INDEX,
|
||||
MATERIALIZED_VIEW
|
||||
}
|
||||
|
||||
private Optional<String> keyspaceName = Optional.absent();
|
||||
private String itemName;
|
||||
private boolean ifExists = true;
|
||||
private final String itemType = "MATERIALIZED VIEW";
|
||||
private Optional<String> keyspaceName = Optional.absent();
|
||||
private String itemName;
|
||||
private boolean ifExists = true;
|
||||
private final String itemType = "MATERIALIZED VIEW";
|
||||
|
||||
public DropMaterializedView(String keyspaceName, String viewName) {
|
||||
this(keyspaceName, viewName, DroppedItem.MATERIALIZED_VIEW);
|
||||
}
|
||||
|
||||
private DropMaterializedView(String keyspaceName, String viewName, DroppedItem itemType) {
|
||||
super(keyspaceName, viewName, Drop.DroppedItem.TABLE);
|
||||
validateNotEmpty(keyspaceName, "Keyspace name");
|
||||
this.keyspaceName = Optional.fromNullable(keyspaceName);
|
||||
this.itemName = viewName;
|
||||
}
|
||||
|
||||
/**
|
||||
* Add the 'IF EXISTS' condition to this DROP statement.
|
||||
*
|
||||
* @return this statement.
|
||||
*/
|
||||
public Drop ifExists() {
|
||||
this.ifExists = true;
|
||||
return this;
|
||||
}
|
||||
|
||||
@Override
|
||||
public String buildInternal() {
|
||||
StringBuilder dropStatement = new StringBuilder("DROP " + itemType + " ");
|
||||
if (ifExists) {
|
||||
dropStatement.append("IF EXISTS ");
|
||||
}
|
||||
if (keyspaceName.isPresent()) {
|
||||
dropStatement.append(keyspaceName.get()).append(".");
|
||||
}
|
||||
|
||||
dropStatement.append(itemName);
|
||||
return dropStatement.toString();
|
||||
public DropMaterializedView(String keyspaceName, String viewName) {
|
||||
this(keyspaceName, viewName, DroppedItem.MATERIALIZED_VIEW);
|
||||
}
|
||||
|
||||
private DropMaterializedView(String keyspaceName, String viewName, DroppedItem itemType) {
|
||||
super(keyspaceName, viewName, Drop.DroppedItem.TABLE);
|
||||
validateNotEmpty(keyspaceName, "Keyspace name");
|
||||
this.keyspaceName = Optional.fromNullable(keyspaceName);
|
||||
this.itemName = viewName;
|
||||
}
|
||||
|
||||
/**
|
||||
* Add the 'IF EXISTS' condition to this DROP statement.
|
||||
*
|
||||
* @return this statement.
|
||||
*/
|
||||
public Drop ifExists() {
|
||||
this.ifExists = true;
|
||||
return this;
|
||||
}
|
||||
|
||||
@Override
|
||||
public String buildInternal() {
|
||||
StringBuilder dropStatement = new StringBuilder("DROP " + itemType + " ");
|
||||
if (ifExists) {
|
||||
dropStatement.append("IF EXISTS ");
|
||||
}
|
||||
if (keyspaceName.isPresent()) {
|
||||
dropStatement.append(keyspaceName.get()).append(".");
|
||||
}
|
||||
|
||||
dropStatement.append(itemName);
|
||||
return dropStatement.toString();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -35,7 +35,7 @@ public enum GetterMethodDetector implements Function<Method, Boolean> {
|
|||
}
|
||||
|
||||
if (Modifier.isStatic(method.getModifiers())) {
|
||||
return false;
|
||||
return false;
|
||||
}
|
||||
|
||||
// Methods marked "Transient" are not mapped, skip them.
|
||||
|
|
|
@ -1,41 +1,38 @@
|
|||
package net.helenus.core;
|
||||
|
||||
import net.helenus.core.reflect.MapExportable;
|
||||
|
||||
import java.time.LocalDateTime;
|
||||
import java.time.ZoneId;
|
||||
import java.util.Date;
|
||||
|
||||
import net.helenus.core.reflect.MapExportable;
|
||||
|
||||
public abstract class AbstractAuditedEntityDraft<E> extends AbstractEntityDraft<E> {
|
||||
|
||||
public AbstractAuditedEntityDraft(MapExportable entity) {
|
||||
super(entity);
|
||||
public AbstractAuditedEntityDraft(MapExportable entity) {
|
||||
super(entity);
|
||||
|
||||
Date in = new Date();
|
||||
LocalDateTime ldt = LocalDateTime.ofInstant(in.toInstant(), ZoneId.systemDefault());
|
||||
Date now = Date.from(ldt.atZone(ZoneId.systemDefault()).toInstant());
|
||||
Date in = new Date();
|
||||
LocalDateTime ldt = LocalDateTime.ofInstant(in.toInstant(), ZoneId.systemDefault());
|
||||
Date now = Date.from(ldt.atZone(ZoneId.systemDefault()).toInstant());
|
||||
|
||||
String who = getCurrentAuditor();
|
||||
String who = getCurrentAuditor();
|
||||
|
||||
if (entity == null) {
|
||||
if (who != null) {
|
||||
set("createdBy", who);
|
||||
}
|
||||
set("createdAt", now);
|
||||
}
|
||||
if (who != null) {
|
||||
set("modifiedBy", who);
|
||||
}
|
||||
set("modifiedAt", now);
|
||||
if (entity == null) {
|
||||
if (who != null) {
|
||||
set("createdBy", who);
|
||||
}
|
||||
set("createdAt", now);
|
||||
}
|
||||
|
||||
protected String getCurrentAuditor() {
|
||||
return null;
|
||||
if (who != null) {
|
||||
set("modifiedBy", who);
|
||||
}
|
||||
set("modifiedAt", now);
|
||||
}
|
||||
|
||||
public Date createdAt() {
|
||||
return (Date) get("createdAt", Date.class);
|
||||
}
|
||||
protected String getCurrentAuditor() {
|
||||
return null;
|
||||
}
|
||||
|
||||
public Date createdAt() {
|
||||
return (Date) get("createdAt", Date.class);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,164 +1,159 @@
|
|||
package net.helenus.core;
|
||||
|
||||
import java.util.*;
|
||||
|
||||
import com.google.common.primitives.Primitives;
|
||||
|
||||
import java.util.*;
|
||||
import net.helenus.core.reflect.DefaultPrimitiveTypes;
|
||||
import net.helenus.core.reflect.Drafted;
|
||||
import net.helenus.core.reflect.MapExportable;
|
||||
import net.helenus.mapping.MappingUtil;
|
||||
|
||||
|
||||
public abstract class AbstractEntityDraft<E> implements Drafted<E> {
|
||||
|
||||
private final Map<String, Object> backingMap = new HashMap<String, Object>();
|
||||
private final MapExportable entity;
|
||||
private final Map<String, Object> entityMap;
|
||||
private final Map<String, Object> backingMap = new HashMap<String, Object>();
|
||||
private final MapExportable entity;
|
||||
private final Map<String, Object> entityMap;
|
||||
|
||||
public AbstractEntityDraft(MapExportable entity) {
|
||||
this.entity = entity;
|
||||
this.entityMap = entity != null ? entity.toMap() : new HashMap<String, Object>();
|
||||
}
|
||||
|
||||
public AbstractEntityDraft(MapExportable entity) {
|
||||
this.entity = entity;
|
||||
this.entityMap = entity != null ? entity.toMap() : new HashMap<String, Object>();
|
||||
}
|
||||
public abstract Class<E> getEntityClass();
|
||||
|
||||
public abstract Class<E> getEntityClass();
|
||||
public E build() {
|
||||
return Helenus.map(getEntityClass(), toMap());
|
||||
}
|
||||
|
||||
public E build() { return Helenus.map(getEntityClass(), toMap()); }
|
||||
@SuppressWarnings("unchecked")
|
||||
protected <T> T get(Getter<T> getter, Class<?> returnType) {
|
||||
return (T) get(this.<T>methodNameFor(getter), returnType);
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
protected <T> T get(Getter<T> getter, Class<?> returnType) {
|
||||
return (T) get(this.<T>methodNameFor(getter), returnType);
|
||||
}
|
||||
@SuppressWarnings("unchecked")
|
||||
protected <T> T get(String key, Class<?> returnType) {
|
||||
T value = (T) backingMap.get(key);
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
protected <T> T get(String key, Class<?> returnType) {
|
||||
T value = (T) backingMap.get(key);
|
||||
if (value == null) {
|
||||
value = (T) entityMap.get(key);
|
||||
if (value == null) {
|
||||
|
||||
if (value == null) {
|
||||
value = (T) entityMap.get(key);
|
||||
if (value == null) {
|
||||
if (Primitives.allPrimitiveTypes().contains(returnType)) {
|
||||
|
||||
if (Primitives.allPrimitiveTypes().contains(returnType)) {
|
||||
DefaultPrimitiveTypes type = DefaultPrimitiveTypes.lookup(returnType);
|
||||
if (type == null) {
|
||||
throw new RuntimeException("unknown primitive type " + returnType);
|
||||
}
|
||||
|
||||
DefaultPrimitiveTypes type = DefaultPrimitiveTypes.lookup(returnType);
|
||||
if (type == null) {
|
||||
throw new RuntimeException("unknown primitive type " + returnType);
|
||||
}
|
||||
|
||||
return (T) type.getDefaultValue();
|
||||
}
|
||||
}
|
||||
return (T) type.getDefaultValue();
|
||||
}
|
||||
|
||||
return value;
|
||||
}
|
||||
}
|
||||
|
||||
protected <T> Object set(Getter<T> getter, Object value) {
|
||||
return set(this.<T>methodNameFor(getter), value);
|
||||
return value;
|
||||
}
|
||||
|
||||
protected <T> Object set(Getter<T> getter, Object value) {
|
||||
return set(this.<T>methodNameFor(getter), value);
|
||||
}
|
||||
|
||||
protected Object set(String key, Object value) {
|
||||
if (key == null || value == null) {
|
||||
return null;
|
||||
}
|
||||
|
||||
protected Object set(String key, Object value) {
|
||||
if (key == null || value == null) {
|
||||
return null;
|
||||
}
|
||||
backingMap.put(key, value);
|
||||
return value;
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
protected <T> T mutate(Getter<T> getter, T value) {
|
||||
return (T) mutate(this.<T>methodNameFor(getter), value);
|
||||
}
|
||||
|
||||
protected Object mutate(String key, Object value) {
|
||||
Objects.requireNonNull(key);
|
||||
|
||||
if (value == null) {
|
||||
return null;
|
||||
}
|
||||
|
||||
if (entity != null) {
|
||||
Map<String, Object> map = entity.toMap();
|
||||
|
||||
if (map.containsKey(key) && !value.equals(map.get(key))) {
|
||||
backingMap.put(key, value);
|
||||
return value;
|
||||
}
|
||||
|
||||
return map.get(key);
|
||||
} else {
|
||||
backingMap.put(key, value);
|
||||
|
||||
return null;
|
||||
}
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
protected <T> T mutate(Getter<T> getter, T value) {
|
||||
return (T) mutate(this.<T>methodNameFor(getter), value);
|
||||
private <T> String methodNameFor(Getter<T> getter) {
|
||||
return MappingUtil.resolveMappingProperty(getter).getProperty().getPropertyName();
|
||||
}
|
||||
|
||||
public <T> Object unset(Getter<T> getter) {
|
||||
return unset(methodNameFor(getter));
|
||||
}
|
||||
|
||||
public Object unset(String key) {
|
||||
if (key != null) {
|
||||
Object value = backingMap.get(key);
|
||||
backingMap.put(key, null);
|
||||
return value;
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
protected Object mutate(String key, Object value) {
|
||||
Objects.requireNonNull(key);
|
||||
public <T> boolean reset(Getter<T> getter, T desiredValue) {
|
||||
return this.<T>reset(this.<T>methodNameFor(getter), desiredValue);
|
||||
}
|
||||
|
||||
if (value == null) {
|
||||
return null;
|
||||
}
|
||||
|
||||
if (entity != null) {
|
||||
Map<String, Object> map = entity.toMap();
|
||||
|
||||
if (map.containsKey(key) && !value.equals(map.get(key))) {
|
||||
backingMap.put(key, value);
|
||||
return value;
|
||||
}
|
||||
|
||||
return map.get(key);
|
||||
} else {
|
||||
backingMap.put(key, value);
|
||||
|
||||
return null;
|
||||
}
|
||||
public <T> boolean reset(String key, T desiredValue) {
|
||||
if (key != null && desiredValue != null) {
|
||||
@SuppressWarnings("unchecked")
|
||||
T currentValue = (T) backingMap.get(key);
|
||||
if (currentValue == null || !currentValue.equals(desiredValue)) {
|
||||
set(key, desiredValue);
|
||||
return true;
|
||||
}
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
private <T> String methodNameFor(Getter<T> getter) {
|
||||
return MappingUtil.resolveMappingProperty(getter)
|
||||
.getProperty()
|
||||
.getPropertyName();
|
||||
@Override
|
||||
public Map<String, Object> toMap() {
|
||||
return toMap(entityMap);
|
||||
}
|
||||
|
||||
public Map<String, Object> toMap(Map<String, Object> entityMap) {
|
||||
Map<String, Object> combined;
|
||||
if (entityMap != null && entityMap.size() > 0) {
|
||||
combined = new HashMap<String, Object>(entityMap.size());
|
||||
for (String key : entityMap.keySet()) {
|
||||
combined.put(key, entityMap.get(key));
|
||||
}
|
||||
} else {
|
||||
combined = new HashMap<String, Object>(backingMap.size());
|
||||
}
|
||||
|
||||
public <T> Object unset(Getter<T> getter) {
|
||||
return unset(methodNameFor(getter));
|
||||
for (String key : mutated()) {
|
||||
combined.put(key, backingMap.get(key));
|
||||
}
|
||||
return combined;
|
||||
}
|
||||
|
||||
public Object unset(String key) {
|
||||
if (key != null) {
|
||||
Object value = backingMap.get(key);
|
||||
backingMap.put(key, null);
|
||||
return value;
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
public <T> boolean reset(Getter<T> getter, T desiredValue) {
|
||||
return this.<T>reset(this.<T>methodNameFor(getter), desiredValue);
|
||||
}
|
||||
|
||||
public <T> boolean reset(String key, T desiredValue) {
|
||||
if (key != null && desiredValue != null) {
|
||||
@SuppressWarnings("unchecked")
|
||||
T currentValue = (T) backingMap.get(key);
|
||||
if (currentValue == null || !currentValue.equals(desiredValue)) {
|
||||
set(key, desiredValue);
|
||||
return true;
|
||||
}
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
@Override
|
||||
public Map<String, Object> toMap() {
|
||||
return toMap(entityMap);
|
||||
}
|
||||
|
||||
public Map<String, Object> toMap(Map<String, Object>entityMap) {
|
||||
Map<String, Object> combined;
|
||||
if (entityMap != null && entityMap.size() > 0) {
|
||||
combined = new HashMap<String, Object>(entityMap.size());
|
||||
for (String key : entityMap.keySet()) {
|
||||
combined.put(key, entityMap.get(key));
|
||||
}
|
||||
} else {
|
||||
combined = new HashMap<String, Object>(backingMap.size());
|
||||
}
|
||||
for (String key : mutated()) {
|
||||
combined.put(key, backingMap.get(key));
|
||||
}
|
||||
return combined;
|
||||
}
|
||||
|
||||
@Override
|
||||
public Set<String> mutated() {
|
||||
return backingMap.keySet();
|
||||
}
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return backingMap.toString();
|
||||
}
|
||||
@Override
|
||||
public Set<String> mutated() {
|
||||
return backingMap.keySet();
|
||||
}
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return backingMap.toString();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -22,7 +22,6 @@ import com.datastax.driver.core.querybuilder.BuiltStatement;
|
|||
import com.google.common.util.concurrent.ListenableFuture;
|
||||
import java.io.PrintStream;
|
||||
import java.util.concurrent.Executor;
|
||||
|
||||
import net.helenus.mapping.value.ColumnValuePreparer;
|
||||
import net.helenus.mapping.value.ColumnValueProvider;
|
||||
import net.helenus.support.HelenusException;
|
||||
|
@ -124,5 +123,4 @@ public abstract class AbstractSessionOperations {
|
|||
void printCql(String cql) {
|
||||
getPrintStream().println(cql);
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -17,10 +17,8 @@ package net.helenus.core;
|
|||
|
||||
import com.diffplug.common.base.Errors;
|
||||
import com.google.common.collect.TreeTraverser;
|
||||
|
||||
import java.util.*;
|
||||
|
||||
|
||||
/** Encapsulates the concept of a "transaction" as a unit-of-work. */
|
||||
public abstract class AbstractUnitOfWork<E extends Exception> implements UnitOfWork, AutoCloseable {
|
||||
private final List<AbstractUnitOfWork<E>> nested = new ArrayList<>();
|
||||
|
@ -70,22 +68,25 @@ public abstract class AbstractUnitOfWork<E extends Exception> implements UnitOfW
|
|||
return null;
|
||||
}
|
||||
|
||||
public Map<String, Set<Object>> getCache() { return cache; }
|
||||
public Map<String, Set<Object>> getCache() {
|
||||
return cache;
|
||||
}
|
||||
|
||||
private Iterator<AbstractUnitOfWork<E>> getChildNodes() {
|
||||
return nested.iterator();
|
||||
}
|
||||
|
||||
/**
|
||||
* Checks to see if the work performed between calling begin and now can be committed or not.
|
||||
*
|
||||
* @return a function from which to chain work that only happens when commit is successful
|
||||
* @throws E when the work overlaps with other concurrent writers.
|
||||
*/
|
||||
/**
|
||||
* Checks to see if the work performed between calling begin and now can be committed or not.
|
||||
*
|
||||
* @return a function from which to chain work that only happens when commit is successful
|
||||
* @throws E when the work overlaps with other concurrent writers.
|
||||
*/
|
||||
public PostCommitFunction<Void, Void> commit() throws E {
|
||||
// All nested UnitOfWork should be committed (not aborted) before calls to commit, check.
|
||||
boolean canCommit = true;
|
||||
TreeTraverser<AbstractUnitOfWork<E>> traverser = TreeTraverser.using(node -> node::getChildNodes);
|
||||
TreeTraverser<AbstractUnitOfWork<E>> traverser =
|
||||
TreeTraverser.using(node -> node::getChildNodes);
|
||||
for (AbstractUnitOfWork<E> uow : traverser.postOrderTraversal(this)) {
|
||||
if (this != uow) {
|
||||
canCommit &= (!uow.aborted && uow.committed);
|
||||
|
@ -112,7 +113,8 @@ public abstract class AbstractUnitOfWork<E extends Exception> implements UnitOfW
|
|||
if (parentCache.containsKey(key)) {
|
||||
// merge the sets
|
||||
Set<Object> ps = parentCache.get(key);
|
||||
ps.addAll(cache.get(key)); //TODO(gburd): review this, likely not correct in all cases as-is.
|
||||
ps.addAll(
|
||||
cache.get(key)); //TODO(gburd): review this, likely not correct in all cases as-is.
|
||||
} else {
|
||||
// add the missing set
|
||||
parentCache.put(key, cache.get(key));
|
||||
|
@ -122,9 +124,12 @@ public abstract class AbstractUnitOfWork<E extends Exception> implements UnitOfW
|
|||
|
||||
// Apply all post-commit functions for
|
||||
if (parent == null) {
|
||||
traverser.postOrderTraversal(this).forEach(uow -> {
|
||||
uow.applyPostCommitFunctions();
|
||||
});
|
||||
traverser
|
||||
.postOrderTraversal(this)
|
||||
.forEach(
|
||||
uow -> {
|
||||
uow.applyPostCommitFunctions();
|
||||
});
|
||||
return new PostCommitFunction(this, null);
|
||||
}
|
||||
}
|
||||
|
@ -137,11 +142,15 @@ public abstract class AbstractUnitOfWork<E extends Exception> implements UnitOfW
|
|||
|
||||
/* Explicitly discard the work and mark it as as such in the log. */
|
||||
public void abort() {
|
||||
TreeTraverser<AbstractUnitOfWork<E>> traverser = TreeTraverser.using(node -> node::getChildNodes);
|
||||
traverser.postOrderTraversal(this).forEach(uow -> {
|
||||
uow.committed = false;
|
||||
uow.aborted = true;
|
||||
});
|
||||
TreeTraverser<AbstractUnitOfWork<E>> traverser =
|
||||
TreeTraverser.using(node -> node::getChildNodes);
|
||||
traverser
|
||||
.postOrderTraversal(this)
|
||||
.forEach(
|
||||
uow -> {
|
||||
uow.committed = false;
|
||||
uow.aborted = true;
|
||||
});
|
||||
// log.record(txn::abort)
|
||||
// cache.invalidateSince(txn::start time)
|
||||
}
|
||||
|
@ -165,5 +174,4 @@ public abstract class AbstractUnitOfWork<E extends Exception> implements UnitOfW
|
|||
public boolean hasCommitted() {
|
||||
return committed;
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -1,8 +1,7 @@
|
|||
package net.helenus.core;
|
||||
|
||||
import java.util.function.Function;
|
||||
|
||||
@FunctionalInterface
|
||||
public interface CommitThunk {
|
||||
void apply();
|
||||
void apply();
|
||||
}
|
||||
|
|
|
@ -142,10 +142,10 @@ public final class Helenus {
|
|||
public static HelenusEntity entity(Class<?> iface) {
|
||||
Metadata metadata = metadataForEntity.get(iface);
|
||||
if (metadata == null) {
|
||||
HelenusSession session = session();
|
||||
if (session != null) {
|
||||
metadata = session.getMetadata();
|
||||
}
|
||||
HelenusSession session = session();
|
||||
if (session != null) {
|
||||
metadata = session.getMetadata();
|
||||
}
|
||||
}
|
||||
return entity(iface, metadata);
|
||||
}
|
||||
|
|
|
@ -15,10 +15,21 @@
|
|||
*/
|
||||
package net.helenus.core;
|
||||
|
||||
import static net.helenus.core.Query.eq;
|
||||
|
||||
import brave.Tracer;
|
||||
import com.codahale.metrics.Meter;
|
||||
import com.codahale.metrics.MetricRegistry;
|
||||
import com.datastax.driver.core.*;
|
||||
import java.io.Closeable;
|
||||
import java.io.PrintStream;
|
||||
import java.lang.reflect.Constructor;
|
||||
import java.lang.reflect.InvocationTargetException;
|
||||
import java.util.Map;
|
||||
import java.util.Objects;
|
||||
import java.util.Optional;
|
||||
import java.util.Set;
|
||||
import java.util.concurrent.Executor;
|
||||
import java.util.function.Function;
|
||||
import net.helenus.core.operation.*;
|
||||
import net.helenus.core.reflect.Drafted;
|
||||
import net.helenus.core.reflect.HelenusPropertyNode;
|
||||
|
@ -33,19 +44,6 @@ import net.helenus.support.Fun.Tuple6;
|
|||
import net.helenus.support.HelenusException;
|
||||
import net.helenus.support.HelenusMappingException;
|
||||
|
||||
import java.io.Closeable;
|
||||
import java.io.PrintStream;
|
||||
import java.lang.reflect.Constructor;
|
||||
import java.lang.reflect.InvocationTargetException;
|
||||
import java.util.Map;
|
||||
import java.util.Objects;
|
||||
import java.util.Optional;
|
||||
import java.util.Set;
|
||||
import java.util.concurrent.Executor;
|
||||
import java.util.function.Function;
|
||||
|
||||
import static net.helenus.core.Query.eq;
|
||||
|
||||
public final class HelenusSession extends AbstractSessionOperations implements Closeable {
|
||||
|
||||
private final int MAX_CACHE_SIZE = 10000;
|
||||
|
@ -69,26 +67,25 @@ public final class HelenusSession extends AbstractSessionOperations implements C
|
|||
private final StatementColumnValuePreparer valuePreparer;
|
||||
private final Metadata metadata;
|
||||
|
||||
|
||||
HelenusSession(
|
||||
Session session,
|
||||
String usingKeyspace,
|
||||
CodecRegistry registry,
|
||||
boolean showCql,
|
||||
PrintStream printStream,
|
||||
SessionRepositoryBuilder sessionRepositoryBuilder,
|
||||
Executor executor,
|
||||
boolean dropSchemaOnClose,
|
||||
ConsistencyLevel consistencyLevel,
|
||||
boolean defaultQueryIdempotency,
|
||||
Class<? extends UnitOfWork> unitOfWorkClass,
|
||||
MetricRegistry metricRegistry,
|
||||
Tracer tracer) {
|
||||
Session session,
|
||||
String usingKeyspace,
|
||||
CodecRegistry registry,
|
||||
boolean showCql,
|
||||
PrintStream printStream,
|
||||
SessionRepositoryBuilder sessionRepositoryBuilder,
|
||||
Executor executor,
|
||||
boolean dropSchemaOnClose,
|
||||
ConsistencyLevel consistencyLevel,
|
||||
boolean defaultQueryIdempotency,
|
||||
Class<? extends UnitOfWork> unitOfWorkClass,
|
||||
MetricRegistry metricRegistry,
|
||||
Tracer tracer) {
|
||||
this.session = session;
|
||||
this.registry = registry == null ? CodecRegistry.DEFAULT_INSTANCE : registry;
|
||||
this.usingKeyspace =
|
||||
Objects.requireNonNull(
|
||||
usingKeyspace, "keyspace needs to be selected before creating session");
|
||||
Objects.requireNonNull(
|
||||
usingKeyspace, "keyspace needs to be selected before creating session");
|
||||
this.showCql = showCql;
|
||||
this.printStream = printStream;
|
||||
this.sessionRepository = sessionRepositoryBuilder.build();
|
||||
|
@ -177,42 +174,53 @@ public final class HelenusSession extends AbstractSessionOperations implements C
|
|||
}
|
||||
|
||||
@Override
|
||||
public boolean getDefaultQueryIdempotency() { return defaultQueryIdempotency; }
|
||||
public boolean getDefaultQueryIdempotency() {
|
||||
return defaultQueryIdempotency;
|
||||
}
|
||||
|
||||
public Metadata getMetadata() { return metadata; }
|
||||
public Metadata getMetadata() {
|
||||
return metadata;
|
||||
}
|
||||
|
||||
public synchronized UnitOfWork begin() {
|
||||
return begin(null);
|
||||
return begin(null);
|
||||
}
|
||||
|
||||
public synchronized UnitOfWork begin(UnitOfWork parent) {
|
||||
try {
|
||||
Class<? extends UnitOfWork> clazz = unitOfWorkClass;
|
||||
Constructor<? extends UnitOfWork> ctor = clazz.getConstructor(HelenusSession.class, UnitOfWork.class);
|
||||
Constructor<? extends UnitOfWork> ctor =
|
||||
clazz.getConstructor(HelenusSession.class, UnitOfWork.class);
|
||||
UnitOfWork uow = ctor.newInstance(this, parent);
|
||||
if (parent != null) {
|
||||
parent.addNestedUnitOfWork(uow);
|
||||
}
|
||||
return uow.begin();
|
||||
}
|
||||
catch (NoSuchMethodException | InvocationTargetException | InstantiationException | IllegalAccessException e) {
|
||||
throw new HelenusException(String.format("Unable to instantiate {} as a UnitOfWork.", unitOfWorkClass.getSimpleName()), e);
|
||||
} catch (NoSuchMethodException
|
||||
| InvocationTargetException
|
||||
| InstantiationException
|
||||
| IllegalAccessException e) {
|
||||
throw new HelenusException(
|
||||
String.format(
|
||||
"Unable to instantiate {} as a UnitOfWork.", unitOfWorkClass.getSimpleName()),
|
||||
e);
|
||||
}
|
||||
}
|
||||
|
||||
public <E> SelectOperation<E> select(E pojo) {
|
||||
Objects.requireNonNull(pojo, "supplied object must be a dsl for a registered entity but cannot be null");
|
||||
Objects.requireNonNull(
|
||||
pojo, "supplied object must be a dsl for a registered entity but cannot be null");
|
||||
ColumnValueProvider valueProvider = getValueProvider();
|
||||
HelenusEntity entity = Helenus.resolve(pojo);
|
||||
Class<?> entityClass = entity.getMappingInterface();
|
||||
|
||||
return new SelectOperation<E>(
|
||||
this,
|
||||
entity,
|
||||
(r) -> {
|
||||
Map<String, Object> map = new ValueProviderMap(r, valueProvider, entity);
|
||||
return (E) Helenus.map(entityClass, map);
|
||||
});
|
||||
this,
|
||||
entity,
|
||||
(r) -> {
|
||||
Map<String, Object> map = new ValueProviderMap(r, valueProvider, entity);
|
||||
return (E) Helenus.map(entityClass, map);
|
||||
});
|
||||
}
|
||||
|
||||
public <E> SelectOperation<E> select(Class<E> entityClass) {
|
||||
|
@ -221,12 +229,12 @@ public final class HelenusSession extends AbstractSessionOperations implements C
|
|||
HelenusEntity entity = Helenus.entity(entityClass);
|
||||
|
||||
return new SelectOperation<E>(
|
||||
this,
|
||||
entity,
|
||||
(r) -> {
|
||||
Map<String, Object> map = new ValueProviderMap(r, valueProvider, entity);
|
||||
return (E) Helenus.map(entityClass, map);
|
||||
});
|
||||
this,
|
||||
entity,
|
||||
(r) -> {
|
||||
Map<String, Object> map = new ValueProviderMap(r, valueProvider, entity);
|
||||
return (E) Helenus.map(entityClass, map);
|
||||
});
|
||||
}
|
||||
|
||||
public SelectOperation<Fun.ArrayTuple> select() {
|
||||
|
@ -239,7 +247,8 @@ public final class HelenusSession extends AbstractSessionOperations implements C
|
|||
}
|
||||
|
||||
public <E> SelectOperation<Row> selectAll(E pojo) {
|
||||
Objects.requireNonNull(pojo, "supplied object must be a dsl for a registered entity but cannot be null");
|
||||
Objects.requireNonNull(
|
||||
pojo, "supplied object must be a dsl for a registered entity but cannot be null");
|
||||
HelenusEntity entity = Helenus.resolve(pojo);
|
||||
return new SelectOperation<Row>(this, entity);
|
||||
}
|
||||
|
@ -411,51 +420,62 @@ public final class HelenusSession extends AbstractSessionOperations implements C
|
|||
|
||||
public <E> UpdateOperation<E> update(Drafted<E> drafted) {
|
||||
if (drafted instanceof AbstractEntityDraft == false) {
|
||||
throw new HelenusMappingException("update of draft objects that don't inherit from AbstractEntityDraft is not yet supported");
|
||||
throw new HelenusMappingException(
|
||||
"update of draft objects that don't inherit from AbstractEntityDraft is not yet supported");
|
||||
}
|
||||
AbstractEntityDraft<E> draft = (AbstractEntityDraft<E>)drafted;
|
||||
AbstractEntityDraft<E> draft = (AbstractEntityDraft<E>) drafted;
|
||||
UpdateOperation update = new UpdateOperation<E>(this, draft);
|
||||
Map<String, Object> map = draft.toMap();
|
||||
Set<String> mutatedProperties = draft.mutated();
|
||||
HelenusEntity entity = Helenus.entity(draft.getEntityClass());
|
||||
|
||||
// Add all the mutated values contained in the draft.
|
||||
entity.getOrderedProperties().forEach(property -> {
|
||||
switch (property.getColumnType()) {
|
||||
case PARTITION_KEY:
|
||||
case CLUSTERING_COLUMN:
|
||||
break;
|
||||
default:
|
||||
String propertyName = property.getPropertyName();
|
||||
if (mutatedProperties.contains(propertyName)) {
|
||||
Object value = map.get(propertyName);
|
||||
Getter<Object> getter = new Getter<Object>() {
|
||||
@Override
|
||||
public Object get() {
|
||||
throw new DslPropertyException(new HelenusPropertyNode(property, Optional.empty()));
|
||||
entity
|
||||
.getOrderedProperties()
|
||||
.forEach(
|
||||
property -> {
|
||||
switch (property.getColumnType()) {
|
||||
case PARTITION_KEY:
|
||||
case CLUSTERING_COLUMN:
|
||||
break;
|
||||
default:
|
||||
String propertyName = property.getPropertyName();
|
||||
if (mutatedProperties.contains(propertyName)) {
|
||||
Object value = map.get(propertyName);
|
||||
Getter<Object> getter =
|
||||
new Getter<Object>() {
|
||||
@Override
|
||||
public Object get() {
|
||||
throw new DslPropertyException(
|
||||
new HelenusPropertyNode(property, Optional.empty()));
|
||||
}
|
||||
};
|
||||
update.set(getter, value);
|
||||
}
|
||||
}
|
||||
};
|
||||
update.set(getter, value);
|
||||
}
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
// Add the partition and clustering keys if they were in the draft (normally the case).
|
||||
entity.getOrderedProperties().forEach(property -> {
|
||||
switch (property.getColumnType()) {
|
||||
case PARTITION_KEY:
|
||||
case CLUSTERING_COLUMN:
|
||||
String propertyName = property.getPropertyName();
|
||||
Object value = map.get(propertyName);
|
||||
Getter<Object> getter = new Getter<Object>() {
|
||||
@Override
|
||||
public Object get() {
|
||||
throw new DslPropertyException(new HelenusPropertyNode(property, Optional.empty()));
|
||||
}
|
||||
};
|
||||
update.where(getter, eq(value));
|
||||
}
|
||||
});
|
||||
entity
|
||||
.getOrderedProperties()
|
||||
.forEach(
|
||||
property -> {
|
||||
switch (property.getColumnType()) {
|
||||
case PARTITION_KEY:
|
||||
case CLUSTERING_COLUMN:
|
||||
String propertyName = property.getPropertyName();
|
||||
Object value = map.get(propertyName);
|
||||
Getter<Object> getter =
|
||||
new Getter<Object>() {
|
||||
@Override
|
||||
public Object get() {
|
||||
throw new DslPropertyException(
|
||||
new HelenusPropertyNode(property, Optional.empty()));
|
||||
}
|
||||
};
|
||||
update.where(getter, eq(value));
|
||||
}
|
||||
});
|
||||
|
||||
return update;
|
||||
}
|
||||
|
@ -478,9 +498,14 @@ public final class HelenusSession extends AbstractSessionOperations implements C
|
|||
}
|
||||
|
||||
public <T> InsertOperation<T> insert(T pojo) {
|
||||
Objects.requireNonNull(pojo, "supplied object must be either an instance of the entity class or a dsl for it, but cannot be null");
|
||||
Objects.requireNonNull(
|
||||
pojo,
|
||||
"supplied object must be either an instance of the entity class or a dsl for it, but cannot be null");
|
||||
HelenusEntity entity = null;
|
||||
try { entity = Helenus.resolve(pojo); } catch (HelenusMappingException e) {}
|
||||
try {
|
||||
entity = Helenus.resolve(pojo);
|
||||
} catch (HelenusMappingException e) {
|
||||
}
|
||||
if (entity != null) {
|
||||
return new InsertOperation<T>(this, entity.getMappingInterface(), true);
|
||||
} else {
|
||||
|
@ -488,7 +513,9 @@ public final class HelenusSession extends AbstractSessionOperations implements C
|
|||
}
|
||||
}
|
||||
|
||||
public <T> InsertOperation<T> insert(Drafted draft) { return insert(draft.build(), draft.mutated()); }
|
||||
public <T> InsertOperation<T> insert(Drafted draft) {
|
||||
return insert(draft.build(), draft.mutated());
|
||||
}
|
||||
|
||||
private <T> InsertOperation<T> insert(T pojo, Set<String> mutations) {
|
||||
Objects.requireNonNull(pojo, "pojo is empty");
|
||||
|
@ -512,9 +539,14 @@ public final class HelenusSession extends AbstractSessionOperations implements C
|
|||
}
|
||||
|
||||
public <T> InsertOperation<T> upsert(T pojo) {
|
||||
Objects.requireNonNull(pojo, "supplied object must be either an instance of the entity class or a dsl for it, but cannot be null");
|
||||
Objects.requireNonNull(
|
||||
pojo,
|
||||
"supplied object must be either an instance of the entity class or a dsl for it, but cannot be null");
|
||||
HelenusEntity entity = null;
|
||||
try { entity = Helenus.resolve(pojo); } catch (HelenusMappingException e) {}
|
||||
try {
|
||||
entity = Helenus.resolve(pojo);
|
||||
} catch (HelenusMappingException e) {
|
||||
}
|
||||
if (entity != null) {
|
||||
return new InsertOperation<T>(this, entity.getMappingInterface(), false);
|
||||
} else {
|
||||
|
@ -587,5 +619,4 @@ public final class HelenusSession extends AbstractSessionOperations implements C
|
|||
break;
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -1,30 +1,29 @@
|
|||
package net.helenus.core;
|
||||
|
||||
|
||||
import java.util.Objects;
|
||||
import java.util.*;
|
||||
import java.util.Objects;
|
||||
|
||||
public class PostCommitFunction<T, R> implements java.util.function.Function<T, R> {
|
||||
|
||||
private final UnitOfWork uow;
|
||||
private final List<CommitThunk> postCommit;
|
||||
private final UnitOfWork uow;
|
||||
private final List<CommitThunk> postCommit;
|
||||
|
||||
PostCommitFunction(UnitOfWork uow, List<CommitThunk> postCommit) {
|
||||
this.uow = uow;
|
||||
this.postCommit = postCommit;
|
||||
}
|
||||
PostCommitFunction(UnitOfWork uow, List<CommitThunk> postCommit) {
|
||||
this.uow = uow;
|
||||
this.postCommit = postCommit;
|
||||
}
|
||||
|
||||
public void andThen(CommitThunk after) {
|
||||
Objects.requireNonNull(after);
|
||||
if (postCommit == null) {
|
||||
after.apply();
|
||||
} else {
|
||||
postCommit.add(after);
|
||||
}
|
||||
public void andThen(CommitThunk after) {
|
||||
Objects.requireNonNull(after);
|
||||
if (postCommit == null) {
|
||||
after.apply();
|
||||
} else {
|
||||
postCommit.add(after);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public R apply(T t) {
|
||||
return null;
|
||||
}
|
||||
@Override
|
||||
public R apply(T t) {
|
||||
return null;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -17,7 +17,6 @@ package net.helenus.core;
|
|||
|
||||
import com.datastax.driver.core.*;
|
||||
import com.datastax.driver.core.IndexMetadata;
|
||||
import com.datastax.driver.core.querybuilder.Clause;
|
||||
import com.datastax.driver.core.querybuilder.IsNotNullClause;
|
||||
import com.datastax.driver.core.querybuilder.QueryBuilder;
|
||||
import com.datastax.driver.core.querybuilder.Select;
|
||||
|
@ -25,8 +24,6 @@ import com.datastax.driver.core.schemabuilder.*;
|
|||
import com.datastax.driver.core.schemabuilder.Create.Options;
|
||||
import java.util.*;
|
||||
import java.util.stream.Collectors;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
import net.helenus.core.reflect.HelenusPropertyNode;
|
||||
import net.helenus.mapping.*;
|
||||
import net.helenus.mapping.ColumnType;
|
||||
|
@ -34,7 +31,6 @@ import net.helenus.mapping.type.OptionalColumnMetadata;
|
|||
import net.helenus.support.CqlUtil;
|
||||
import net.helenus.support.HelenusMappingException;
|
||||
|
||||
import static com.datastax.driver.core.querybuilder.QueryBuilder.raw;
|
||||
|
||||
public final class SchemaUtil {
|
||||
|
||||
|
@ -152,7 +148,8 @@ public final class SchemaUtil {
|
|||
return SchemaBuilder.dropType(type.getTypeName()).ifExists();
|
||||
}
|
||||
|
||||
public static SchemaStatement createMaterializedView(String keyspace, String viewName, HelenusEntity entity) {
|
||||
public static SchemaStatement createMaterializedView(
|
||||
String keyspace, String viewName, HelenusEntity entity) {
|
||||
if (entity.getType() != HelenusEntityType.VIEW) {
|
||||
throw new HelenusMappingException("expected view entity " + entity);
|
||||
}
|
||||
|
@ -163,49 +160,53 @@ public final class SchemaUtil {
|
|||
|
||||
List<HelenusPropertyNode> props = new ArrayList<HelenusPropertyNode>();
|
||||
entity
|
||||
.getOrderedProperties()
|
||||
.stream()
|
||||
.map(p -> new HelenusPropertyNode(p, Optional.empty()))
|
||||
.forEach(p -> props.add(p));
|
||||
.getOrderedProperties()
|
||||
.stream()
|
||||
.map(p -> new HelenusPropertyNode(p, Optional.empty()))
|
||||
.forEach(p -> props.add(p));
|
||||
|
||||
Select.Selection selection = QueryBuilder.select();
|
||||
|
||||
for (HelenusPropertyNode prop : props) {
|
||||
String columnName = prop.getColumnName();
|
||||
selection = selection.column(columnName);
|
||||
String columnName = prop.getColumnName();
|
||||
selection = selection.column(columnName);
|
||||
}
|
||||
String tableName = Helenus.entity(entity.getMappingInterface().getInterfaces()[0])
|
||||
.getName().toCql();
|
||||
String tableName =
|
||||
Helenus.entity(entity.getMappingInterface().getInterfaces()[0]).getName().toCql();
|
||||
Select.Where where = selection.from(tableName).where();
|
||||
List<String> p = new ArrayList<String>(props.size());
|
||||
List<String> c = new ArrayList<String>(props.size());
|
||||
|
||||
for (HelenusPropertyNode prop : props) {
|
||||
String columnName = prop.getColumnName();
|
||||
switch(prop.getProperty().getColumnType()) {
|
||||
String columnName = prop.getColumnName();
|
||||
switch (prop.getProperty().getColumnType()) {
|
||||
case PARTITION_KEY:
|
||||
p.add(columnName);
|
||||
where = where.and(new IsNotNullClause(columnName));
|
||||
break;
|
||||
p.add(columnName);
|
||||
where = where.and(new IsNotNullClause(columnName));
|
||||
break;
|
||||
case CLUSTERING_COLUMN:
|
||||
c.add(columnName);
|
||||
where = where.and(new IsNotNullClause(columnName));
|
||||
break;
|
||||
c.add(columnName);
|
||||
where = where.and(new IsNotNullClause(columnName));
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
String primaryKey = "PRIMARY KEY ("
|
||||
String primaryKey =
|
||||
"PRIMARY KEY ("
|
||||
+ ((p.size() > 1) ? "(" + String.join(", ", p) + ")" : p.get(0))
|
||||
+ ((c.size() > 0) ? ", " + ((c.size() > 1) ? "(" + String.join(", ", c) + ")" : c.get(0)) : "")
|
||||
+ ((c.size() > 0)
|
||||
? ", " + ((c.size() > 1) ? "(" + String.join(", ", c) + ")" : c.get(0))
|
||||
: "")
|
||||
+ ")";
|
||||
|
||||
return new CreateMaterializedView(keyspace, viewName, where, primaryKey);
|
||||
}
|
||||
|
||||
public static SchemaStatement dropMaterializedView(String keyspace, String viewName, HelenusEntity entity) {
|
||||
return new DropMaterializedView(keyspace, viewName);
|
||||
public static SchemaStatement dropMaterializedView(
|
||||
String keyspace, String viewName, HelenusEntity entity) {
|
||||
return new DropMaterializedView(keyspace, viewName);
|
||||
}
|
||||
|
||||
public static SchemaStatement createTable(HelenusEntity entity) {
|
||||
|
|
|
@ -19,6 +19,12 @@ import brave.Tracer;
|
|||
import com.codahale.metrics.MetricRegistry;
|
||||
import com.datastax.driver.core.*;
|
||||
import com.google.common.util.concurrent.MoreExecutors;
|
||||
import java.io.IOException;
|
||||
import java.io.PrintStream;
|
||||
import java.util.*;
|
||||
import java.util.concurrent.Executor;
|
||||
import java.util.concurrent.Executors;
|
||||
import java.util.function.Consumer;
|
||||
import net.helenus.core.reflect.DslExportable;
|
||||
import net.helenus.mapping.HelenusEntity;
|
||||
import net.helenus.mapping.HelenusEntityType;
|
||||
|
@ -29,13 +35,6 @@ import net.helenus.support.Either;
|
|||
import net.helenus.support.HelenusException;
|
||||
import net.helenus.support.PackageUtil;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.PrintStream;
|
||||
import java.util.*;
|
||||
import java.util.concurrent.Executor;
|
||||
import java.util.concurrent.Executors;
|
||||
import java.util.function.Consumer;
|
||||
|
||||
public final class SessionInitializer extends AbstractSessionOperations {
|
||||
|
||||
private final Session session;
|
||||
|
@ -131,12 +130,12 @@ public final class SessionInitializer extends AbstractSessionOperations {
|
|||
}
|
||||
|
||||
public SessionInitializer idempotentQueryExecution(boolean idempotent) {
|
||||
this.idempotent = idempotent;
|
||||
return this;
|
||||
this.idempotent = idempotent;
|
||||
return this;
|
||||
}
|
||||
|
||||
public boolean getDefaultQueryIdempotency() {
|
||||
return idempotent;
|
||||
return idempotent;
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -185,9 +184,10 @@ public final class SessionInitializer extends AbstractSessionOperations {
|
|||
PackageUtil.getClasses(packageName)
|
||||
.stream()
|
||||
.filter(c -> c.isInterface() && !c.isAnnotation())
|
||||
.forEach(clazz -> {
|
||||
initList.add(Either.right(clazz));
|
||||
});
|
||||
.forEach(
|
||||
clazz -> {
|
||||
initList.add(Either.right(clazz));
|
||||
});
|
||||
} catch (IOException | ClassNotFoundException e) {
|
||||
throw new HelenusException("fail to add package " + packageName, e);
|
||||
}
|
||||
|
@ -267,18 +267,19 @@ public final class SessionInitializer extends AbstractSessionOperations {
|
|||
|
||||
Objects.requireNonNull(usingKeyspace, "please define keyspace by 'use' operator");
|
||||
|
||||
initList.forEach((either) -> {
|
||||
Class<?> iface = null;
|
||||
if (either.isLeft()) {
|
||||
initList.forEach(
|
||||
(either) -> {
|
||||
Class<?> iface = null;
|
||||
if (either.isLeft()) {
|
||||
iface = MappingUtil.getMappingInterface(either.getLeft());
|
||||
} else {
|
||||
} else {
|
||||
iface = either.getRight();
|
||||
}
|
||||
}
|
||||
|
||||
DslExportable dsl = (DslExportable) Helenus.dsl(iface);
|
||||
dsl.setCassandraMetadataForHelenusSesion(session.getCluster().getMetadata());
|
||||
sessionRepository.add(dsl);
|
||||
});
|
||||
DslExportable dsl = (DslExportable) Helenus.dsl(iface);
|
||||
dsl.setCassandraMetadataForHelenusSesion(session.getCluster().getMetadata());
|
||||
sessionRepository.add(dsl);
|
||||
});
|
||||
|
||||
TableOperations tableOps = new TableOperations(this, dropUnusedColumns, dropUnusedIndexes);
|
||||
UserTypeOperations userTypeOps = new UserTypeOperations(this, dropUnusedColumns);
|
||||
|
@ -289,18 +290,18 @@ public final class SessionInitializer extends AbstractSessionOperations {
|
|||
// Drop view first, otherwise a `DROP TABLE ...` will fail as the type is still referenced
|
||||
// by a view.
|
||||
sessionRepository
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.VIEW)
|
||||
.forEach(e -> tableOps.dropView(e));
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.VIEW)
|
||||
.forEach(e -> tableOps.dropView(e));
|
||||
|
||||
// Drop tables second, before DROP TYPE otherwise a `DROP TYPE ...` will fail as the type is
|
||||
// still referenced by a table.
|
||||
sessionRepository
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.TABLE)
|
||||
.forEach(e -> tableOps.dropTable(e));
|
||||
sessionRepository
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.TABLE)
|
||||
.forEach(e -> tableOps.dropTable(e));
|
||||
|
||||
eachUserTypeInReverseOrder(userTypeOps, e -> userTypeOps.dropUserType(e));
|
||||
|
||||
|
@ -314,11 +315,11 @@ public final class SessionInitializer extends AbstractSessionOperations {
|
|||
.filter(e -> e.getType() == HelenusEntityType.TABLE)
|
||||
.forEach(e -> tableOps.createTable(e));
|
||||
|
||||
sessionRepository
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.VIEW)
|
||||
.forEach(e -> tableOps.createView(e));
|
||||
sessionRepository
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.VIEW)
|
||||
.forEach(e -> tableOps.createView(e));
|
||||
|
||||
break;
|
||||
|
||||
|
@ -331,29 +332,28 @@ public final class SessionInitializer extends AbstractSessionOperations {
|
|||
.filter(e -> e.getType() == HelenusEntityType.TABLE)
|
||||
.forEach(e -> tableOps.validateTable(getTableMetadata(e), e));
|
||||
|
||||
break;
|
||||
break;
|
||||
|
||||
case UPDATE:
|
||||
eachUserTypeInOrder(userTypeOps, e -> userTypeOps.updateUserType(getUserType(e), e));
|
||||
|
||||
sessionRepository
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.VIEW)
|
||||
.forEach(e -> tableOps.dropView(e));
|
||||
sessionRepository
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.VIEW)
|
||||
.forEach(e -> tableOps.dropView(e));
|
||||
|
||||
|
||||
sessionRepository
|
||||
sessionRepository
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.TABLE)
|
||||
.forEach(e -> tableOps.updateTable(getTableMetadata(e), e));
|
||||
|
||||
sessionRepository
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.VIEW)
|
||||
.forEach(e -> tableOps.createView(e));
|
||||
sessionRepository
|
||||
.entities()
|
||||
.stream()
|
||||
.filter(e -> e.getType() == HelenusEntityType.VIEW)
|
||||
.forEach(e -> tableOps.createView(e));
|
||||
break;
|
||||
}
|
||||
|
||||
|
|
|
@ -47,7 +47,10 @@ public final class TableOperations {
|
|||
|
||||
if (tmd == null) {
|
||||
throw new HelenusException(
|
||||
"table does not exists " + entity.getName() + "for entity " + entity.getMappingInterface());
|
||||
"table does not exists "
|
||||
+ entity.getName()
|
||||
+ "for entity "
|
||||
+ entity.getMappingInterface());
|
||||
}
|
||||
|
||||
List<SchemaStatement> list = SchemaUtil.alterTable(tmd, entity, dropUnusedColumns);
|
||||
|
@ -66,7 +69,7 @@ public final class TableOperations {
|
|||
public void updateTable(TableMetadata tmd, HelenusEntity entity) {
|
||||
if (tmd == null) {
|
||||
createTable(entity);
|
||||
return;
|
||||
return;
|
||||
}
|
||||
|
||||
executeBatch(SchemaUtil.alterTable(tmd, entity, dropUnusedColumns));
|
||||
|
@ -74,26 +77,31 @@ public final class TableOperations {
|
|||
}
|
||||
|
||||
public void createView(HelenusEntity entity) {
|
||||
sessionOps.execute(SchemaUtil.createMaterializedView(sessionOps.usingKeyspace(), entity.getName().toCql(), entity), true);
|
||||
// executeBatch(SchemaUtil.createIndexes(entity)); NOTE: Unfortunately C* 3.10 does not yet support 2i on materialized views.
|
||||
sessionOps.execute(
|
||||
SchemaUtil.createMaterializedView(
|
||||
sessionOps.usingKeyspace(), entity.getName().toCql(), entity),
|
||||
true);
|
||||
// executeBatch(SchemaUtil.createIndexes(entity)); NOTE: Unfortunately C* 3.10 does not yet support 2i on materialized views.
|
||||
}
|
||||
|
||||
public void dropView(HelenusEntity entity) {
|
||||
sessionOps.execute(SchemaUtil.dropMaterializedView(sessionOps.usingKeyspace(), entity.getName().toCql(), entity), true);
|
||||
sessionOps.execute(
|
||||
SchemaUtil.dropMaterializedView(
|
||||
sessionOps.usingKeyspace(), entity.getName().toCql(), entity),
|
||||
true);
|
||||
}
|
||||
|
||||
public void updateView(TableMetadata tmd, HelenusEntity entity) {
|
||||
if (tmd == null) {
|
||||
createTable(entity);
|
||||
return;
|
||||
}
|
||||
|
||||
executeBatch(SchemaUtil.alterTable(tmd, entity, dropUnusedColumns));
|
||||
executeBatch(SchemaUtil.alterIndexes(tmd, entity, dropUnusedIndexes));
|
||||
if (tmd == null) {
|
||||
createTable(entity);
|
||||
return;
|
||||
}
|
||||
|
||||
executeBatch(SchemaUtil.alterTable(tmd, entity, dropUnusedColumns));
|
||||
executeBatch(SchemaUtil.alterIndexes(tmd, entity, dropUnusedIndexes));
|
||||
}
|
||||
|
||||
private void executeBatch(List<SchemaStatement> list) {
|
||||
private void executeBatch(List<SchemaStatement> list) {
|
||||
|
||||
list.forEach(
|
||||
s -> {
|
||||
|
|
|
@ -15,44 +15,42 @@
|
|||
*/
|
||||
package net.helenus.core;
|
||||
|
||||
import net.helenus.support.Either;
|
||||
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
|
||||
public interface UnitOfWork<E extends Exception> extends AutoCloseable {
|
||||
|
||||
/**
|
||||
* Marks the beginning of a transactional section of work. Will write a record to the shared
|
||||
* write-ahead log.
|
||||
*
|
||||
* @return the handle used to commit or abort the work.
|
||||
*/
|
||||
UnitOfWork begin();
|
||||
/**
|
||||
* Marks the beginning of a transactional section of work. Will write a record to the shared
|
||||
* write-ahead log.
|
||||
*
|
||||
* @return the handle used to commit or abort the work.
|
||||
*/
|
||||
UnitOfWork begin();
|
||||
|
||||
UnitOfWork addNestedUnitOfWork(UnitOfWork uow);
|
||||
UnitOfWork addNestedUnitOfWork(UnitOfWork uow);
|
||||
|
||||
/**
|
||||
* Checks to see if the work performed between calling begin and now can be committed or not.
|
||||
*
|
||||
* @return a function from which to chain work that only happens when commit is successful
|
||||
* @throws E when the work overlaps with other concurrent writers.
|
||||
*/
|
||||
PostCommitFunction<Void, Void> commit() throws E;
|
||||
/**
|
||||
* Checks to see if the work performed between calling begin and now can be committed or not.
|
||||
*
|
||||
* @return a function from which to chain work that only happens when commit is successful
|
||||
* @throws E when the work overlaps with other concurrent writers.
|
||||
*/
|
||||
PostCommitFunction<Void, Void> commit() throws E;
|
||||
|
||||
/**
|
||||
* Explicitly abort the work within this unit of work. Any nested aborted unit of work
|
||||
* will trigger the entire unit of work to commit.
|
||||
*/
|
||||
void abort();
|
||||
/**
|
||||
* Explicitly abort the work within this unit of work. Any nested aborted unit of work will
|
||||
* trigger the entire unit of work to commit.
|
||||
*/
|
||||
void abort();
|
||||
|
||||
boolean hasAborted();
|
||||
|
||||
boolean hasAborted();
|
||||
boolean hasCommitted();
|
||||
|
||||
boolean hasCommitted();
|
||||
//Either<Object, Set<Object>> cacheLookup(String key);
|
||||
Set<Object> cacheLookup(String key);
|
||||
|
||||
//Either<Object, Set<Object>> cacheLookup(String key);
|
||||
Set<Object> cacheLookup(String key);
|
||||
|
||||
Map<String, Set<Object>> getCache();
|
||||
Map<String, Set<Object>> getCache();
|
||||
}
|
||||
|
|
|
@ -19,9 +19,8 @@ import net.helenus.support.HelenusException;
|
|||
|
||||
class UnitOfWorkImpl extends AbstractUnitOfWork<HelenusException> {
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
public UnitOfWorkImpl(HelenusSession session, UnitOfWork parent) {
|
||||
super(session, (AbstractUnitOfWork<HelenusException>) parent);
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
public UnitOfWorkImpl(HelenusSession session, UnitOfWork parent) {
|
||||
super(session, (AbstractUnitOfWork<HelenusException>) parent);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,11 +19,11 @@ import java.util.LinkedHashMap;
|
|||
import java.util.LinkedList;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
|
||||
import net.helenus.core.*;
|
||||
import net.helenus.mapping.HelenusProperty;
|
||||
|
||||
public abstract class AbstractFilterOptionalOperation<E, O extends AbstractFilterOptionalOperation<E, O>>
|
||||
public abstract class AbstractFilterOptionalOperation<
|
||||
E, O extends AbstractFilterOptionalOperation<E, O>>
|
||||
extends AbstractOptionalOperation<E, O> {
|
||||
|
||||
protected Map<HelenusProperty, Filter<?>> filters = null;
|
||||
|
|
|
@ -19,11 +19,11 @@ import java.util.LinkedHashMap;
|
|||
import java.util.LinkedList;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
|
||||
import net.helenus.core.*;
|
||||
import net.helenus.mapping.HelenusProperty;
|
||||
|
||||
public abstract class AbstractFilterStreamOperation<E, O extends AbstractFilterStreamOperation<E, O>>
|
||||
public abstract class AbstractFilterStreamOperation<
|
||||
E, O extends AbstractFilterStreamOperation<E, O>>
|
||||
extends AbstractStreamOperation<E, O> {
|
||||
|
||||
protected Map<HelenusProperty, Filter<?>> filters = null;
|
||||
|
|
|
@ -17,8 +17,6 @@ package net.helenus.core.operation;
|
|||
|
||||
import com.codahale.metrics.Timer;
|
||||
import com.datastax.driver.core.ResultSet;
|
||||
|
||||
import java.util.Objects;
|
||||
import java.util.concurrent.CompletableFuture;
|
||||
import net.helenus.core.AbstractSessionOperations;
|
||||
import net.helenus.core.UnitOfWork;
|
||||
|
@ -40,7 +38,6 @@ public abstract class AbstractOperation<E, O extends AbstractOperation<E, O>>
|
|||
return new PreparedOperation<E>(prepareStatement(), this);
|
||||
}
|
||||
|
||||
|
||||
public E sync() {
|
||||
final Timer.Context context = requestLatency.time();
|
||||
try {
|
||||
|
@ -72,5 +69,4 @@ public abstract class AbstractOperation<E, O extends AbstractOperation<E, O>>
|
|||
if (uow == null) return async();
|
||||
return CompletableFuture.<E>supplyAsync(() -> sync(uow));
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -21,12 +21,10 @@ import com.datastax.driver.core.ResultSet;
|
|||
import com.google.common.base.Function;
|
||||
import com.google.common.util.concurrent.Futures;
|
||||
import com.google.common.util.concurrent.ListenableFuture;
|
||||
|
||||
import java.util.HashSet;
|
||||
import java.util.Optional;
|
||||
import java.util.Set;
|
||||
import java.util.concurrent.CompletableFuture;
|
||||
|
||||
import net.helenus.core.AbstractSessionOperations;
|
||||
import net.helenus.core.UnitOfWork;
|
||||
|
||||
|
@ -114,5 +112,4 @@ public abstract class AbstractOptionalOperation<E, O extends AbstractOptionalOpe
|
|||
if (uow == null) return async();
|
||||
return CompletableFuture.<Optional<E>>supplyAsync(() -> sync(uow));
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -32,7 +32,8 @@ import net.helenus.support.HelenusException;
|
|||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
public abstract class AbstractStatementOperation<E, O extends AbstractStatementOperation<E, O>> extends Operation<E> {
|
||||
public abstract class AbstractStatementOperation<E, O extends AbstractStatementOperation<E, O>>
|
||||
extends Operation<E> {
|
||||
|
||||
final Logger logger = LoggerFactory.getLogger(getClass());
|
||||
|
||||
|
@ -55,7 +56,6 @@ public abstract class AbstractStatementOperation<E, O extends AbstractStatementO
|
|||
this.idempotent = sessionOperations.getDefaultQueryIdempotency();
|
||||
}
|
||||
|
||||
|
||||
public O ignoreCache(boolean enabled) {
|
||||
enableCache = enabled;
|
||||
return (O) this;
|
||||
|
@ -93,8 +93,8 @@ public abstract class AbstractStatementOperation<E, O extends AbstractStatementO
|
|||
}
|
||||
|
||||
public O isIdempotent(boolean idempotent) {
|
||||
this.idempotent = idempotent;
|
||||
return (O) this;
|
||||
this.idempotent = idempotent;
|
||||
return (O) this;
|
||||
}
|
||||
|
||||
public O downgradingConsistencyRetryPolicy() {
|
||||
|
@ -232,7 +232,7 @@ public abstract class AbstractStatementOperation<E, O extends AbstractStatementO
|
|||
}
|
||||
|
||||
if (idempotent) {
|
||||
statement.setIdempotent(true);
|
||||
statement.setIdempotent(true);
|
||||
}
|
||||
|
||||
return statement;
|
||||
|
|
|
@ -21,20 +21,11 @@ import com.datastax.driver.core.ResultSet;
|
|||
import com.google.common.base.Function;
|
||||
import com.google.common.util.concurrent.Futures;
|
||||
import com.google.common.util.concurrent.ListenableFuture;
|
||||
|
||||
import java.util.Map;
|
||||
import java.util.Objects;
|
||||
import java.util.Optional;
|
||||
import java.util.Set;
|
||||
import java.util.concurrent.CompletableFuture;
|
||||
import java.util.stream.Collectors;
|
||||
import java.util.stream.Stream;
|
||||
import net.helenus.core.AbstractSessionOperations;
|
||||
import net.helenus.core.Helenus;
|
||||
import net.helenus.core.UnitOfWork;
|
||||
import net.helenus.mapping.HelenusEntity;
|
||||
import net.helenus.mapping.value.ColumnValueProvider;
|
||||
import net.helenus.mapping.value.ValueProviderMap;
|
||||
|
||||
public abstract class AbstractStreamOperation<E, O extends AbstractStreamOperation<E, O>>
|
||||
extends AbstractStatementOperation<E, O> {
|
||||
|
@ -113,5 +104,4 @@ public abstract class AbstractStreamOperation<E, O extends AbstractStreamOperati
|
|||
if (uow == null) return async();
|
||||
return CompletableFuture.<Stream<E>>supplyAsync(() -> sync(uow));
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -34,7 +34,9 @@ public final class BoundOptionalOperation<E>
|
|||
}
|
||||
|
||||
@Override
|
||||
public Optional<E> transform(ResultSet resultSet) { return delegate.transform(resultSet); }
|
||||
public Optional<E> transform(ResultSet resultSet) {
|
||||
return delegate.transform(resultSet);
|
||||
}
|
||||
|
||||
@Override
|
||||
public Statement buildStatement(boolean cached) {
|
||||
|
|
|
@ -26,14 +26,17 @@ public final class BoundStreamOperation<E>
|
|||
private final BoundStatement boundStatement;
|
||||
private final AbstractStreamOperation<E, ?> delegate;
|
||||
|
||||
public BoundStreamOperation(BoundStatement boundStatement, AbstractStreamOperation<E, ?> operation) {
|
||||
public BoundStreamOperation(
|
||||
BoundStatement boundStatement, AbstractStreamOperation<E, ?> operation) {
|
||||
super(operation.sessionOps);
|
||||
this.boundStatement = boundStatement;
|
||||
this.delegate = operation;
|
||||
}
|
||||
|
||||
@Override
|
||||
public String getStatementCacheKey() { return delegate.getStatementCacheKey(); }
|
||||
public String getStatementCacheKey() {
|
||||
return delegate.getStatementCacheKey();
|
||||
}
|
||||
|
||||
@Override
|
||||
public Stream<E> transform(ResultSet resultSet) {
|
||||
|
@ -41,5 +44,7 @@ public final class BoundStreamOperation<E>
|
|||
}
|
||||
|
||||
@Override
|
||||
public Statement buildStatement(boolean cached) { return boundStatement; }
|
||||
public Statement buildStatement(boolean cached) {
|
||||
return boundStatement;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,10 +19,9 @@ import com.datastax.driver.core.ResultSet;
|
|||
import com.datastax.driver.core.querybuilder.BuiltStatement;
|
||||
import com.datastax.driver.core.querybuilder.Insert;
|
||||
import com.datastax.driver.core.querybuilder.QueryBuilder;
|
||||
import com.google.common.base.Joiner;
|
||||
import java.util.*;
|
||||
import java.util.function.Function;
|
||||
|
||||
import com.google.common.base.Joiner;
|
||||
import net.helenus.core.AbstractSessionOperations;
|
||||
import net.helenus.core.Getter;
|
||||
import net.helenus.core.Helenus;
|
||||
|
@ -41,7 +40,8 @@ public final class InsertOperation<T> extends AbstractOperation<T, InsertOperati
|
|||
|
||||
private HelenusEntity entity;
|
||||
|
||||
private final List<Fun.Tuple2<HelenusPropertyNode, Object>> values = new ArrayList<Fun.Tuple2<HelenusPropertyNode, Object>>();
|
||||
private final List<Fun.Tuple2<HelenusPropertyNode, Object>> values =
|
||||
new ArrayList<Fun.Tuple2<HelenusPropertyNode, Object>>();
|
||||
private final T pojo;
|
||||
private final Class<?> resultType;
|
||||
private boolean ifNotExists;
|
||||
|
@ -57,7 +57,8 @@ public final class InsertOperation<T> extends AbstractOperation<T, InsertOperati
|
|||
this.resultType = ResultSet.class;
|
||||
}
|
||||
|
||||
public InsertOperation(AbstractSessionOperations sessionOperations, Class<?> resultType, boolean ifNotExists) {
|
||||
public InsertOperation(
|
||||
AbstractSessionOperations sessionOperations, Class<?> resultType, boolean ifNotExists) {
|
||||
super(sessionOperations);
|
||||
|
||||
this.ifNotExists = ifNotExists;
|
||||
|
@ -172,7 +173,7 @@ public final class InsertOperation<T> extends AbstractOperation<T, InsertOperati
|
|||
// Some values man need to be converted (e.g. from String to Enum). This is done
|
||||
// within the BeanColumnValueProvider below.
|
||||
Optional<Function<Object, Object>> converter =
|
||||
prop.getReadConverter(sessionOps.getSessionRepository());
|
||||
prop.getReadConverter(sessionOps.getSessionRepository());
|
||||
if (converter.isPresent()) {
|
||||
backingMap.put(key, converter.get().apply(backingMap.get(key)));
|
||||
}
|
||||
|
@ -232,23 +233,25 @@ public final class InsertOperation<T> extends AbstractOperation<T, InsertOperati
|
|||
public String getStatementCacheKey() {
|
||||
List<String> keys = new ArrayList<>(values.size());
|
||||
values.forEach(
|
||||
t -> {
|
||||
HelenusPropertyNode prop = t._1;
|
||||
switch (prop.getProperty().getColumnType()) {
|
||||
case PARTITION_KEY:
|
||||
case CLUSTERING_COLUMN:
|
||||
keys.add(prop.getColumnName() + "==" + t._2.toString());
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
});
|
||||
t -> {
|
||||
HelenusPropertyNode prop = t._1;
|
||||
switch (prop.getProperty().getColumnType()) {
|
||||
case PARTITION_KEY:
|
||||
case CLUSTERING_COLUMN:
|
||||
keys.add(prop.getColumnName() + "==" + t._2.toString());
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
});
|
||||
return entity.getName() + ": " + Joiner.on(",").join(keys);
|
||||
}
|
||||
|
||||
@Override
|
||||
public T sync(UnitOfWork uow) {
|
||||
if (uow == null) { return sync(); }
|
||||
if (uow == null) {
|
||||
return sync();
|
||||
}
|
||||
T result = super.sync(uow);
|
||||
Class<?> iface = entity.getMappingInterface();
|
||||
if (resultType == iface) {
|
||||
|
@ -261,5 +264,4 @@ public final class InsertOperation<T> extends AbstractOperation<T, InsertOperati
|
|||
}
|
||||
return result;
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -1,73 +1,79 @@
|
|||
package net.helenus.core.operation;
|
||||
|
||||
import java.util.concurrent.ExecutionException;
|
||||
|
||||
import brave.Span;
|
||||
import brave.Tracer;
|
||||
import brave.propagation.TraceContext;
|
||||
import com.codahale.metrics.Meter;
|
||||
import com.codahale.metrics.MetricRegistry;
|
||||
import com.codahale.metrics.Timer;
|
||||
import com.datastax.driver.core.ResultSet;
|
||||
import com.datastax.driver.core.ResultSetFuture;
|
||||
import com.datastax.driver.core.Statement;
|
||||
|
||||
import brave.Span;
|
||||
import brave.Tracer;
|
||||
import brave.propagation.TraceContext;
|
||||
import java.util.concurrent.ExecutionException;
|
||||
import net.helenus.core.AbstractSessionOperations;
|
||||
import net.helenus.core.UnitOfWork;
|
||||
|
||||
|
||||
public abstract class Operation<E> {
|
||||
|
||||
protected final AbstractSessionOperations sessionOps;
|
||||
protected final Meter uowCacheHits;
|
||||
protected final Meter uowCacheMiss;
|
||||
protected final Timer requestLatency;
|
||||
protected final AbstractSessionOperations sessionOps;
|
||||
protected final Meter uowCacheHits;
|
||||
protected final Meter uowCacheMiss;
|
||||
protected final Timer requestLatency;
|
||||
|
||||
Operation(AbstractSessionOperations sessionOperations) {
|
||||
this.sessionOps = sessionOperations;
|
||||
MetricRegistry metrics = sessionOperations.getMetricRegistry();
|
||||
this.uowCacheHits = metrics.meter("net.helenus.UOW-cache-hits");
|
||||
this.uowCacheMiss = metrics.meter("net.helenus.UOW-cache-miss");
|
||||
this.requestLatency = metrics.timer("net.helenus.request-latency");
|
||||
Operation(AbstractSessionOperations sessionOperations) {
|
||||
this.sessionOps = sessionOperations;
|
||||
MetricRegistry metrics = sessionOperations.getMetricRegistry();
|
||||
this.uowCacheHits = metrics.meter("net.helenus.UOW-cache-hits");
|
||||
this.uowCacheMiss = metrics.meter("net.helenus.UOW-cache-miss");
|
||||
this.requestLatency = metrics.timer("net.helenus.request-latency");
|
||||
}
|
||||
|
||||
public ResultSet execute(
|
||||
AbstractSessionOperations session,
|
||||
UnitOfWork uow,
|
||||
TraceContext traceContext,
|
||||
boolean showValues,
|
||||
boolean cached) {
|
||||
|
||||
// Start recording in a Zipkin sub-span our execution time to perform this operation.
|
||||
Tracer tracer = session.getZipkinTracer();
|
||||
Span span = null;
|
||||
if (tracer != null && traceContext != null) {
|
||||
span = tracer.newChild(traceContext);
|
||||
}
|
||||
|
||||
public ResultSet execute(AbstractSessionOperations session, UnitOfWork uow, TraceContext traceContext, boolean showValues, boolean cached) {
|
||||
try {
|
||||
|
||||
// Start recording in a Zipkin sub-span our execution time to perform this operation.
|
||||
Tracer tracer = session.getZipkinTracer();
|
||||
Span span = null;
|
||||
if (tracer != null && traceContext != null) {
|
||||
span = tracer.newChild(traceContext);
|
||||
}
|
||||
if (span != null) {
|
||||
span.name("cassandra");
|
||||
span.start();
|
||||
}
|
||||
|
||||
try {
|
||||
Statement statement = options(buildStatement(cached));
|
||||
ResultSetFuture futureResultSet = session.executeAsync(statement, showValues);
|
||||
return futureResultSet.get();
|
||||
|
||||
if (span != null) {
|
||||
span.name("cassandra");
|
||||
span.start();
|
||||
}
|
||||
} catch (InterruptedException | ExecutionException e) {
|
||||
|
||||
Statement statement = options(buildStatement(cached));
|
||||
ResultSetFuture futureResultSet = session.executeAsync(statement, showValues);
|
||||
return futureResultSet.get();
|
||||
throw new RuntimeException(e);
|
||||
|
||||
} catch (InterruptedException | ExecutionException e) {
|
||||
} finally {
|
||||
|
||||
throw new RuntimeException(e);
|
||||
|
||||
} finally {
|
||||
|
||||
if (span != null) {
|
||||
span.finish();
|
||||
}
|
||||
|
||||
}
|
||||
if (span != null) {
|
||||
span.finish();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
public Statement options(Statement statement) { return statement; }
|
||||
public Statement options(Statement statement) {
|
||||
return statement;
|
||||
}
|
||||
|
||||
public Statement buildStatement(boolean cached) { return null; }
|
||||
|
||||
public String getStatementCacheKey() { return null; }
|
||||
public Statement buildStatement(boolean cached) {
|
||||
return null;
|
||||
}
|
||||
|
||||
public String getStatementCacheKey() {
|
||||
return null;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -17,7 +17,6 @@ package net.helenus.core.operation;
|
|||
|
||||
import com.datastax.driver.core.BoundStatement;
|
||||
import com.datastax.driver.core.PreparedStatement;
|
||||
import java.util.regex.Pattern;
|
||||
|
||||
public final class PreparedStreamOperation<E> {
|
||||
|
||||
|
|
|
@ -38,7 +38,9 @@ public final class SelectFirstOperation<E>
|
|||
}
|
||||
|
||||
@Override
|
||||
public String getStatementCacheKey() { return delegate.getStatementCacheKey(); }
|
||||
public String getStatementCacheKey() {
|
||||
return delegate.getStatementCacheKey();
|
||||
}
|
||||
|
||||
@Override
|
||||
public BuiltStatement buildStatement(boolean cached) {
|
||||
|
|
|
@ -36,7 +36,9 @@ public final class SelectFirstTransformingOperation<R, E>
|
|||
}
|
||||
|
||||
@Override
|
||||
public String getStatementCacheKey() { return delegate.getStatementCacheKey(); }
|
||||
public String getStatementCacheKey() {
|
||||
return delegate.getStatementCacheKey();
|
||||
}
|
||||
|
||||
@Override
|
||||
public BuiltStatement buildStatement(boolean cached) {
|
||||
|
|
|
@ -23,13 +23,12 @@ import com.datastax.driver.core.querybuilder.QueryBuilder;
|
|||
import com.datastax.driver.core.querybuilder.Select;
|
||||
import com.datastax.driver.core.querybuilder.Select.Selection;
|
||||
import com.datastax.driver.core.querybuilder.Select.Where;
|
||||
import com.google.common.base.Joiner;
|
||||
import com.google.common.collect.Iterables;
|
||||
import java.util.*;
|
||||
import java.util.function.Function;
|
||||
import java.util.stream.Stream;
|
||||
import java.util.stream.StreamSupport;
|
||||
|
||||
import com.google.common.base.Joiner;
|
||||
import com.google.common.collect.Iterables;
|
||||
import net.helenus.core.*;
|
||||
import net.helenus.core.reflect.HelenusPropertyNode;
|
||||
import net.helenus.mapping.HelenusEntity;
|
||||
|
@ -55,7 +54,7 @@ public final class SelectOperation<E> extends AbstractFilterStreamOperation<E, S
|
|||
super(sessionOperations);
|
||||
|
||||
this.rowMapper =
|
||||
new Function<Row, E>() {
|
||||
new Function<Row, E>() {
|
||||
|
||||
@Override
|
||||
public E apply(Row source) {
|
||||
|
@ -203,7 +202,6 @@ public final class SelectOperation<E> extends AbstractFilterStreamOperation<E, S
|
|||
switch (prop.getProperty().getColumnType()) {
|
||||
case PARTITION_KEY:
|
||||
case CLUSTERING_COLUMN:
|
||||
|
||||
Filter filter = filters.get(prop.getProperty());
|
||||
if (filter != null) {
|
||||
keys.add(filter.toString());
|
||||
|
@ -239,14 +237,14 @@ public final class SelectOperation<E> extends AbstractFilterStreamOperation<E, S
|
|||
entity = prop.getEntity();
|
||||
} else if (entity != prop.getEntity()) {
|
||||
throw new HelenusMappingException(
|
||||
"you can select columns only from a single entity "
|
||||
+ entity.getMappingInterface()
|
||||
+ " or "
|
||||
+ prop.getEntity().getMappingInterface());
|
||||
"you can select columns only from a single entity "
|
||||
+ entity.getMappingInterface()
|
||||
+ " or "
|
||||
+ prop.getEntity().getMappingInterface());
|
||||
}
|
||||
|
||||
if (cached) {
|
||||
switch(prop.getProperty().getColumnType()) {
|
||||
switch (prop.getProperty().getColumnType()) {
|
||||
case PARTITION_KEY:
|
||||
case CLUSTERING_COLUMN:
|
||||
break;
|
||||
|
@ -305,10 +303,14 @@ public final class SelectOperation<E> extends AbstractFilterStreamOperation<E, S
|
|||
@Override
|
||||
public Stream<E> transform(ResultSet resultSet) {
|
||||
if (rowMapper != null) {
|
||||
return StreamSupport.stream(Spliterators.spliteratorUnknownSize(resultSet.iterator(), Spliterator.ORDERED), false).map(rowMapper);
|
||||
return StreamSupport.stream(
|
||||
Spliterators.spliteratorUnknownSize(resultSet.iterator(), Spliterator.ORDERED), false)
|
||||
.map(rowMapper);
|
||||
} else {
|
||||
return (Stream<E>)
|
||||
StreamSupport.stream(Spliterators.spliteratorUnknownSize(resultSet.iterator(), Spliterator.ORDERED),false);
|
||||
StreamSupport.stream(
|
||||
Spliterators.spliteratorUnknownSize(resultSet.iterator(), Spliterator.ORDERED),
|
||||
false);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -36,7 +36,9 @@ public final class SelectTransformingOperation<R, E>
|
|||
}
|
||||
|
||||
@Override
|
||||
public String getStatementCacheKey() { return delegate.getStatementCacheKey(); }
|
||||
public String getStatementCacheKey() {
|
||||
return delegate.getStatementCacheKey();
|
||||
}
|
||||
|
||||
@Override
|
||||
public BuiltStatement buildStatement(boolean cached) {
|
||||
|
@ -47,5 +49,4 @@ public final class SelectTransformingOperation<R, E>
|
|||
public Stream<R> transform(ResultSet resultSet) {
|
||||
return delegate.transform(resultSet).map(fn);
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -15,15 +15,13 @@
|
|||
*/
|
||||
package net.helenus.core.operation;
|
||||
|
||||
import java.util.*;
|
||||
import java.util.function.Function;
|
||||
|
||||
import com.datastax.driver.core.ResultSet;
|
||||
import com.datastax.driver.core.querybuilder.Assignment;
|
||||
import com.datastax.driver.core.querybuilder.BuiltStatement;
|
||||
import com.datastax.driver.core.querybuilder.QueryBuilder;
|
||||
import com.datastax.driver.core.querybuilder.Update;
|
||||
|
||||
import java.util.*;
|
||||
import java.util.function.Function;
|
||||
import net.helenus.core.*;
|
||||
import net.helenus.core.reflect.HelenusPropertyNode;
|
||||
import net.helenus.mapping.HelenusEntity;
|
||||
|
@ -43,19 +41,21 @@ public final class UpdateOperation<E> extends AbstractFilterOperation<E, UpdateO
|
|||
private int[] ttl;
|
||||
private long[] timestamp;
|
||||
|
||||
public UpdateOperation(AbstractSessionOperations sessionOperations){
|
||||
public UpdateOperation(AbstractSessionOperations sessionOperations) {
|
||||
super(sessionOperations);
|
||||
this.draft = null;
|
||||
this.draftMap = null;
|
||||
}
|
||||
|
||||
public UpdateOperation(AbstractSessionOperations sessionOperations, AbstractEntityDraft<E> draft) {
|
||||
public UpdateOperation(
|
||||
AbstractSessionOperations sessionOperations, AbstractEntityDraft<E> draft) {
|
||||
super(sessionOperations);
|
||||
this.draft = draft;
|
||||
this.draftMap = draft.toMap();
|
||||
}
|
||||
|
||||
public UpdateOperation(AbstractSessionOperations sessionOperations, HelenusPropertyNode p, Object v) {
|
||||
public UpdateOperation(
|
||||
AbstractSessionOperations sessionOperations, HelenusPropertyNode p, Object v) {
|
||||
super(sessionOperations);
|
||||
this.draft = null;
|
||||
this.draftMap = null;
|
||||
|
@ -152,7 +152,7 @@ public final class UpdateOperation<E> extends AbstractFilterOperation<E, UpdateO
|
|||
|
||||
if (draft != null) {
|
||||
String key = p.getProperty().getPropertyName();
|
||||
List<V> list = (List<V>)draftMap.get(key);
|
||||
List<V> list = (List<V>) draftMap.get(key);
|
||||
list.add(0, value);
|
||||
}
|
||||
|
||||
|
@ -194,7 +194,7 @@ public final class UpdateOperation<E> extends AbstractFilterOperation<E, UpdateO
|
|||
|
||||
if (draft != null) {
|
||||
String key = p.getProperty().getPropertyName();
|
||||
List<V> list = (List<V>)draftMap.get(key);
|
||||
List<V> list = (List<V>) draftMap.get(key);
|
||||
if (idx < 0) {
|
||||
list.add(0, value);
|
||||
} else if (idx > list.size()) {
|
||||
|
@ -222,7 +222,7 @@ public final class UpdateOperation<E> extends AbstractFilterOperation<E, UpdateO
|
|||
|
||||
if (draft != null) {
|
||||
String key = p.getProperty().getPropertyName();
|
||||
List<V> list = (List<V>)draftMap.get(key);
|
||||
List<V> list = (List<V>) draftMap.get(key);
|
||||
list.add(value);
|
||||
}
|
||||
|
||||
|
@ -579,7 +579,9 @@ public final class UpdateOperation<E> extends AbstractFilterOperation<E, UpdateO
|
|||
|
||||
@Override
|
||||
public E sync(UnitOfWork uow) {
|
||||
if (uow == null) { return sync(); }
|
||||
if (uow == null) {
|
||||
return sync();
|
||||
}
|
||||
E result = super.sync(uow);
|
||||
if (draft != null) {
|
||||
String key = getStatementCacheKey();
|
||||
|
@ -591,5 +593,4 @@ public final class UpdateOperation<E> extends AbstractFilterOperation<E, UpdateO
|
|||
}
|
||||
return result;
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -1,6 +1,5 @@
|
|||
package net.helenus.core.reflect;
|
||||
|
||||
import net.helenus.mapping.HelenusEntity;
|
||||
|
||||
import java.util.Set;
|
||||
|
||||
|
|
|
@ -67,46 +67,46 @@ public class DslInvocationHandler<E> implements InvocationHandler {
|
|||
}
|
||||
|
||||
private HelenusEntity init(Metadata metadata) {
|
||||
HelenusEntity entity = new HelenusMappingEntity(iface, metadata);
|
||||
HelenusEntity entity = new HelenusMappingEntity(iface, metadata);
|
||||
|
||||
for (HelenusProperty prop : entity.getOrderedProperties()) {
|
||||
for (HelenusProperty prop : entity.getOrderedProperties()) {
|
||||
|
||||
map.put(prop.getGetterMethod(), prop);
|
||||
map.put(prop.getGetterMethod(), prop);
|
||||
|
||||
AbstractDataType type = prop.getDataType();
|
||||
Class<?> javaType = prop.getJavaType();
|
||||
AbstractDataType type = prop.getDataType();
|
||||
Class<?> javaType = prop.getJavaType();
|
||||
|
||||
if (type instanceof UDTDataType && !UDTValue.class.isAssignableFrom(javaType)) {
|
||||
if (type instanceof UDTDataType && !UDTValue.class.isAssignableFrom(javaType)) {
|
||||
|
||||
Object childDsl =
|
||||
Helenus.dsl(
|
||||
javaType,
|
||||
classLoader,
|
||||
Optional.of(new HelenusPropertyNode(prop, parent)),
|
||||
metadata);
|
||||
Object childDsl =
|
||||
Helenus.dsl(
|
||||
javaType,
|
||||
classLoader,
|
||||
Optional.of(new HelenusPropertyNode(prop, parent)),
|
||||
metadata);
|
||||
|
||||
udtMap.put(prop.getGetterMethod(), childDsl);
|
||||
}
|
||||
|
||||
if (type instanceof DTDataType) {
|
||||
DTDataType dataType = (DTDataType) type;
|
||||
|
||||
if (dataType.getDataType() instanceof TupleType
|
||||
&& !TupleValue.class.isAssignableFrom(javaType)) {
|
||||
|
||||
Object childDsl =
|
||||
Helenus.dsl(
|
||||
javaType,
|
||||
classLoader,
|
||||
Optional.of(new HelenusPropertyNode(prop, parent)),
|
||||
metadata);
|
||||
|
||||
tupleMap.put(prop.getGetterMethod(), childDsl);
|
||||
}
|
||||
}
|
||||
udtMap.put(prop.getGetterMethod(), childDsl);
|
||||
}
|
||||
|
||||
return entity;
|
||||
if (type instanceof DTDataType) {
|
||||
DTDataType dataType = (DTDataType) type;
|
||||
|
||||
if (dataType.getDataType() instanceof TupleType
|
||||
&& !TupleValue.class.isAssignableFrom(javaType)) {
|
||||
|
||||
Object childDsl =
|
||||
Helenus.dsl(
|
||||
javaType,
|
||||
classLoader,
|
||||
Optional.of(new HelenusPropertyNode(prop, parent)),
|
||||
metadata);
|
||||
|
||||
tupleMap.put(prop.getGetterMethod(), childDsl);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
return entity;
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -127,10 +127,10 @@ public class DslInvocationHandler<E> implements InvocationHandler {
|
|||
}
|
||||
|
||||
if (DslExportable.SET_METADATA_METHOD.equals(methodName)
|
||||
&& args.length == 1
|
||||
&& args[0] instanceof Metadata) {
|
||||
&& args.length == 1
|
||||
&& args[0] instanceof Metadata) {
|
||||
if (metadata == null) {
|
||||
this.setCassandraMetadataForHelenusSesion((Metadata) args[0]);
|
||||
this.setCassandraMetadataForHelenusSesion((Metadata) args[0]);
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
|
|
@ -23,7 +23,6 @@ import java.lang.reflect.Method;
|
|||
import java.lang.reflect.Proxy;
|
||||
import java.util.Collections;
|
||||
import java.util.Map;
|
||||
|
||||
import net.helenus.core.Helenus;
|
||||
import net.helenus.mapping.annotation.Transient;
|
||||
import net.helenus.support.HelenusException;
|
||||
|
@ -94,7 +93,7 @@ public class MapperInvocationHandler<E> implements InvocationHandler, Serializab
|
|||
}
|
||||
|
||||
if ("dsl".equals(methodName)) {
|
||||
return Helenus.dsl(iface);
|
||||
return Helenus.dsl(iface);
|
||||
}
|
||||
|
||||
if (MapExportable.TO_MAP_METHOD.equals(methodName)) {
|
||||
|
|
|
@ -55,21 +55,22 @@ public final class HelenusMappingEntity implements HelenusEntity {
|
|||
|
||||
Map<String, Method> methods = new HashMap<String, Method>();
|
||||
for (Method m : iface.getDeclaredMethods()) {
|
||||
methods.put(m.getName(), m);
|
||||
methods.put(m.getName(), m);
|
||||
}
|
||||
|
||||
for (Class<?> c : ClassUtils.getAllInterfaces(iface)) {
|
||||
if (c.getDeclaredAnnotation(Table.class) != null || c.getDeclaredAnnotation(InheritedTable.class) != null) {
|
||||
if (c.getDeclaredAnnotation(Table.class) != null
|
||||
|| c.getDeclaredAnnotation(InheritedTable.class) != null) {
|
||||
for (Method m : c.getDeclaredMethods()) {
|
||||
Method o = methods.get(m.getName());
|
||||
if (o != null) {
|
||||
// Prefer overridden method implementation.
|
||||
if (o.getDeclaringClass().isAssignableFrom(m.getDeclaringClass())) {
|
||||
methods.put(m.getName(), m);
|
||||
}
|
||||
} else {
|
||||
methods.put(m.getName(), m);
|
||||
Method o = methods.get(m.getName());
|
||||
if (o != null) {
|
||||
// Prefer overridden method implementation.
|
||||
if (o.getDeclaringClass().isAssignableFrom(m.getDeclaringClass())) {
|
||||
methods.put(m.getName(), m);
|
||||
}
|
||||
} else {
|
||||
methods.put(m.getName(), m);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,7 +1,5 @@
|
|||
package net.helenus.mapping.annotation;
|
||||
|
||||
import net.helenus.core.Getter;
|
||||
|
||||
import java.lang.annotation.*;
|
||||
|
||||
/**
|
||||
|
@ -14,40 +12,39 @@ import java.lang.annotation.*;
|
|||
* <p>This is useful when you need to perform IN or SORT/ORDER-BY queries and to do so you'll need
|
||||
* different materialized table on disk in Cassandra.
|
||||
*
|
||||
* <p>For each @Table annotated interface Helenus will create/update/verify Cassandra Materialized Views
|
||||
* and some indexes if needed on startup.
|
||||
* <p>For each @Table annotated interface Helenus will create/update/verify Cassandra Materialized
|
||||
* Views and some indexes if needed on startup.
|
||||
*/
|
||||
@Inherited
|
||||
@Retention(RetentionPolicy.RUNTIME)
|
||||
@Target({ElementType.TYPE})
|
||||
public @interface CoveringIndex {
|
||||
|
||||
/**
|
||||
* Defined the name of the index. By default the entity name with column name as suffix.
|
||||
*
|
||||
* @return name of the covering index
|
||||
*/
|
||||
String name() default "";
|
||||
/**
|
||||
* Defined the name of the index. By default the entity name with column name as suffix.
|
||||
*
|
||||
* @return name of the covering index
|
||||
*/
|
||||
String name() default "";
|
||||
|
||||
/**
|
||||
* Set of fields in this entity to replicate in the index.
|
||||
*
|
||||
* @return array of the string names of the fields.
|
||||
*/
|
||||
String[] covering() default "";
|
||||
/**
|
||||
* Set of fields in this entity to replicate in the index.
|
||||
*
|
||||
* @return array of the string names of the fields.
|
||||
*/
|
||||
String[] covering() default "";
|
||||
|
||||
/**
|
||||
* Set of fields to use as the partition keys for this projection.
|
||||
*
|
||||
* @return array of the string names of the fields.
|
||||
*/
|
||||
String[] partitionKeys() default "";
|
||||
|
||||
/**
|
||||
* Set of fields to use as the clustering columns for this projection.
|
||||
*
|
||||
* @return array of the string names of the fields.
|
||||
*/
|
||||
String[] clusteringColumns() default "";
|
||||
/**
|
||||
* Set of fields to use as the partition keys for this projection.
|
||||
*
|
||||
* @return array of the string names of the fields.
|
||||
*/
|
||||
String[] partitionKeys() default "";
|
||||
|
||||
/**
|
||||
* Set of fields to use as the clustering columns for this projection.
|
||||
*
|
||||
* @return array of the string names of the fields.
|
||||
*/
|
||||
String[] clusteringColumns() default "";
|
||||
}
|
||||
|
|
|
@ -25,29 +25,28 @@ import java.lang.annotation.*;
|
|||
* <p>This is useful when you need to perform IN or SORT/ORDER-BY queries and to do so you'll need
|
||||
* different materialized table on disk in Cassandra.
|
||||
*
|
||||
* <p>For each @Table annotated interface Helenus will create/update/verify Cassandra Materialized Views
|
||||
* and some indexes if needed on startup.
|
||||
* <p>For each @Table annotated interface Helenus will create/update/verify Cassandra Materialized
|
||||
* Views and some indexes if needed on startup.
|
||||
*/
|
||||
@Inherited
|
||||
@Retention(RetentionPolicy.RUNTIME)
|
||||
@Target({ElementType.TYPE})
|
||||
public @interface MaterializedView {
|
||||
|
||||
/**
|
||||
* Default value is the SimpleName of the interface normalized to underscore
|
||||
*
|
||||
* @return name of the type
|
||||
*/
|
||||
String value() default "";
|
||||
|
||||
/**
|
||||
* For reserved words in Cassandra we need quotation in CQL queries. This property marks that the
|
||||
* name of the type needs to be quoted.
|
||||
*
|
||||
* <p>Default value is false, we are quoting only selected names.
|
||||
*
|
||||
* @return true if name have to be quoted
|
||||
*/
|
||||
boolean forceQuote() default false;
|
||||
/**
|
||||
* Default value is the SimpleName of the interface normalized to underscore
|
||||
*
|
||||
* @return name of the type
|
||||
*/
|
||||
String value() default "";
|
||||
|
||||
/**
|
||||
* For reserved words in Cassandra we need quotation in CQL queries. This property marks that the
|
||||
* name of the type needs to be quoted.
|
||||
*
|
||||
* <p>Default value is false, we are quoting only selected names.
|
||||
*
|
||||
* @return true if name have to be quoted
|
||||
*/
|
||||
boolean forceQuote() default false;
|
||||
}
|
||||
|
|
|
@ -2,6 +2,7 @@ package net.helenus.mapping.javatype;
|
|||
|
||||
public abstract class AbstractCollectionJavaType extends AbstractJavaType {
|
||||
|
||||
public static boolean isCollectionType() { return true; }
|
||||
|
||||
public static boolean isCollectionType() {
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -33,7 +33,9 @@ import net.helenus.support.HelenusMappingException;
|
|||
|
||||
public abstract class AbstractJavaType {
|
||||
|
||||
public static boolean isCollectionType() { return false; }
|
||||
public static boolean isCollectionType() {
|
||||
return false;
|
||||
}
|
||||
|
||||
public abstract Class<?> getJavaClass();
|
||||
|
||||
|
|
|
@ -18,7 +18,6 @@ package net.helenus.mapping.javatype;
|
|||
import com.datastax.driver.core.*;
|
||||
import java.lang.reflect.Method;
|
||||
import java.lang.reflect.Type;
|
||||
import java.util.AbstractCollection;
|
||||
import java.util.Map;
|
||||
import java.util.Optional;
|
||||
import java.util.function.Function;
|
||||
|
|
|
@ -4,10 +4,11 @@ import net.helenus.mapping.ColumnType;
|
|||
|
||||
public abstract class AbstractCollectionDataType extends AbstractDataType {
|
||||
|
||||
public AbstractCollectionDataType(ColumnType columnType) {
|
||||
super(columnType);
|
||||
}
|
||||
|
||||
public boolean isCollectionType() { return true; }
|
||||
public AbstractCollectionDataType(ColumnType columnType) {
|
||||
super(columnType);
|
||||
}
|
||||
|
||||
public boolean isCollectionType() {
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -55,6 +55,7 @@ public abstract class AbstractDataType {
|
|||
"wrong column type " + columnType + " for UserDefinedType in columnName " + columnName);
|
||||
}
|
||||
|
||||
public boolean isCollectionType() { return false; }
|
||||
|
||||
public boolean isCollectionType() {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -18,8 +18,6 @@ package net.helenus.mapping.type;
|
|||
import com.datastax.driver.core.DataType;
|
||||
import com.datastax.driver.core.UserType;
|
||||
import com.datastax.driver.core.schemabuilder.*;
|
||||
|
||||
import java.util.AbstractCollection;
|
||||
import java.util.List;
|
||||
import net.helenus.mapping.ColumnType;
|
||||
import net.helenus.mapping.IdentityName;
|
||||
|
|
|
@ -33,7 +33,7 @@ public enum BeanColumnValueProvider implements ColumnValueProvider {
|
|||
try {
|
||||
value = getter.invoke(bean, new Object[] {});
|
||||
} catch (InvocationTargetException e) {
|
||||
if (e.getCause() != null ) {
|
||||
if (e.getCause() != null) {
|
||||
throw new HelenusException("getter threw an exception", e.getCause());
|
||||
}
|
||||
} catch (ReflectiveOperationException e) {
|
||||
|
|
|
@ -67,11 +67,8 @@ public class CollectionTest extends AbstractEmbeddedCassandraTest {
|
|||
|
||||
// read full object
|
||||
|
||||
Customer actual = session.<Customer>select(customer)
|
||||
.where(customer::id, eq(id))
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
Customer actual =
|
||||
session.<Customer>select(customer).where(customer::id, eq(id)).single().sync().orElse(null);
|
||||
Assert.assertEquals(id, actual.id());
|
||||
Assert.assertEquals(aliases, actual.aliases());
|
||||
Assert.assertNull(actual.names());
|
||||
|
@ -90,11 +87,8 @@ public class CollectionTest extends AbstractEmbeddedCassandraTest {
|
|||
|
||||
session.update().set(customer::aliases, expected).where(customer::id, eq(id)).sync();
|
||||
|
||||
actual = session.<Customer>select(customer)
|
||||
.where(customer::id, eq(id))
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
actual =
|
||||
session.<Customer>select(customer).where(customer::id, eq(id)).single().sync().orElse(null);
|
||||
|
||||
Assert.assertEquals(id, actual.id());
|
||||
Assert.assertEquals(expected, actual.aliases());
|
||||
|
@ -170,11 +164,8 @@ public class CollectionTest extends AbstractEmbeddedCassandraTest {
|
|||
|
||||
// read full object
|
||||
|
||||
Customer actual = session.<Customer>select(customer)
|
||||
.where(customer::id, eq(id))
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
Customer actual =
|
||||
session.<Customer>select(customer).where(customer::id, eq(id)).single().sync().orElse(null);
|
||||
|
||||
Assert.assertEquals(id, actual.id());
|
||||
Assert.assertEquals(names, actual.names());
|
||||
|
@ -200,11 +191,8 @@ public class CollectionTest extends AbstractEmbeddedCassandraTest {
|
|||
|
||||
session.update().set(customer::names, expected).where(customer::id, eq(id)).sync();
|
||||
|
||||
actual = session.<Customer>select(customer)
|
||||
.where(customer::id, eq(id))
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
actual =
|
||||
session.<Customer>select(customer).where(customer::id, eq(id)).single().sync().orElse(null);
|
||||
|
||||
Assert.assertEquals(id, actual.id());
|
||||
Assert.assertEquals(expected, actual.names());
|
||||
|
@ -306,10 +294,8 @@ public class CollectionTest extends AbstractEmbeddedCassandraTest {
|
|||
|
||||
// read full object
|
||||
|
||||
Customer actual = session.<Customer>select(customer)
|
||||
.where(customer::id, eq(id)).single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
Customer actual =
|
||||
session.<Customer>select(customer).where(customer::id, eq(id)).single().sync().orElse(null);
|
||||
|
||||
Assert.assertEquals(id, actual.id());
|
||||
Assert.assertEquals(props, actual.properties());
|
||||
|
@ -343,11 +329,8 @@ public class CollectionTest extends AbstractEmbeddedCassandraTest {
|
|||
|
||||
session.update().set(customer::properties, expected).where(customer::id, eq(id)).sync();
|
||||
|
||||
actual = session.<Customer>select(customer)
|
||||
.where(customer::id, eq(id))
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
actual =
|
||||
session.<Customer>select(customer).where(customer::id, eq(id)).single().sync().orElse(null);
|
||||
Assert.assertEquals(id, actual.id());
|
||||
Assert.assertEquals(expected, actual.properties());
|
||||
|
||||
|
|
|
@ -18,76 +18,72 @@ package net.helenus.test.integration.core.draft;
|
|||
import java.util.ArrayList;
|
||||
import java.util.HashMap;
|
||||
import java.util.HashSet;
|
||||
|
||||
import net.helenus.core.Helenus;
|
||||
import net.helenus.core.HelenusSession;
|
||||
import net.helenus.test.integration.build.AbstractEmbeddedCassandraTest;
|
||||
import org.junit.Assert;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
|
||||
import net.helenus.core.Helenus;
|
||||
import net.helenus.core.HelenusSession;
|
||||
import net.helenus.test.integration.build.AbstractEmbeddedCassandraTest;
|
||||
|
||||
import static net.helenus.core.Query.eq;
|
||||
|
||||
public class EntityDraftBuilderTest extends AbstractEmbeddedCassandraTest {
|
||||
|
||||
static Supply supply;
|
||||
static HelenusSession session;
|
||||
|
||||
static Supply supply;
|
||||
static HelenusSession session;
|
||||
@BeforeClass
|
||||
public static void beforeTest() {
|
||||
session = Helenus.init(getSession()).showCql().add(Supply.class).autoCreateDrop().get();
|
||||
supply = session.dsl(Supply.class);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFoo() throws Exception {
|
||||
Supply.Draft draft = null;
|
||||
|
||||
@BeforeClass
|
||||
public static void beforeTest() {
|
||||
session = Helenus.init(getSession())
|
||||
.showCql()
|
||||
.add(Supply.class)
|
||||
.autoCreateDrop()
|
||||
.get();
|
||||
supply = session.dsl(Supply.class);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFoo() throws Exception {
|
||||
Supply.Draft draft = null;
|
||||
|
||||
draft = Supply.draft("APAC")
|
||||
.code("WIDGET-002")
|
||||
.description("Our second Widget!")
|
||||
.demand(new HashMap<String, Long>() {{
|
||||
draft =
|
||||
Supply.draft("APAC")
|
||||
.code("WIDGET-002")
|
||||
.description("Our second Widget!")
|
||||
.demand(
|
||||
new HashMap<String, Long>() {
|
||||
{
|
||||
put("APAC", 100L);
|
||||
put("EMEA", 10000L);
|
||||
put("NORAM", 2000000L);
|
||||
}})
|
||||
.shipments(new HashSet<String>() {{
|
||||
}
|
||||
})
|
||||
.shipments(
|
||||
new HashSet<String>() {
|
||||
{
|
||||
add("HMS Puddle in transit to APAC, 100 units.");
|
||||
add("Frigate Jimmy in transit to EMEA, 10000 units.");
|
||||
}})
|
||||
.suppliers(new ArrayList<String>() {{
|
||||
}
|
||||
})
|
||||
.suppliers(
|
||||
new ArrayList<String>() {
|
||||
{
|
||||
add("Puddle, Inc.");
|
||||
add("Jimmy Town, LTD.");
|
||||
}});
|
||||
}
|
||||
});
|
||||
|
||||
Supply s1 = session.<Supply>insert(draft)
|
||||
.sync();
|
||||
Supply s1 = session.<Supply>insert(draft).sync();
|
||||
|
||||
// List
|
||||
Supply s2 = session.<Supply>update(s1.update())
|
||||
.prepend(supply::suppliers, "Pignose Supply, LLC.")
|
||||
.sync();
|
||||
Assert.assertEquals(s2.suppliers().get(0), "Pignose Supply, LLC.");
|
||||
// List
|
||||
Supply s2 =
|
||||
session
|
||||
.<Supply>update(s1.update())
|
||||
.prepend(supply::suppliers, "Pignose Supply, LLC.")
|
||||
.sync();
|
||||
Assert.assertEquals(s2.suppliers().get(0), "Pignose Supply, LLC.");
|
||||
|
||||
// Set
|
||||
String shipment = "Pignose, on the way! (1M units)";
|
||||
Supply s3 = session.<Supply>update(s2.update())
|
||||
.add(supply::shipments, shipment)
|
||||
.sync();
|
||||
Assert.assertTrue(s3.shipments().contains(shipment));
|
||||
// Set
|
||||
String shipment = "Pignose, on the way! (1M units)";
|
||||
Supply s3 = session.<Supply>update(s2.update()).add(supply::shipments, shipment).sync();
|
||||
Assert.assertTrue(s3.shipments().contains(shipment));
|
||||
|
||||
// Map
|
||||
Supply s4 = session.<Supply>update(s3.update())
|
||||
.put(supply::demand, "NORAM", 10L)
|
||||
.sync();
|
||||
Assert.assertEquals((long)s4.demand().get("NORAM"), 10L);
|
||||
|
||||
}
|
||||
// Map
|
||||
Supply s4 = session.<Supply>update(s3.update()).put(supply::demand, "NORAM", 10L).sync();
|
||||
Assert.assertEquals((long) s4.demand().get("NORAM"), 10L);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,78 +1,93 @@
|
|||
package net.helenus.test.integration.core.draft;
|
||||
|
||||
import java.util.UUID;
|
||||
|
||||
import net.helenus.core.AbstractAuditedEntityDraft;
|
||||
import net.helenus.core.Helenus;
|
||||
import net.helenus.core.reflect.MapExportable;
|
||||
import net.helenus.mapping.annotation.*;
|
||||
|
||||
|
||||
@Table
|
||||
public interface Inventory {
|
||||
|
||||
static Inventory inventory = Helenus.dsl(Inventory.class);
|
||||
static Inventory inventory = Helenus.dsl(Inventory.class);
|
||||
|
||||
@PartitionKey UUID id();
|
||||
@Column("emea") @Types.Counter long EMEA();
|
||||
@Column("noram") @Types.Counter long NORAM();
|
||||
@Column("apac") @Types.Counter long APAC();
|
||||
@PartitionKey
|
||||
UUID id();
|
||||
|
||||
@Transient static Draft draft(UUID id) { return new Draft(id); }
|
||||
@Column("emea")
|
||||
@Types.Counter
|
||||
long EMEA();
|
||||
|
||||
@Transient default Draft update() { return new Draft(this); }
|
||||
@Column("noram")
|
||||
@Types.Counter
|
||||
long NORAM();
|
||||
|
||||
@Column("apac")
|
||||
@Types.Counter
|
||||
long APAC();
|
||||
|
||||
class Draft extends AbstractAuditedEntityDraft<Inventory> {
|
||||
@Transient
|
||||
static Draft draft(UUID id) {
|
||||
return new Draft(id);
|
||||
}
|
||||
|
||||
// Entity/Draft pattern-enabling methods:
|
||||
Draft(UUID id) {
|
||||
super(null);
|
||||
@Transient
|
||||
default Draft update() {
|
||||
return new Draft(this);
|
||||
}
|
||||
|
||||
// Primary Key:
|
||||
set(inventory::id, id);
|
||||
}
|
||||
class Draft extends AbstractAuditedEntityDraft<Inventory> {
|
||||
|
||||
Draft(Inventory inventory) {
|
||||
super((MapExportable) inventory);
|
||||
}
|
||||
|
||||
public Class<Inventory> getEntityClass() { return Inventory.class; }
|
||||
|
||||
protected String getCurrentAuditor() { return "unknown"; }
|
||||
|
||||
// Immutable properties:
|
||||
public UUID id() {
|
||||
return this.<UUID>get(inventory::id, UUID.class);
|
||||
}
|
||||
|
||||
public long EMEA() {
|
||||
return this.<Long>get(inventory::EMEA, long.class);
|
||||
}
|
||||
|
||||
public Draft EMEA(long count) {
|
||||
mutate(inventory::EMEA, count);
|
||||
return this;
|
||||
}
|
||||
|
||||
public long APAC() {
|
||||
return this.<Long>get(inventory::APAC, long.class);
|
||||
}
|
||||
|
||||
public Draft APAC(long count) {
|
||||
mutate(inventory::APAC, count);
|
||||
return this;
|
||||
}
|
||||
|
||||
public long NORAM() {
|
||||
return this.<Long>get(inventory::NORAM, long.class);
|
||||
}
|
||||
|
||||
public Draft NORAM(long count) {
|
||||
mutate(inventory::NORAM, count);
|
||||
return this;
|
||||
}
|
||||
// Entity/Draft pattern-enabling methods:
|
||||
Draft(UUID id) {
|
||||
super(null);
|
||||
|
||||
// Primary Key:
|
||||
set(inventory::id, id);
|
||||
}
|
||||
|
||||
Draft(Inventory inventory) {
|
||||
super((MapExportable) inventory);
|
||||
}
|
||||
|
||||
public Class<Inventory> getEntityClass() {
|
||||
return Inventory.class;
|
||||
}
|
||||
|
||||
protected String getCurrentAuditor() {
|
||||
return "unknown";
|
||||
}
|
||||
|
||||
// Immutable properties:
|
||||
public UUID id() {
|
||||
return this.<UUID>get(inventory::id, UUID.class);
|
||||
}
|
||||
|
||||
public long EMEA() {
|
||||
return this.<Long>get(inventory::EMEA, long.class);
|
||||
}
|
||||
|
||||
public Draft EMEA(long count) {
|
||||
mutate(inventory::EMEA, count);
|
||||
return this;
|
||||
}
|
||||
|
||||
public long APAC() {
|
||||
return this.<Long>get(inventory::APAC, long.class);
|
||||
}
|
||||
|
||||
public Draft APAC(long count) {
|
||||
mutate(inventory::APAC, count);
|
||||
return this;
|
||||
}
|
||||
|
||||
public long NORAM() {
|
||||
return this.<Long>get(inventory::NORAM, long.class);
|
||||
}
|
||||
|
||||
public Draft NORAM(long count) {
|
||||
mutate(inventory::NORAM, count);
|
||||
return this;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,129 +1,145 @@
|
|||
package net.helenus.test.integration.core.draft;
|
||||
|
||||
import java.lang.reflect.Proxy;
|
||||
import com.datastax.driver.core.utils.UUIDs;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
import java.util.UUID;
|
||||
|
||||
import com.datastax.driver.core.utils.UUIDs;
|
||||
|
||||
import net.helenus.core.AbstractEntityDraft;
|
||||
import net.helenus.core.Helenus;
|
||||
import net.helenus.core.reflect.MapExportable;
|
||||
import net.helenus.mapping.annotation.*;
|
||||
|
||||
|
||||
@Table
|
||||
public interface Supply {
|
||||
|
||||
static Supply supply = Helenus.dsl(Supply.class);
|
||||
static Supply supply = Helenus.dsl(Supply.class);
|
||||
|
||||
@PartitionKey UUID id();
|
||||
@ClusteringColumn(ordinal=0) default String region() { return "NORAM"; }
|
||||
@PartitionKey
|
||||
UUID id();
|
||||
|
||||
@Index(caseSensitive = false) String code();
|
||||
@Index String description(); // @IndexText == lucene index
|
||||
@Index Map<String, Long> demand();
|
||||
@Index List<String> suppliers();
|
||||
@Index Set<String> shipments();
|
||||
@ClusteringColumn(ordinal = 0)
|
||||
default String region() {
|
||||
return "NORAM";
|
||||
}
|
||||
|
||||
@Transient static Draft draft(String region) { return new Draft(region); }
|
||||
@Index(caseSensitive = false)
|
||||
String code();
|
||||
|
||||
@Transient default Draft update() { return new Draft(this); }
|
||||
@Index
|
||||
String description(); // @IndexText == lucene index
|
||||
|
||||
@Index
|
||||
Map<String, Long> demand();
|
||||
|
||||
class Draft extends AbstractEntityDraft<Supply> {
|
||||
@Index
|
||||
List<String> suppliers();
|
||||
|
||||
// Entity/Draft pattern-enabling methods:
|
||||
Draft(String region) {
|
||||
super(null);
|
||||
@Index
|
||||
Set<String> shipments();
|
||||
|
||||
// Primary Key:
|
||||
set(supply::id, UUIDs.timeBased());
|
||||
set(supply::region, region);
|
||||
}
|
||||
@Transient
|
||||
static Draft draft(String region) {
|
||||
return new Draft(region);
|
||||
}
|
||||
|
||||
Draft(Supply supply) {
|
||||
super((MapExportable) supply);
|
||||
}
|
||||
@Transient
|
||||
default Draft update() {
|
||||
return new Draft(this);
|
||||
}
|
||||
|
||||
public Class<Supply> getEntityClass() { return Supply.class; }
|
||||
class Draft extends AbstractEntityDraft<Supply> {
|
||||
|
||||
// Immutable properties:
|
||||
public UUID id() {
|
||||
return this.<UUID>get(supply::id, UUID.class);
|
||||
}
|
||||
|
||||
public String region() {
|
||||
return this.<String>get(supply::region, String.class);
|
||||
}
|
||||
|
||||
// Mutable properties:
|
||||
public String code() {
|
||||
return this.<String>get(supply::code, String.class);
|
||||
}
|
||||
|
||||
public Draft code(String code) {
|
||||
mutate(supply::code, code);
|
||||
return this;
|
||||
}
|
||||
|
||||
public Draft setCode(String code) {
|
||||
return code(code);
|
||||
}
|
||||
|
||||
public String description() {
|
||||
return this.<String>get(supply::description, String.class);
|
||||
}
|
||||
|
||||
public Draft description(String description) {
|
||||
mutate(supply::description, description);
|
||||
return this;
|
||||
}
|
||||
|
||||
public Draft setDescription(String description) {
|
||||
return description(description);
|
||||
}
|
||||
|
||||
public Map<String, Long> demand() {
|
||||
return this.<Map<String, Long>>get(supply::demand, Map.class);
|
||||
}
|
||||
|
||||
public Draft demand(Map<String, Long> demand) {
|
||||
mutate(supply::demand, demand);
|
||||
return this;
|
||||
}
|
||||
|
||||
public Draft setDemand(Map<String, Long> demand) {
|
||||
return demand(demand);
|
||||
}
|
||||
|
||||
public List<String> suppliers() {
|
||||
return this.<List<String>>get(supply::suppliers, List.class);
|
||||
}
|
||||
|
||||
public Draft suppliers(List<String> suppliers) {
|
||||
mutate(supply::suppliers, suppliers);
|
||||
return this;
|
||||
}
|
||||
|
||||
public Draft setSuppliers(List<String> suppliers) {
|
||||
return suppliers(suppliers);
|
||||
}
|
||||
|
||||
public Set<String> shipments() {
|
||||
return this.<Set<String>>get(supply::shipments, Set.class);
|
||||
}
|
||||
|
||||
public Draft shipments(Set<String> shipments) {
|
||||
mutate(supply::shipments, shipments);
|
||||
return this;
|
||||
}
|
||||
|
||||
public Draft setshipments(Set<String> shipments) {
|
||||
return shipments(shipments);
|
||||
}
|
||||
// Entity/Draft pattern-enabling methods:
|
||||
Draft(String region) {
|
||||
super(null);
|
||||
|
||||
// Primary Key:
|
||||
set(supply::id, UUIDs.timeBased());
|
||||
set(supply::region, region);
|
||||
}
|
||||
|
||||
Draft(Supply supply) {
|
||||
super((MapExportable) supply);
|
||||
}
|
||||
|
||||
public Class<Supply> getEntityClass() {
|
||||
return Supply.class;
|
||||
}
|
||||
|
||||
// Immutable properties:
|
||||
public UUID id() {
|
||||
return this.<UUID>get(supply::id, UUID.class);
|
||||
}
|
||||
|
||||
public String region() {
|
||||
return this.<String>get(supply::region, String.class);
|
||||
}
|
||||
|
||||
// Mutable properties:
|
||||
public String code() {
|
||||
return this.<String>get(supply::code, String.class);
|
||||
}
|
||||
|
||||
public Draft code(String code) {
|
||||
mutate(supply::code, code);
|
||||
return this;
|
||||
}
|
||||
|
||||
public Draft setCode(String code) {
|
||||
return code(code);
|
||||
}
|
||||
|
||||
public String description() {
|
||||
return this.<String>get(supply::description, String.class);
|
||||
}
|
||||
|
||||
public Draft description(String description) {
|
||||
mutate(supply::description, description);
|
||||
return this;
|
||||
}
|
||||
|
||||
public Draft setDescription(String description) {
|
||||
return description(description);
|
||||
}
|
||||
|
||||
public Map<String, Long> demand() {
|
||||
return this.<Map<String, Long>>get(supply::demand, Map.class);
|
||||
}
|
||||
|
||||
public Draft demand(Map<String, Long> demand) {
|
||||
mutate(supply::demand, demand);
|
||||
return this;
|
||||
}
|
||||
|
||||
public Draft setDemand(Map<String, Long> demand) {
|
||||
return demand(demand);
|
||||
}
|
||||
|
||||
public List<String> suppliers() {
|
||||
return this.<List<String>>get(supply::suppliers, List.class);
|
||||
}
|
||||
|
||||
public Draft suppliers(List<String> suppliers) {
|
||||
mutate(supply::suppliers, suppliers);
|
||||
return this;
|
||||
}
|
||||
|
||||
public Draft setSuppliers(List<String> suppliers) {
|
||||
return suppliers(suppliers);
|
||||
}
|
||||
|
||||
public Set<String> shipments() {
|
||||
return this.<Set<String>>get(supply::shipments, Set.class);
|
||||
}
|
||||
|
||||
public Draft shipments(Set<String> shipments) {
|
||||
mutate(supply::shipments, shipments);
|
||||
return this;
|
||||
}
|
||||
|
||||
public Draft setshipments(Set<String> shipments) {
|
||||
return shipments(shipments);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -65,7 +65,8 @@ public class HierarchyTest extends AbstractEmbeddedCassandraTest {
|
|||
.value(cat::nickname, "garfield")
|
||||
.value(cat::eatable, false)
|
||||
.sync();
|
||||
Optional<Cat> animal = session.select(Cat.class).where(cat::nickname, eq("garfield")).single().sync();
|
||||
Optional<Cat> animal =
|
||||
session.select(Cat.class).where(cat::nickname, eq("garfield")).single().sync();
|
||||
Assert.assertTrue(animal.isPresent());
|
||||
|
||||
Cat cat = animal.get();
|
||||
|
|
|
@ -20,6 +20,7 @@ import net.helenus.mapping.annotation.InheritedTable;
|
|||
@InheritedTable
|
||||
public interface Mammal extends Animal {
|
||||
|
||||
default boolean warmBlodded() { return true; }
|
||||
|
||||
default boolean warmBlodded() {
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -18,21 +18,17 @@ package net.helenus.test.integration.core.simple;
|
|||
import static net.helenus.core.Query.eq;
|
||||
|
||||
import com.datastax.driver.core.ResultSet;
|
||||
import java.util.*;
|
||||
import net.helenus.core.Helenus;
|
||||
import net.helenus.core.HelenusSession;
|
||||
import net.helenus.core.Operator;
|
||||
import net.helenus.core.operation.UpdateOperation;
|
||||
import net.helenus.core.reflect.Drafted;
|
||||
import net.helenus.mapping.HelenusEntity;
|
||||
import net.helenus.support.Fun;
|
||||
import net.helenus.test.integration.build.AbstractEmbeddedCassandraTest;
|
||||
import org.junit.Assert;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
|
||||
import java.util.*;
|
||||
|
||||
|
||||
public class SimpleUserTest extends AbstractEmbeddedCassandraTest {
|
||||
|
||||
static User user;
|
||||
|
@ -102,11 +98,7 @@ public class SimpleUserTest extends AbstractEmbeddedCassandraTest {
|
|||
|
||||
// select as object
|
||||
|
||||
actual = session.<User>select(user)
|
||||
.where(user::id, eq(100L))
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
actual = session.<User>select(user).where(user::id, eq(100L)).single().sync().orElse(null);
|
||||
assertUsers(newUser, actual);
|
||||
|
||||
// select by columns
|
||||
|
@ -192,11 +184,7 @@ public class SimpleUserTest extends AbstractEmbeddedCassandraTest {
|
|||
|
||||
Assert.assertEquals("_albert", name);
|
||||
|
||||
User u2 = session.<User>select(user)
|
||||
.where(user::id, eq(100L))
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
User u2 = session.<User>select(user).where(user::id, eq(100L)).single().sync().orElse(null);
|
||||
|
||||
Assert.assertEquals(Long.valueOf(100L), u2.id());
|
||||
Assert.assertEquals("albert", u2.name());
|
||||
|
@ -204,31 +192,27 @@ public class SimpleUserTest extends AbstractEmbeddedCassandraTest {
|
|||
|
||||
//
|
||||
User greg =
|
||||
session
|
||||
.<User>insert(user)
|
||||
.value(user::name, "greg")
|
||||
.value(user::age, 44)
|
||||
.value(user::type, UserType.USER)
|
||||
.value(user::id, 1234L)
|
||||
.sync();
|
||||
session
|
||||
.<User>insert(user)
|
||||
.value(user::name, "greg")
|
||||
.value(user::age, 44)
|
||||
.value(user::type, UserType.USER)
|
||||
.value(user::id, 1234L)
|
||||
.sync();
|
||||
|
||||
Optional<User> maybeGreg =
|
||||
session
|
||||
.<User>select(user)
|
||||
.where(user::id, eq(1234L))
|
||||
.single()
|
||||
.sync();
|
||||
session.<User>select(user).where(user::id, eq(1234L)).single().sync();
|
||||
|
||||
// INSERT
|
||||
|
||||
session
|
||||
.update()
|
||||
.set(user::name, null)
|
||||
.set(user::age, null)
|
||||
.set(user::type, null)
|
||||
.where(user::id, eq(100L))
|
||||
.zipkinContext(null)
|
||||
.sync();
|
||||
.update()
|
||||
.set(user::name, null)
|
||||
.set(user::age, null)
|
||||
.set(user::type, null)
|
||||
.where(user::id, eq(100L))
|
||||
.zipkinContext(null)
|
||||
.sync();
|
||||
|
||||
Fun.Tuple3<String, Integer, UserType> tuple =
|
||||
session
|
||||
|
@ -252,21 +236,16 @@ public class SimpleUserTest extends AbstractEmbeddedCassandraTest {
|
|||
|
||||
public void testZipkin() throws Exception {
|
||||
session
|
||||
.update()
|
||||
.set(user::name, null)
|
||||
.set(user::age, null)
|
||||
.set(user::type, null)
|
||||
.where(user::id, eq(100L))
|
||||
.zipkinContext(null)
|
||||
.sync();
|
||||
|
||||
.update()
|
||||
.set(user::name, null)
|
||||
.set(user::age, null)
|
||||
.set(user::type, null)
|
||||
.where(user::id, eq(100L))
|
||||
.zipkinContext(null)
|
||||
.sync();
|
||||
|
||||
UpdateOperation<ResultSet> update = session.update();
|
||||
update
|
||||
.set(user::name, null)
|
||||
.zipkinContext(null)
|
||||
.sync();
|
||||
|
||||
update.set(user::name, null).zipkinContext(null).sync();
|
||||
}
|
||||
|
||||
private void assertUsers(User expected, User actual) {
|
||||
|
|
|
@ -42,11 +42,8 @@ public class TupleMapTest extends TupleCollectionTest {
|
|||
|
||||
// read full object
|
||||
|
||||
Book actual = session.<Book>select(book)
|
||||
.where(book::id, Query.eq(id))
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
Book actual =
|
||||
session.<Book>select(book).where(book::id, Query.eq(id)).single().sync().orElse(null);
|
||||
Assert.assertEquals(id, actual.id());
|
||||
assertEqualMaps(writers, actual.writers());
|
||||
Assert.assertNull(actual.reviewers());
|
||||
|
@ -77,11 +74,7 @@ public class TupleMapTest extends TupleCollectionTest {
|
|||
|
||||
session.update().set(book::writers, expected).where(book::id, Query.eq(id)).sync();
|
||||
|
||||
actual = session.<Book>select(book)
|
||||
.where(book::id, Query.eq(id))
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
actual = session.<Book>select(book).where(book::id, Query.eq(id)).single().sync().orElse(null);
|
||||
|
||||
Assert.assertEquals(id, actual.id());
|
||||
assertEqualMaps(expected, actual.writers());
|
||||
|
@ -96,7 +89,8 @@ public class TupleMapTest extends TupleCollectionTest {
|
|||
expected.put(third, unk);
|
||||
session.update().put(book::writers, third, unk).where(book::id, Query.eq(id)).sync();
|
||||
|
||||
actualMap = session.select(book::writers).where(book::id, Query.eq(id)).sync().findFirst().get()._1;
|
||||
actualMap =
|
||||
session.select(book::writers).where(book::id, Query.eq(id)).sync().findFirst().get()._1;
|
||||
assertEqualMaps(expected, actualMap);
|
||||
|
||||
// putAll operation
|
||||
|
|
|
@ -15,6 +15,9 @@
|
|||
*/
|
||||
package net.helenus.test.integration.core.unitofwork;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.List;
|
||||
import net.helenus.core.Helenus;
|
||||
import net.helenus.core.HelenusSession;
|
||||
import net.helenus.core.UnitOfWork;
|
||||
|
@ -23,102 +26,125 @@ import org.junit.Assert;
|
|||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.List;
|
||||
|
||||
|
||||
public class AndThenOrderTest extends AbstractEmbeddedCassandraTest {
|
||||
|
||||
static HelenusSession session;
|
||||
static HelenusSession session;
|
||||
|
||||
@BeforeClass
|
||||
public static void beforeTest() {
|
||||
session = Helenus.init(getSession())
|
||||
.showCql()
|
||||
.autoCreateDrop()
|
||||
.get();
|
||||
}
|
||||
@BeforeClass
|
||||
public static void beforeTest() {
|
||||
session = Helenus.init(getSession()).showCql().autoCreateDrop().get();
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testAndThenOrdering() throws Exception {
|
||||
List<String> q = new ArrayList<String>(5);
|
||||
UnitOfWork uow1, uow2, uow3, uow4, uow5;
|
||||
@Test
|
||||
public void testAndThenOrdering() throws Exception {
|
||||
List<String> q = new ArrayList<String>(5);
|
||||
UnitOfWork uow1, uow2, uow3, uow4, uow5;
|
||||
|
||||
uow5 = session.begin();
|
||||
uow3 = session.begin(uow5);
|
||||
uow1 = session.begin(uow3);
|
||||
uow1.commit().andThen(() -> { q.add("1"); });
|
||||
uow2 = session.begin(uow3);
|
||||
uow2.commit().andThen(() -> { q.add("2"); });
|
||||
uow3.commit().andThen(() -> { q.add("3"); });
|
||||
uow4 = session.begin(uow5);
|
||||
uow4.commit().andThen(() -> { q.add("4"); });
|
||||
uow5.commit().andThen(() -> { q.add("5"); });
|
||||
uow5 = session.begin();
|
||||
uow3 = session.begin(uow5);
|
||||
uow1 = session.begin(uow3);
|
||||
uow1.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
q.add("1");
|
||||
});
|
||||
uow2 = session.begin(uow3);
|
||||
uow2.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
q.add("2");
|
||||
});
|
||||
uow3.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
q.add("3");
|
||||
});
|
||||
uow4 = session.begin(uow5);
|
||||
uow4.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
q.add("4");
|
||||
});
|
||||
uow5.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
q.add("5");
|
||||
});
|
||||
|
||||
System.out.println(q);
|
||||
Assert.assertTrue(Arrays.equals(q.toArray(new String[5]), new String[] {"1", "2", "3", "4", "5"}));
|
||||
System.out.println(q);
|
||||
Assert.assertTrue(
|
||||
Arrays.equals(q.toArray(new String[5]), new String[] {"1", "2", "3", "4", "5"}));
|
||||
}
|
||||
|
||||
}
|
||||
@Test
|
||||
public void testExceptionWithinAndThen() throws Exception {
|
||||
List<String> q = new ArrayList<String>(5);
|
||||
UnitOfWork uow1, uow2, uow3, uow4, uow5;
|
||||
|
||||
@Test
|
||||
public void testExceptionWithinAndThen() throws Exception {
|
||||
List<String> q = new ArrayList<String>(5);
|
||||
UnitOfWork uow1, uow2, uow3, uow4, uow5;
|
||||
|
||||
uow5 = session.begin();
|
||||
uow4 = session.begin(uow5);
|
||||
try {
|
||||
uow3 = session.begin(uow4);
|
||||
uow1 = session.begin(uow3);
|
||||
uow1.commit().andThen(() -> {
|
||||
uow5 = session.begin();
|
||||
uow4 = session.begin(uow5);
|
||||
try {
|
||||
uow3 = session.begin(uow4);
|
||||
uow1 = session.begin(uow3);
|
||||
uow1.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
q.add("1");
|
||||
});
|
||||
uow2 = session.begin(uow3);
|
||||
uow2.commit().andThen(() -> {
|
||||
});
|
||||
uow2 = session.begin(uow3);
|
||||
uow2.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
q.add("2");
|
||||
});
|
||||
uow3.commit().andThen(() -> {
|
||||
});
|
||||
uow3.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
q.add("3");
|
||||
});
|
||||
uow4.commit().andThen(() -> {
|
||||
});
|
||||
uow4.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
q.add("4");
|
||||
});
|
||||
throw new Exception();
|
||||
} catch (Exception e) {
|
||||
uow4.abort();
|
||||
}
|
||||
uow5.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
q.add("5");
|
||||
});
|
||||
throw new Exception();
|
||||
} catch(Exception e) {
|
||||
uow4.abort();
|
||||
}
|
||||
uow5.commit().andThen(() -> { q.add("5"); });
|
||||
|
||||
System.out.println(q);
|
||||
Assert.assertTrue(q.isEmpty() == true);
|
||||
System.out.println(q);
|
||||
Assert.assertTrue(q.isEmpty() == true);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testClosableWillAbortWhenNotCommitted() throws Exception {
|
||||
UnitOfWork unitOfWork;
|
||||
try (UnitOfWork uow = session.begin()) {
|
||||
unitOfWork = uow;
|
||||
Assert.assertFalse(uow.hasAborted());
|
||||
}
|
||||
Assert.assertTrue(unitOfWork.hasAborted());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testClosableWillAbortWhenNotCommitted() throws Exception {
|
||||
UnitOfWork unitOfWork;
|
||||
try(UnitOfWork uow = session.begin()) {
|
||||
unitOfWork = uow;
|
||||
Assert.assertFalse(uow.hasAborted());
|
||||
}
|
||||
Assert.assertTrue(unitOfWork.hasAborted());
|
||||
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testClosable() throws Exception {
|
||||
UnitOfWork unitOfWork;
|
||||
try(UnitOfWork uow = session.begin()) {
|
||||
unitOfWork = uow;
|
||||
Assert.assertFalse(uow.hasAborted());
|
||||
uow.commit().andThen(() -> {
|
||||
@Test
|
||||
public void testClosable() throws Exception {
|
||||
UnitOfWork unitOfWork;
|
||||
try (UnitOfWork uow = session.begin()) {
|
||||
unitOfWork = uow;
|
||||
Assert.assertFalse(uow.hasAborted());
|
||||
uow.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
Assert.assertFalse(uow.hasAborted());
|
||||
Assert.assertTrue(uow.hasCommitted());
|
||||
});
|
||||
}
|
||||
Assert.assertFalse(unitOfWork.hasAborted());
|
||||
Assert.assertTrue(unitOfWork.hasCommitted());
|
||||
});
|
||||
}
|
||||
|
||||
Assert.assertFalse(unitOfWork.hasAborted());
|
||||
Assert.assertTrue(unitOfWork.hasCommitted());
|
||||
}
|
||||
}
|
||||
|
|
|
@ -15,16 +15,13 @@
|
|||
*/
|
||||
package net.helenus.test.integration.core.unitofwork;
|
||||
|
||||
import net.helenus.mapping.annotation.*;
|
||||
|
||||
import com.datastax.driver.core.DataType.Name;
|
||||
import java.util.Set;
|
||||
|
||||
import net.helenus.mapping.annotation.*;
|
||||
|
||||
@UDT
|
||||
public interface Directory extends FilesystemNode {
|
||||
|
||||
@Types.Set(Name.TIMEUUID)
|
||||
Set<FilesystemNode> inodes();
|
||||
|
||||
@Types.Set(Name.TIMEUUID)
|
||||
Set<FilesystemNode> inodes();
|
||||
}
|
||||
|
|
|
@ -17,11 +17,9 @@ package net.helenus.test.integration.core.unitofwork;
|
|||
|
||||
import net.helenus.mapping.annotation.*;
|
||||
|
||||
|
||||
@UDT
|
||||
public interface File extends FilesystemNode {
|
||||
|
||||
@Column
|
||||
byte[] data();
|
||||
|
||||
@Column
|
||||
byte[] data();
|
||||
}
|
||||
|
|
|
@ -20,6 +20,5 @@ import net.helenus.mapping.annotation.UDT;
|
|||
@UDT
|
||||
public interface FileAttributes {
|
||||
|
||||
String owner();
|
||||
|
||||
String owner();
|
||||
}
|
||||
|
|
|
@ -15,20 +15,18 @@
|
|||
*/
|
||||
package net.helenus.test.integration.core.unitofwork;
|
||||
|
||||
import net.helenus.mapping.annotation.*;
|
||||
|
||||
import java.util.UUID;
|
||||
import net.helenus.mapping.annotation.*;
|
||||
|
||||
@Table("fs")
|
||||
public interface FilesystemNode {
|
||||
|
||||
@PartitionKey
|
||||
UUID inode();
|
||||
@PartitionKey
|
||||
UUID inode();
|
||||
|
||||
@ClusteringColumn
|
||||
String name();
|
||||
|
||||
@Column
|
||||
FileAttributes attr();
|
||||
@ClusteringColumn
|
||||
String name();
|
||||
|
||||
@Column
|
||||
FileAttributes attr();
|
||||
}
|
||||
|
|
|
@ -15,7 +15,10 @@
|
|||
*/
|
||||
package net.helenus.test.integration.core.unitofwork;
|
||||
|
||||
import static net.helenus.core.Query.eq;
|
||||
|
||||
import com.datastax.driver.core.utils.UUIDs;
|
||||
import java.util.UUID;
|
||||
import net.bytebuddy.utility.RandomString;
|
||||
import net.helenus.core.Helenus;
|
||||
import net.helenus.core.HelenusSession;
|
||||
|
@ -29,167 +32,161 @@ import org.junit.Assert;
|
|||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
|
||||
import java.util.UUID;
|
||||
|
||||
import static net.helenus.core.Query.eq;
|
||||
|
||||
|
||||
@Table
|
||||
@Cacheable
|
||||
interface Widget {
|
||||
@PartitionKey
|
||||
UUID id();
|
||||
@Column
|
||||
String name();
|
||||
}
|
||||
@PartitionKey
|
||||
UUID id();
|
||||
|
||||
@Column
|
||||
String name();
|
||||
}
|
||||
|
||||
public class UnitOfWorkTest extends AbstractEmbeddedCassandraTest {
|
||||
|
||||
static Widget widget;
|
||||
static HelenusSession session;
|
||||
static Widget widget;
|
||||
static HelenusSession session;
|
||||
|
||||
@BeforeClass
|
||||
public static void beforeTest() {
|
||||
session = Helenus.init(getSession()).showCql().add(Widget.class).autoCreateDrop().get();
|
||||
widget = session.dsl(Widget.class);
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
public static void beforeTest() {
|
||||
session = Helenus.init(getSession())
|
||||
.showCql()
|
||||
.add(Widget.class)
|
||||
.autoCreateDrop()
|
||||
.get();
|
||||
widget = session.dsl(Widget.class);
|
||||
}
|
||||
@Test
|
||||
public void testSelectAfterSelect() throws Exception {
|
||||
Widget w1, w2;
|
||||
UUID key = UUIDs.timeBased();
|
||||
|
||||
@Test
|
||||
public void testSelectAfterSelect() throws Exception {
|
||||
Widget w1, w2;
|
||||
UUID key = UUIDs.timeBased();
|
||||
// This should inserted Widget, but not cache it.
|
||||
session
|
||||
.<Widget>insert(widget)
|
||||
.value(widget::id, key)
|
||||
.value(widget::name, RandomString.make(20))
|
||||
.sync();
|
||||
|
||||
// This should inserted Widget, but not cache it.
|
||||
session.<Widget>insert(widget)
|
||||
.value(widget::id, key)
|
||||
.value(widget::name, RandomString.make(20))
|
||||
.sync();
|
||||
try (UnitOfWork uow = session.begin()) {
|
||||
|
||||
try (UnitOfWork uow = session.begin()) {
|
||||
// This should read from the database and return a Widget.
|
||||
w1 =
|
||||
session.<Widget>select(widget).where(widget::id, eq(key)).single().sync(uow).orElse(null);
|
||||
|
||||
// This should read from the database and return a Widget.
|
||||
w1 = session.<Widget>select(widget)
|
||||
.where(widget::id, eq(key))
|
||||
.single()
|
||||
.sync(uow)
|
||||
.orElse(null);
|
||||
|
||||
// This should read from the cache and get the same instance of a Widget.
|
||||
w2 = session.<Widget>select(widget)
|
||||
.where(widget::id, eq(key))
|
||||
.single()
|
||||
.sync(uow)
|
||||
.orElse(null);
|
||||
|
||||
uow.commit()
|
||||
.andThen(() -> {
|
||||
Assert.assertEquals(w1, w2);
|
||||
});
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testSelectAfterNestedSelect() throws Exception {
|
||||
Widget w1, w2, w3, w4;
|
||||
UUID key1 = UUIDs.timeBased();
|
||||
UUID key2 = UUIDs.timeBased();
|
||||
|
||||
// This should inserted Widget, and not cache it in uow1.
|
||||
try (UnitOfWork uow1 = session.begin()) {
|
||||
w1 = session.<Widget>insert(widget)
|
||||
.value(widget::id, key1)
|
||||
.value(widget::name, RandomString.make(20))
|
||||
.sync(uow1);
|
||||
|
||||
try (UnitOfWork uow2 = session.begin(uow1)) {
|
||||
|
||||
// This should read from uow1's cache and return the same Widget.
|
||||
w2 = session.<Widget>select(widget)
|
||||
.where(widget::id, eq(key1))
|
||||
.single()
|
||||
.sync(uow2)
|
||||
.orElse(null);
|
||||
// This should read from the cache and get the same instance of a Widget.
|
||||
w2 =
|
||||
session.<Widget>select(widget).where(widget::id, eq(key)).single().sync(uow).orElse(null);
|
||||
|
||||
uow.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
Assert.assertEquals(w1, w2);
|
||||
|
||||
w3 = session.<Widget>insert(widget)
|
||||
.value(widget::id, key2)
|
||||
.value(widget::name, RandomString.make(20))
|
||||
.sync(uow2);
|
||||
|
||||
uow2.commit()
|
||||
.andThen(() -> {
|
||||
Assert.assertEquals(w1, w2);
|
||||
});
|
||||
}
|
||||
|
||||
// This should read from the cache and get the same instance of a Widget.
|
||||
w4 = session.<Widget>select(widget)
|
||||
.where(widget::id, eq(key2))
|
||||
.single()
|
||||
.sync(uow1)
|
||||
.orElse(null);
|
||||
|
||||
uow1.commit()
|
||||
.andThen(() -> {
|
||||
Assert.assertEquals(w3, w4);
|
||||
});
|
||||
}
|
||||
|
||||
});
|
||||
}
|
||||
/*
|
||||
@Test
|
||||
public void testSelectAfterInsertProperlyCachesEntity() throws Exception {
|
||||
Widget w1, w2, w3, w4;
|
||||
UUID key = UUIDs.timeBased();
|
||||
}
|
||||
|
||||
try (UnitOfWork uow = session.begin()) {
|
||||
@Test
|
||||
public void testSelectAfterNestedSelect() throws Exception {
|
||||
Widget w1, w2, w3, w4;
|
||||
UUID key1 = UUIDs.timeBased();
|
||||
UUID key2 = UUIDs.timeBased();
|
||||
|
||||
// This should cache the inserted Widget.
|
||||
w1 = session.<Widget>insert(widget)
|
||||
.value(widget::id, key)
|
||||
.value(widget::name, RandomString.make(20))
|
||||
.sync(uow);
|
||||
// This should inserted Widget, and not cache it in uow1.
|
||||
try (UnitOfWork uow1 = session.begin()) {
|
||||
w1 =
|
||||
session
|
||||
.<Widget>insert(widget)
|
||||
.value(widget::id, key1)
|
||||
.value(widget::name, RandomString.make(20))
|
||||
.sync(uow1);
|
||||
|
||||
// This should read from the cache and get the same instance of a Widget.
|
||||
w2 = session.<Widget>select(widget)
|
||||
.where(widget::id, eq(key))
|
||||
.single()
|
||||
.sync(uow)
|
||||
.orElse(null);
|
||||
try (UnitOfWork uow2 = session.begin(uow1)) {
|
||||
|
||||
uow.commit()
|
||||
.andThen(() -> {
|
||||
Assert.assertEquals(w1, w2);
|
||||
});
|
||||
}
|
||||
|
||||
// This should read the widget from the session cache and maintain object identity.
|
||||
w3 = session.<Widget>select(widget)
|
||||
.where(widget::id, eq(key))
|
||||
// This should read from uow1's cache and return the same Widget.
|
||||
w2 =
|
||||
session
|
||||
.<Widget>select(widget)
|
||||
.where(widget::id, eq(key1))
|
||||
.single()
|
||||
.sync()
|
||||
.sync(uow2)
|
||||
.orElse(null);
|
||||
|
||||
Assert.assertEquals(w1, w3);
|
||||
Assert.assertEquals(w1, w2);
|
||||
|
||||
// This should read the widget from the database, no object identity but values should match.
|
||||
w4 = session.<Widget>select(widget)
|
||||
.where(widget::id, eq(key))
|
||||
.ignoreCache()
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
w3 =
|
||||
session
|
||||
.<Widget>insert(widget)
|
||||
.value(widget::id, key2)
|
||||
.value(widget::name, RandomString.make(20))
|
||||
.sync(uow2);
|
||||
|
||||
Assert.assertNotEquals(w1, w4);
|
||||
Assert.assertTrue(w1.equals(w4));
|
||||
uow2.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
Assert.assertEquals(w1, w2);
|
||||
});
|
||||
}
|
||||
|
||||
// This should read from the cache and get the same instance of a Widget.
|
||||
w4 =
|
||||
session
|
||||
.<Widget>select(widget)
|
||||
.where(widget::id, eq(key2))
|
||||
.single()
|
||||
.sync(uow1)
|
||||
.orElse(null);
|
||||
|
||||
uow1.commit()
|
||||
.andThen(
|
||||
() -> {
|
||||
Assert.assertEquals(w3, w4);
|
||||
});
|
||||
}
|
||||
*/
|
||||
}
|
||||
/*
|
||||
@Test
|
||||
public void testSelectAfterInsertProperlyCachesEntity() throws Exception {
|
||||
Widget w1, w2, w3, w4;
|
||||
UUID key = UUIDs.timeBased();
|
||||
|
||||
try (UnitOfWork uow = session.begin()) {
|
||||
|
||||
// This should cache the inserted Widget.
|
||||
w1 = session.<Widget>insert(widget)
|
||||
.value(widget::id, key)
|
||||
.value(widget::name, RandomString.make(20))
|
||||
.sync(uow);
|
||||
|
||||
// This should read from the cache and get the same instance of a Widget.
|
||||
w2 = session.<Widget>select(widget)
|
||||
.where(widget::id, eq(key))
|
||||
.single()
|
||||
.sync(uow)
|
||||
.orElse(null);
|
||||
|
||||
uow.commit()
|
||||
.andThen(() -> {
|
||||
Assert.assertEquals(w1, w2);
|
||||
});
|
||||
}
|
||||
|
||||
// This should read the widget from the session cache and maintain object identity.
|
||||
w3 = session.<Widget>select(widget)
|
||||
.where(widget::id, eq(key))
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
|
||||
Assert.assertEquals(w1, w3);
|
||||
|
||||
// This should read the widget from the database, no object identity but values should match.
|
||||
w4 = session.<Widget>select(widget)
|
||||
.where(widget::id, eq(key))
|
||||
.ignoreCache()
|
||||
.single()
|
||||
.sync()
|
||||
.orElse(null);
|
||||
|
||||
Assert.assertNotEquals(w1, w4);
|
||||
Assert.assertTrue(w1.equals(w4));
|
||||
}
|
||||
*/
|
||||
}
|
||||
|
|
|
@ -1,24 +1,29 @@
|
|||
package net.helenus.test.integration.core.views;
|
||||
|
||||
import java.util.Date;
|
||||
import java.util.UUID;
|
||||
import net.helenus.mapping.annotation.ClusteringColumn;
|
||||
import net.helenus.mapping.annotation.CoveringIndex;
|
||||
import net.helenus.mapping.annotation.PartitionKey;
|
||||
import net.helenus.mapping.annotation.Table;
|
||||
|
||||
import java.util.Date;
|
||||
import java.util.UUID;
|
||||
|
||||
@Table
|
||||
@CoveringIndex(name="cyclist_mv",
|
||||
covering={"age", "birthday", "country"},
|
||||
partitionKeys={"age", "cid"},
|
||||
clusteringColumns={})
|
||||
@CoveringIndex(
|
||||
name = "cyclist_mv",
|
||||
covering = {"age", "birthday", "country"},
|
||||
partitionKeys = {"age", "cid"},
|
||||
clusteringColumns = {}
|
||||
)
|
||||
public interface Cyclist {
|
||||
@ClusteringColumn
|
||||
UUID cid();
|
||||
String name();
|
||||
@PartitionKey
|
||||
int age();
|
||||
Date birthday();
|
||||
String country();
|
||||
@ClusteringColumn
|
||||
UUID cid();
|
||||
|
||||
String name();
|
||||
|
||||
@PartitionKey
|
||||
int age();
|
||||
|
||||
Date birthday();
|
||||
|
||||
String country();
|
||||
}
|
||||
|
|
|
@ -1,20 +1,22 @@
|
|||
package net.helenus.test.integration.core.views;
|
||||
|
||||
import java.util.Date;
|
||||
import java.util.UUID;
|
||||
import net.helenus.mapping.annotation.ClusteringColumn;
|
||||
import net.helenus.mapping.annotation.Index;
|
||||
import net.helenus.mapping.annotation.MaterializedView;
|
||||
import net.helenus.mapping.annotation.PartitionKey;
|
||||
|
||||
import java.util.Date;
|
||||
import java.util.UUID;
|
||||
|
||||
@MaterializedView
|
||||
public interface CyclistsByAge extends Cyclist {
|
||||
@PartitionKey
|
||||
UUID cid();
|
||||
@ClusteringColumn
|
||||
int age();
|
||||
Date birthday();
|
||||
@Index
|
||||
String country();
|
||||
@PartitionKey
|
||||
UUID cid();
|
||||
|
||||
@ClusteringColumn
|
||||
int age();
|
||||
|
||||
Date birthday();
|
||||
|
||||
@Index
|
||||
String country();
|
||||
}
|
||||
|
|
|
@ -15,66 +15,62 @@
|
|||
*/
|
||||
package net.helenus.test.integration.core.views;
|
||||
|
||||
import static net.helenus.core.Query.eq;
|
||||
|
||||
import java.text.ParseException;
|
||||
import java.text.SimpleDateFormat;
|
||||
import java.util.*;
|
||||
|
||||
import net.helenus.test.integration.core.views.Cyclist;
|
||||
import net.helenus.test.integration.core.views.CyclistsByAge;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
|
||||
import net.helenus.core.Helenus;
|
||||
import net.helenus.core.HelenusSession;
|
||||
import net.helenus.test.integration.build.AbstractEmbeddedCassandraTest;
|
||||
|
||||
import static net.helenus.core.Query.eq;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
|
||||
// See: https://docs.datastax.com/en/cql/3.3/cql/cql_using/useCreateMV.html
|
||||
// https://docs.datastax.com/en/cql/3.3/cql/cql_reference/cqlCreateMaterializedView.html
|
||||
// https://www.datastax.com/dev/blog/materialized-view-performance-in-cassandra-3-x
|
||||
public class MaterializedViewTest extends AbstractEmbeddedCassandraTest {
|
||||
|
||||
static Cyclist cyclist;
|
||||
static HelenusSession session;
|
||||
|
||||
static Cyclist cyclist;
|
||||
static HelenusSession session;
|
||||
|
||||
static Date dateFromString(String dateInString) {
|
||||
SimpleDateFormat formatter = new SimpleDateFormat("dd-MMM-yyyy");
|
||||
try {
|
||||
return formatter.parse(dateInString);
|
||||
} catch (ParseException e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
return null;
|
||||
static Date dateFromString(String dateInString) {
|
||||
SimpleDateFormat formatter = new SimpleDateFormat("dd-MMM-yyyy");
|
||||
try {
|
||||
return formatter.parse(dateInString);
|
||||
} catch (ParseException e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
public static void beforeTest() {
|
||||
session = Helenus.init(getSession())
|
||||
.showCql()
|
||||
.add(Cyclist.class)
|
||||
.add(CyclistsByAge.class)
|
||||
.autoCreateDrop()
|
||||
.get();
|
||||
cyclist = session.dsl(Cyclist.class);
|
||||
@BeforeClass
|
||||
public static void beforeTest() {
|
||||
session =
|
||||
Helenus.init(getSession())
|
||||
.showCql()
|
||||
.add(Cyclist.class)
|
||||
.add(CyclistsByAge.class)
|
||||
.autoCreateDrop()
|
||||
.get();
|
||||
cyclist = session.dsl(Cyclist.class);
|
||||
|
||||
session
|
||||
.insert(cyclist)
|
||||
.value(cyclist::cid, UUID.randomUUID())
|
||||
.value(cyclist::age, 18)
|
||||
.value(cyclist::birthday, dateFromString("1997-02-08"))
|
||||
.value(cyclist::country, "Netherlands")
|
||||
.value(cyclist::name, "Pascal EENKHOORN")
|
||||
.sync();
|
||||
}
|
||||
session
|
||||
.insert(cyclist)
|
||||
.value(cyclist::cid, UUID.randomUUID())
|
||||
.value(cyclist::age, 18)
|
||||
.value(cyclist::birthday, dateFromString("1997-02-08"))
|
||||
.value(cyclist::country, "Netherlands")
|
||||
.value(cyclist::name, "Pascal EENKHOORN")
|
||||
.sync();
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testMv() throws Exception {
|
||||
session
|
||||
.select(Cyclist.class)
|
||||
.<CyclistsByAge>from(CyclistsByAge.class)
|
||||
.where(cyclist::age, eq(18))
|
||||
.sync();
|
||||
|
||||
}
|
||||
@Test
|
||||
public void testMv() throws Exception {
|
||||
session
|
||||
.select(Cyclist.class)
|
||||
.<CyclistsByAge>from(CyclistsByAge.class)
|
||||
.where(cyclist::age, eq(18))
|
||||
.sync();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,7 +19,6 @@ import java.util.Date;
|
|||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
import net.helenus.core.reflect.Drafted;
|
||||
import net.helenus.mapping.HelenusEntity;
|
||||
import net.helenus.mapping.annotation.*;
|
||||
|
||||
@Table
|
||||
|
|
Loading…
Reference in a new issue