Skip to content

Commit

Permalink
Reverted unfortunate word change
Browse files Browse the repository at this point in the history
  • Loading branch information
systay committed Sep 6, 2016
1 parent f4f26ab commit 78e5bcb
Show file tree
Hide file tree
Showing 60 changed files with 104 additions and 104 deletions.
Expand Up @@ -271,7 +271,7 @@ else if ( statement.equalsIgnoreCase( ROLLBACK ) )
else if( spi.isPeriodicCommit( statement ) ) else if( spi.isPeriodicCommit( statement ) )
{ {
throw new QueryExecutionKernelException( new InvalidSemanticsException( throw new QueryExecutionKernelException( new InvalidSemanticsException(
"Executing stream that use periodic commit in an " + "Executing queries that use periodic commit in an " +
"open transaction is not possible." ) ); "open transaction is not possible." ) );
} }
else else
Expand Down
Expand Up @@ -518,7 +518,7 @@ public void shouldNotSupportUsingPeriodicCommitInTransaction() throws Exception


// Then // Then
assertThat( recorder.nextResponse(), failedWithStatus( Status.Statement.SemanticError ) ); assertThat( recorder.nextResponse(), failedWithStatus( Status.Statement.SemanticError ) );
// "Executing stream that use periodic commit in an open transaction is not possible." // "Executing queries that use periodic commit in an open transaction is not possible."
} }


@Test @Test
Expand Down
Expand Up @@ -422,7 +422,7 @@ class LoadCsvAcceptanceTest
result.toList should equal(List(Map("field" -> "something"))) result.toList should equal(List(Map("field" -> "something")))
} }


test("eager stream should be handled correctly") { test("eager queries should be handled correctly") {
val urls = csvUrls({ val urls = csvUrls({
writer => writer =>
writer.println("id,title,country,year") writer.println("id,title,country,year")
Expand Down
Expand Up @@ -340,7 +340,7 @@ return p""")


// Not TCK material -- indexes // Not TCK material -- indexes


test("should handle stream that cant be index solved because expressions lack dependencies") { test("should handle queries that cant be index solved because expressions lack dependencies") {
// given // given
val a = createLabeledNode(Map("property" -> 42), "Label") val a = createLabeledNode(Map("property" -> 42), "Label")
val b = createLabeledNode(Map("property" -> 42), "Label") val b = createLabeledNode(Map("property" -> 42), "Label")
Expand All @@ -358,7 +358,7 @@ return p""")
result.toList should equal (List(Map("a" -> a, "b" -> b))) result.toList should equal (List(Map("a" -> a, "b" -> b)))
} }


test("should handle stream that cant be index solved because expressions lack dependencies with two disjoin patterns") { test("should handle queries that cant be index solved because expressions lack dependencies with two disjoin patterns") {
// given // given
val a = createLabeledNode(Map("property" -> 42), "Label") val a = createLabeledNode(Map("property" -> 42), "Label")
val b = createLabeledNode(Map("property" -> 42), "Label") val b = createLabeledNode(Map("property" -> 42), "Label")
Expand All @@ -378,7 +378,7 @@ return p""")
)) ))
} }


test("should use the index for property existence stream (with exists) for cost when asked for it") { test("should use the index for property existence queries (with exists) for cost when asked for it") {
// given // given
val n = createLabeledNode(Map("email" -> "me@mine"), "User") val n = createLabeledNode(Map("email" -> "me@mine"), "User")
val m = createLabeledNode(Map("email" -> "you@yours"), "User") val m = createLabeledNode(Map("email" -> "you@yours"), "User")
Expand All @@ -393,7 +393,7 @@ return p""")
result.executionPlanDescription().toString should include("NodeIndexScan") result.executionPlanDescription().toString should include("NodeIndexScan")
} }


test("should use the index for property existence stream (with IS NOT NULL) for cost when asked for it") { test("should use the index for property existence queries (with IS NOT NULL) for cost when asked for it") {
// given // given
val n = createLabeledNode(Map("email" -> "me@mine"), "User") val n = createLabeledNode(Map("email" -> "me@mine"), "User")
val m = createLabeledNode(Map("email" -> "you@yours"), "User") val m = createLabeledNode(Map("email" -> "you@yours"), "User")
Expand All @@ -420,7 +420,7 @@ return p""")
Seq(n, m, p) Seq(n, m, p)
} }


test("should use the index for property existence stream when cardinality prefers it") { test("should use the index for property existence queries when cardinality prefers it") {
// given // given
val nodes = setupIndexScanTest() val nodes = setupIndexScanTest()


Expand All @@ -432,7 +432,7 @@ return p""")
result.executionPlanDescription().toString should include("NodeIndexScan") result.executionPlanDescription().toString should include("NodeIndexScan")
} }


test("should not use the index for property existence stream when cardinality does not prefer it") { test("should not use the index for property existence queries when cardinality does not prefer it") {
// given // given
val nodes = setupIndexScanTest() val nodes = setupIndexScanTest()


Expand All @@ -444,7 +444,7 @@ return p""")
result.executionPlanDescription().toString should include("NodeByLabelScan") result.executionPlanDescription().toString should include("NodeByLabelScan")
} }


test("should not use the index for property existence stream when property value predicate exists") { test("should not use the index for property existence queries when property value predicate exists") {
// given // given
val nodes = setupIndexScanTest() val nodes = setupIndexScanTest()


Expand All @@ -457,7 +457,7 @@ return p""")
result.executionPlanDescription().toString should not include "NodeIndexScan" result.executionPlanDescription().toString should not include "NodeIndexScan"
} }


test("should use the index for property existence stream for rule when asked for it") { test("should use the index for property existence queries for rule when asked for it") {
// given // given
val n = createLabeledNode(Map("email" -> "me@mine"), "User") val n = createLabeledNode(Map("email" -> "me@mine"), "User")
val m = createLabeledNode(Map("email" -> "you@yours"), "User") val m = createLabeledNode(Map("email" -> "you@yours"), "User")
Expand All @@ -472,7 +472,7 @@ return p""")
result.executionPlanDescription().toString should include("SchemaIndex") result.executionPlanDescription().toString should include("SchemaIndex")
} }


test("should not use the index for property existence stream for rule when not asking for it") { test("should not use the index for property existence queries for rule when not asking for it") {
// given // given
val n = createLabeledNode(Map("email" -> "me@mine"), "User") val n = createLabeledNode(Map("email" -> "me@mine"), "User")
val m = createLabeledNode(Map("email" -> "you@yours"), "User") val m = createLabeledNode(Map("email" -> "you@yours"), "User")
Expand Down
Expand Up @@ -354,7 +354,7 @@ class SemanticErrorAcceptanceTest extends ExecutionEngineFunSuite {
) )
} }


test("aggregation inside looping stream is not allowed") { test("aggregation inside looping queries is not allowed") {


val mess = "Can't use aggregating expressions inside of expressions executing over lists" val mess = "Can't use aggregating expressions inside of expressions executing over lists"
executeAndEnsureError( executeAndEnsureError(
Expand Down
Expand Up @@ -118,7 +118,7 @@ class SyntaxExceptionAcceptanceTest extends ExecutionEngineFunSuite {
) )
} }


test("handles multiline stream") { test("handles multiline queries") {
test( test(
"""start """start
a=node(0), a=node(0),
Expand Down
Expand Up @@ -104,7 +104,7 @@ class UniqueIndexAcceptanceTest extends ExecutionEngineFunSuite with NewPlannerT
result shouldNot use("NodeUniqueIndexSeek(Locking)") result shouldNot use("NodeUniqueIndexSeek(Locking)")
} }


test("should use locking unique index for merge stream") { test("should use locking unique index for merge queries") {
//GIVEN //GIVEN
createLabeledNode(Map("name" -> "Andres"), "Person") createLabeledNode(Map("name" -> "Andres"), "Person")
graph.createConstraint("Person", "name") graph.createConstraint("Person", "name")
Expand All @@ -117,7 +117,7 @@ class UniqueIndexAcceptanceTest extends ExecutionEngineFunSuite with NewPlannerT
result should use("NodeUniqueIndexSeek(Locking)") result should use("NodeUniqueIndexSeek(Locking)")
} }


test("should use locking unique index for mixed read write stream") { test("should use locking unique index for mixed read write queries") {
//GIVEN //GIVEN
createLabeledNode(Map("name" -> "Andres"), "Person") createLabeledNode(Map("name" -> "Andres"), "Person")
graph.createConstraint("Person", "name") graph.createConstraint("Person", "name")
Expand Down
Expand Up @@ -431,7 +431,7 @@ class UsingAcceptanceTest extends ExecutionEngineFunSuite with NewPlannerTestSup
result.executionPlanDescription() should includeOnlyOneHashJoinOn("c") result.executionPlanDescription() should includeOnlyOneHashJoinOn("c")
} }


test(s"$plannerName should be able to use join hints for stream with var length pattern") { test(s"$plannerName should be able to use join hints for queries with var length pattern") {
val a = createLabeledNode(Map("prop" -> "foo"), "Foo") val a = createLabeledNode(Map("prop" -> "foo"), "Foo")
val b = createNode() val b = createNode()
val c = createNode() val c = createNode()
Expand Down
Expand Up @@ -30,7 +30,7 @@ import org.neo4j.cypher.internal.frontend.v3_1.{InputPosition, Rewriter, Scope,
// A prepared query captures all information that has been derived so far as part // A prepared query captures all information that has been derived so far as part
// of this processing // of this processing
// //
// Currently there are two types of prepared stream: // Currently there are two types of prepared queries:
// //
// - PreparedQuerySyntax (can be constructed without db access) // - PreparedQuerySyntax (can be constructed without db access)
// - PreparedQuerySemantics (construction may requires db access for resolving procedure signatures) // - PreparedQuerySemantics (construction may requires db access for resolving procedure signatures)
Expand Down
Expand Up @@ -175,7 +175,7 @@ final case class RangeLessThan[+V](bounds: Bounds[V]) extends HalfOpenSeekRange[
PrefixRange is used to describe intervals on string values for prefix search. PrefixRange is used to describe intervals on string values for prefix search.
This is practical for two reasons: This is practical for two reasons:
- It directly maps on prefix stream of index implementations - It directly maps on prefix queries of index implementations
- It removes the need to construct a proper upper bound value for an interval that - It removes the need to construct a proper upper bound value for an interval that
would describe the prefix search (which can be difficult due to unicode issues) would describe the prefix search (which can be difficult due to unicode issues)
*/ */
Expand Down
Expand Up @@ -359,8 +359,8 @@ object ClauseConverters {


/* /*
When encountering a WITH that is not an event horizon, and we have no optional matches in the current QueryGraph, When encountering a WITH that is not an event horizon, and we have no optional matches in the current QueryGraph,
we simply continue building on the current PlannerQuery. Our ASTRewriters rewrite stream in such a way that we simply continue building on the current PlannerQuery. Our ASTRewriters rewrite queries in such a way that
a lot of stream have these WITH clauses. a lot of queries have these WITH clauses.
Handles: ... WITH * [WHERE <predicate>] ... Handles: ... WITH * [WHERE <predicate>] ...
*/ */
Expand Down Expand Up @@ -467,7 +467,7 @@ object ClauseConverters {
val nonHints = items.collect { case Left(item) => item } val nonHints = items.collect { case Left(item) => item }


if (nonHints.nonEmpty) { if (nonHints.nonEmpty) {
// all other start stream is delegated to legacy planner // all other start queries is delegated to legacy planner
throw new CantHandleQueryException() throw new CantHandleQueryException()
} }


Expand Down
Expand Up @@ -63,7 +63,7 @@ object StatementConverters {
case _: UnionDistinct => true case _: UnionDistinct => true
} }
val plannedQueries: Seq[PlannerQueryBuilder] = queries.reverseMap(x => toPlannerQueryBuilder(x, semanticTable)) val plannedQueries: Seq[PlannerQueryBuilder] = queries.reverseMap(x => toPlannerQueryBuilder(x, semanticTable))
//UNION requires all stream to return the same variables //UNION requires all queries to return the same variables
assert(plannedQueries.nonEmpty) assert(plannedQueries.nonEmpty)
val returns = plannedQueries.head.returns val returns = plannedQueries.head.returns
assert(plannedQueries.forall(_.returns == returns)) assert(plannedQueries.forall(_.returns == returns))
Expand Down
Expand Up @@ -190,7 +190,7 @@ case class PartiallySolvedQuery(returns: Seq[QueryToken[ReturnColumn]],


def containsAggregation: Boolean = aggregation.nonEmpty || tail.exists(_.containsAggregation) def containsAggregation: Boolean = aggregation.nonEmpty || tail.exists(_.containsAggregation)


/* This methods is used to rewrite the stream from the end of the query line to the beginning of it */ /* This methods is used to rewrite the queries from the end of the query line to the beginning of it */
def rewriteFromTheTail(f: PartiallySolvedQuery => PartiallySolvedQuery): PartiallySolvedQuery = def rewriteFromTheTail(f: PartiallySolvedQuery => PartiallySolvedQuery): PartiallySolvedQuery =
f(copy(tail = tail.map(_.rewriteFromTheTail(f)))) f(copy(tail = tail.map(_.rewriteFromTheTail(f))))


Expand Down
Expand Up @@ -27,7 +27,7 @@ import org.neo4j.cypher.internal.compiler.v3_1.symbols.SymbolTable


/** /**
* This class should get rid of any extra columns built up while building the execution plan, that weren't in the * This class should get rid of any extra columns built up while building the execution plan, that weren't in the
* stream return clause. * queries return clause.
*/ */
class ColumnFilterBuilder extends PlanBuilder { class ColumnFilterBuilder extends PlanBuilder {
def apply(plan: ExecutionPlanInProgress, ctx: PlanContext)(implicit pipeMonitor: PipeMonitor) = { def apply(plan: ExecutionPlanInProgress, ctx: PlanContext)(implicit pipeMonitor: PipeMonitor) = {
Expand Down
Expand Up @@ -47,7 +47,7 @@ class DistinctBuilder extends PlanBuilder {
plan.query.returns.flatMap(_.token.expressions(plan.pipe.symbols)).toMap plan.query.returns.flatMap(_.token.expressions(plan.pipe.symbols)).toMap


def canWorkWith(plan: ExecutionPlanInProgress, ctx: PlanContext)(implicit pipeMonitor: PipeMonitor) = { def canWorkWith(plan: ExecutionPlanInProgress, ctx: PlanContext)(implicit pipeMonitor: PipeMonitor) = {
plan.query.aggregateToDo && //The parser marks DISTINCT stream as aggregates. Revisit? plan.query.aggregateToDo && //The parser marks DISTINCT queries as aggregates. Revisit?
plan.query.aggregation.isEmpty && //It's an aggregate query without aggregate expressions plan.query.aggregation.isEmpty && //It's an aggregate query without aggregate expressions
plan.query.readyToAggregate && plan.query.readyToAggregate &&
plan.query.returns.exists { plan.query.returns.exists {
Expand Down
Expand Up @@ -26,7 +26,7 @@ import org.neo4j.cypher.internal.compiler.v3_1.pipes.PipeMonitor
import org.neo4j.cypher.internal.compiler.v3_1.spi.PlanContext import org.neo4j.cypher.internal.compiler.v3_1.spi.PlanContext


/* /*
This builder is concerned with finding stream without start items and without index hints, and This builder is concerned with finding queries without start items and without index hints, and
choosing a start point to use. choosing a start point to use.
To do this, three things are done. To do this, three things are done.
Expand Down
Expand Up @@ -28,7 +28,7 @@ import org.neo4j.cypher.internal.frontend.v3_1.ast._
import org.neo4j.cypher.internal.frontend.v3_1.symbols.TypeSpec import org.neo4j.cypher.internal.frontend.v3_1.symbols.TypeSpec


/** /**
* This planner takes on stream that requires no planning such as procedures and schema commands * This planner takes on queries that requires no planning such as procedures and schema commands
* *
* @param delegate The plan builder to delegate to * @param delegate The plan builder to delegate to
*/ */
Expand Down
Expand Up @@ -63,7 +63,7 @@ case class CreateNode(key: String, properties: Map[String, Expression], labels:




/* /*
Parameters coming in from the outside in stream using parameters like this: Parameters coming in from the outside in queries using parameters like this:
CREATE (n {param}) CREATE (n {param})
Expand Down
Expand Up @@ -96,7 +96,7 @@ case class CostBasedExecutablePlanBuilder(monitors: Monitors,


val pipeBuildContext = PipeExecutionBuilderContext(metrics.cardinality, semanticTable, plannerName) val pipeBuildContext = PipeExecutionBuilderContext(metrics.cardinality, semanticTable, plannerName)


//Check for unresolved tokens for read-only stream //Check for unresolved tokens for read-only queries
if (plan.solved.all(_.queryGraph.readOnly)) checkForUnresolvedTokens(ast, semanticTable).foreach(notificationLogger += _) if (plan.solved.all(_.queryGraph.readOnly)) checkForUnresolvedTokens(ast, semanticTable).foreach(notificationLogger += _)


(periodicCommit, plan, pipeBuildContext) (periodicCommit, plan, pipeBuildContext)
Expand Down
Expand Up @@ -25,7 +25,7 @@ import org.neo4j.cypher.internal.compiler.v3_1.planner.{CardinalityEstimation, P
RollUp is the inverse of the Unwind operator. For each row passed in from the LHS, the whole RHS is executed. RollUp is the inverse of the Unwind operator. For each row passed in from the LHS, the whole RHS is executed.
For each row produced by the RHS, a single column value is extracted and inserted into a collection. For each row produced by the RHS, a single column value is extracted and inserted into a collection.
It is used for sub stream that return collections, such as pattern expressions (returns a collection of paths) and It is used for sub queries that return collections, such as pattern expressions (returns a collection of paths) and
pattern comprehension. pattern comprehension.
Note about nullableIdentifiers: when any of these identifiers is null, the collection should be null. Note about nullableIdentifiers: when any of these identifiers is null, the collection should be null.
Expand Down
Expand Up @@ -43,7 +43,7 @@ object rewriteProcedureCalls {
resolveCalls andThen fakeStandaloneCallDeclarations resolveCalls andThen fakeStandaloneCallDeclarations
} }


// Current procedure calling syntax allows simplified short-hand syntax for stream // Current procedure calling syntax allows simplified short-hand syntax for queries
// that only consist of a standalone procedure call. In all other cases attempts to // that only consist of a standalone procedure call. In all other cases attempts to
// use the simplified syntax lead to errors during semantic checking. // use the simplified syntax lead to errors during semantic checking.
// //
Expand Down
Expand Up @@ -290,7 +290,7 @@ class InlineProjectionsTest extends CypherFunSuite with AstRewritingTestSupport
} }


// FIXME: 2014-4-30 Stefan: This is not yet supported by the inline rewriter // FIXME: 2014-4-30 Stefan: This is not yet supported by the inline rewriter
test("should refuse to inline stream containing update clauses by throwing CantHandleQueryException") { test("should refuse to inline queries containing update clauses by throwing CantHandleQueryException") {
evaluating { evaluating {
projectionInlinedAst( projectionInlinedAst(
"""CREATE (n) """CREATE (n)
Expand Down
Expand Up @@ -93,7 +93,7 @@ class LiteralReplacementTest extends CypherFunSuite {
) )
} }


test("should not rewrite stream that already have params in them") { test("should not rewrite queries that already have params in them") {
assertRewrite( assertRewrite(
"CREATE (a:Person {name:'Jakub', age:{age} })", "CREATE (a:Person {name:'Jakub', age:{age} })",
"CREATE (a:Person {name:'Jakub', age:{age} })", "CREATE (a:Person {name:'Jakub', age:{age} })",
Expand Down
Expand Up @@ -30,7 +30,7 @@ import org.neo4j.cypher.internal.frontend.v3_1.test_helpers.CypherFunSuite
class ExecutionWorkflowBuilderTest extends CypherFunSuite { class ExecutionWorkflowBuilderTest extends CypherFunSuite {
val PlannerName = IDPPlannerName val PlannerName = IDPPlannerName


test("produces eager results for updating stream") { test("produces eager results for updating queries") {
// GIVEN // GIVEN
val pipe = mock[Pipe] val pipe = mock[Pipe]
when(pipe.createResults(any())).thenReturn(Iterator.empty) when(pipe.createResults(any())).thenReturn(Iterator.empty)
Expand All @@ -48,7 +48,7 @@ class ExecutionWorkflowBuilderTest extends CypherFunSuite {
result.asInstanceOf[PipeExecutionResult].result shouldBe a[EagerResultIterator] result.asInstanceOf[PipeExecutionResult].result shouldBe a[EagerResultIterator]
} }


test("produces lazy results for non-updating stream") { test("produces lazy results for non-updating queries") {
// GIVEN // GIVEN
val pipe = mock[Pipe] val pipe = mock[Pipe]
when(pipe.createResults(any())).thenReturn(Iterator.empty) when(pipe.createResults(any())).thenReturn(Iterator.empty)
Expand All @@ -65,7 +65,7 @@ class ExecutionWorkflowBuilderTest extends CypherFunSuite {
result.asInstanceOf[PipeExecutionResult].result should not be an[EagerResultIterator] result.asInstanceOf[PipeExecutionResult].result should not be an[EagerResultIterator]
} }


test("produces explain results for EXPLAIN stream") { test("produces explain results for EXPLAIN queries") {
// GIVEN // GIVEN
val pipe = mock[Pipe] val pipe = mock[Pipe]
when(pipe.createResults(any())).thenReturn(Iterator.empty) when(pipe.createResults(any())).thenReturn(Iterator.empty)
Expand Down
Expand Up @@ -119,7 +119,7 @@ class NodeFetchStrategyTest extends CypherFunSuite {
inner.run() inner.run()
} }


test("should select schema index for range stream") { test("should select schema index for range queries") {
object inner extends AstConstructionTestSupport { object inner extends AstConstructionTestSupport {


def run(): Unit = { def run(): Unit = {
Expand Down
Expand Up @@ -293,7 +293,7 @@ class StartPointChoosingBuilderTest extends BuilderTest {
inner.run() inner.run()
} }


test("should pick any index available for range stream") { test("should pick any index available for range queries") {
object inner extends AstConstructionTestSupport { object inner extends AstConstructionTestSupport {


def run() = { def run() = {
Expand Down
Expand Up @@ -29,7 +29,7 @@ class AssumeIndependenceQueryGraphCardinalityModelTest extends RandomizedCardina
import ABCD._ import ABCD._
import org.scalatest.prop.TableDrivenPropertyChecks._ import org.scalatest.prop.TableDrivenPropertyChecks._


test("all stream") { test("all queries") {
val queries = Table.apply[String, Double]( val queries = Table.apply[String, Double](
("query", "expected cardinality"), ("query", "expected cardinality"),
"MATCH (n)" "MATCH (n)"
Expand Down
Expand Up @@ -47,7 +47,7 @@ class GraphStatisticsSnapshotTest extends CypherFunSuite {
} }
} }


test("records stream and its observed values") { test("records queries and its observed values") {
val snapshot = MutableGraphStatisticsSnapshot() val snapshot = MutableGraphStatisticsSnapshot()
val instrumentedStatistics = InstrumentedGraphStatistics(graphStatistics(), snapshot) val instrumentedStatistics = InstrumentedGraphStatistics(graphStatistics(), snapshot)
instrumentedStatistics.nodesWithLabelCardinality(None) instrumentedStatistics.nodesWithLabelCardinality(None)
Expand Down
Expand Up @@ -118,7 +118,7 @@ class ExecutionEngine(val queryService: GraphDatabaseQueryService, logProvider:
private def parsePreParsedQuery(preParsedQuery: PreParsedQuery, tracer: CompilationPhaseTracer): ParsedQuery = { private def parsePreParsedQuery(preParsedQuery: PreParsedQuery, tracer: CompilationPhaseTracer): ParsedQuery = {
parsedQueries.get(preParsedQuery.statementWithVersionAndPlanner).getOrElse { parsedQueries.get(preParsedQuery.statementWithVersionAndPlanner).getOrElse {
val parsedQuery = compiler.parseQuery(preParsedQuery, tracer) val parsedQuery = compiler.parseQuery(preParsedQuery, tracer)
//don't cache failed stream //don't cache failed queries
if (!parsedQuery.hasErrors) parsedQueries.put(preParsedQuery.statementWithVersionAndPlanner, parsedQuery) if (!parsedQuery.hasErrors) parsedQueries.put(preParsedQuery.statementWithVersionAndPlanner, parsedQuery)
parsedQuery parsedQuery
} }
Expand Down
Expand Up @@ -42,7 +42,7 @@ import scala.collection.immutable.Map


class ExecutionEngineIT extends CypherFunSuite with GraphIcing { class ExecutionEngineIT extends CypherFunSuite with GraphIcing {


test("by default when using cypher 2.3 some stream should default to COST") { test("by default when using cypher 2.3 some queries should default to COST") {
//given //given
val db = new TestGraphDatabaseFactory() val db = new TestGraphDatabaseFactory()
.newImpermanentDatabaseBuilder() .newImpermanentDatabaseBuilder()
Expand All @@ -60,7 +60,7 @@ class ExecutionEngineIT extends CypherFunSuite with GraphIcing {
plan2.getArguments.get("planner-impl") should equal(CostBasedPlannerName.default.name) plan2.getArguments.get("planner-impl") should equal(CostBasedPlannerName.default.name)
} }


test("by default when using cypher 3.0 some stream should default to COST") { test("by default when using cypher 3.0 some queries should default to COST") {
//given //given
val db = new TestGraphDatabaseFactory() val db = new TestGraphDatabaseFactory()
.newImpermanentDatabaseBuilder() .newImpermanentDatabaseBuilder()
Expand All @@ -78,7 +78,7 @@ class ExecutionEngineIT extends CypherFunSuite with GraphIcing {
plan2.getArguments.get("planner-impl") should equal(CostBasedPlannerName.default.name) plan2.getArguments.get("planner-impl") should equal(CostBasedPlannerName.default.name)
} }


test("by default when using cypher 3.1 some stream should default to COST") { test("by default when using cypher 3.1 some queries should default to COST") {
//given //given
val db = new TestGraphDatabaseFactory() val db = new TestGraphDatabaseFactory()
.newImpermanentDatabaseBuilder() .newImpermanentDatabaseBuilder()
Expand Down

0 comments on commit 78e5bcb

Please sign in to comment.