JVM Library to translate GraphQL queries and mutations to Neo4j’s Cypher
How does it work
This library
-
parses a GraphQL schema and
-
uses the information of the annotated schema to translate GraphQL queries and parameters into Cypher queries and parameters.
Those Cypher queries can then executed, e.g via the Neo4j-Java-Driver (or other JVM drivers) against the graph database and the results can be returned directly to the caller.
The request, result and error handling is not part of this library, but we provide demo programs on how to use it in different languages.
Note
|
All the supported features are listed and explained below, more detailed docs will be added in time. |
Examples
For complex examples take a look at our example projects
API compatibility to @neo4j/graphql
Since the javascript pendant of this library (neo4j-graphql-js) has majored into a neo4j product, we want to migrate our augmented schema, to match as much as possible to the one of the @neo4j/graphql
.
Therefore, we created a list of issues to track progress.
We will try to make the migration as smooth as possible. For this purpose we will support the old, and the new way of schema augmentation until the next major release.
To already test the new features, you can enable them via some setting in the SchemaConfig
FAQ
How does this relate to the other neo4j graphql libraries?
The GRANDstack is a full-stack package that integrates React frontends via GraphQL through neo4j-graphql-js
with Neo4j.
Similar to neo4j-graphql-js
this library focuses on query translation, just for the JVM instead of Node.js.
It does not provide a server (except as examples) or other facilities but is meant to be used as a dependency included for a single purpose.
We plan to replace the code in the current Neo4j server plugin neo4j-graphql
with a single call to this library.
The server plugin could still exist as an example that shows how to handle request-response and error-handling, and perhaps some minimal schema management but be slimmed down to a tiny piece of code.
How does this related to graphql-java
This library uses graphql-java
under the hood for parsing of schema and queries, and managing the GraphQL state and context.
But not for nested field resolvers or data fetching.
If you wanted, you could combine graphql-java
resolvers with this library to have a part of your schema handled by Neo4j.
Thanks a lot to the maintainers of graphql-java
for the awesome library.
Note
|
We also use neo4j-opencypher-dsl provided graciously by the spring-data-neo4j-rx project to generate our cypher queries.
|
Usage
You can use the library as dependency: org.neo4j:neo4j-graphql-java:1.4.0
in any JVM program.
The basic usage should be:
val schema =
"""
type Person {
name: ID!
age: Int
}
# Optional if you use generated queries
type Query {
person : [Person]
personByName(name:ID) : Person
}"""
val query = """ { p:personByName(name:"Joe") { age } } """
val schema = SchemaBuilder.buildSchema(idl)
val ctx = QueryContext()
val (cypher, params) = Translator(schema).translate(query, params, ctx)
// generated Cypher
cypher == "MATCH (p:Person) WHERE p.name = $pName RETURN p {.age} as p"
You find more usage examples in the:
Demo
Here is a minimalistic example in Groovy using the Neo4j-Java driver and Spark-Java as webserver.
It is running against a Neo4j instance at bolt://localhost
(username: neo4j
, password: s3cr3t
) containing the :play movies
graph.
(You can also use a Kotlin based server example.)
In case you wand to bind the neo4j driver directly to the graphql schema you can use the DataFetchingInterceptor to intercept the cypher queries.
// Simplistic GraphQL Server using SparkJava
@Grapes([
@Grab('com.sparkjava:spark-core:2.7.2'),
@Grab('org.neo4j.driver:neo4j-java-driver:1.7.2'),
@Grab('com.google.code.gson:gson:2.8.5'),
@Grab('org.neo4j:neo4j-graphql-java:1.4.0')
])
import spark.*
import static spark.Spark.*
import com.google.gson.Gson
import org.neo4j.graphql.*
import org.neo4j.driver.v1.*
schema = """
type Person {
name: ID!
born: Int
actedIn: [Movie] @relation(name:"ACTED_IN")
}
type Movie {
title: ID!
released: Int
tagline: String
}
type Query {
person : [Person]
}
"""
gson = new Gson()
render = (ResponseTransformer)gson.&toJson
def query(value) { gson.fromJson(value,Map.class)["query"] }
graphql = new Translator(SchemaBuilder.buildSchema(schema))
def translate(query) { graphql.translate(query) }
driver = GraphDatabase.driver("bolt://localhost",AuthTokens.basic("neo4j","s3cr3t"))
def run(cypher) { driver.session().withCloseable {
it.run(cypher.query, Values.value(cypher.params)).list{ it.asMap() }}}
post("/graphql","application/json", { req, res -> run(translate(query(req.body())).first()) }, render);
Run the example with:
groovy docs/Server.groovy
and use http://localhost:4567/graphql as your GraphQL URL.
It uses a schema of:
type Person {
name: ID!
born: Int
actedIn: [Movie] @relation(name:"ACTED_IN")
}
type Movie {
title: ID!
released: Int
tagline: String
}
type Query {
person : [Person]
}
And can run queries like:
{
person(first:3) {
name
born
actedIn(first:2) {
title
}
}
}
You can also test it with curl
curl -XPOST http://localhost:4567/graphql -d'{"query":"{person {name}}"}'
This example doesn’t handle introspection queries, but the one in the test directory does.
Advanced Queries
{ person(filter: {name_starts_with: "L"}, orderBy: "born_asc", first: 5, offset: 2) { name born actedIn(first: 1) { title } } }
{ person(filter: {name_starts_with: "J", born_gte: 1970}, first:2) { name born actedIn(first:1) { title released } } }
Features
Current
-
parse SDL schema
-
resolve query fields via result types
-
handle arguments as equality comparisons for top level and nested fields
-
handle relationships via @relation directive on schema fields
-
@relation directive on types for rich relationships (from, to fields for start & end node)
-
handle first, offset arguments
-
argument types: string, int, float, array
-
request parameter support
-
parametrization for cypher query
-
aliases
-
inline and named fragments
-
auto-generate query fields for all objects
-
@cypher directive for fields to compute field values, support arguments
-
@cypher directive for top level queries and mutations, supports arguments
-
@cypher directives can have a
passThrough:true
argument, that gives sole responsibility for the nested query result for this field to your Cypher query. You will have to provide all data/structure required by client queries. Otherwise, we assume if you return object-types that you will return the appropriate nodes from your statement. -
auto-generate mutation fields for all objects to create, update, delete
-
date(time)
-
interfaces
-
complex filter parameters, with optional query optimization strategy
-
scalars
-
spatial
-
skip limit params
-
sorting (nested)
-
ignoring fields
Next
-
input types
-
unions
Documentation
Parse SDL schema
-
Currently schemas with object types, enums, fragments and Query types are parsed and handled.
-
@relation
directives on fields and types for rich relationships -
@cypher
directives on fields and top-level query and mutation fields. -
The configurable augmentation auto-generates queries and mutations (create,update,delete) for all types.
-
Supports the built-in scalars for GraphQL.
-
For arguments input types in many places and filters from GraphCool/Prisma.
Resolve query Fields via Result Types
For query fields that result in object types (even if wrapped in list/non-null), the appropriate object type is determined via the schema and used to translate the query.
e.g.
type Query {
person: [Person]
}
# query "person" is resolved to and via "Person"
type Person {
name : String
}
Handle Arguments as Equality Comparisons for Top Level and Nested Fields
If you add a simple argument to your top-level query or nested related fields, those will be translated to direct equality comparisons.
person(name:"Joe", age:42) {
name
}
to an equivalent of
MATCH (person:Person) WHERE person.name = 'Joe' AND person.age = 42 RETURN person { .name } AS person
The literal values are turned into Cypher query parameters.
Handle Relationships via @relation Directive on Schema Fields
If you want to represent a relationship from the graph in GraphQL you have to add a @relation
directive which contains the relationship-type and the direction.
The default direction for a relationship is 'OUT'. Other values are 'IN' and 'BOTH'.
So you can use different domain names in your GraphQL fields that are independent of your graph model.
type Person {
name : String
actedIn: [Movie] @relation(name:"ACTED_IN", direction:OUT)
}
person(name:"Keanu Reeves") {
name
actedIn {
title
}
}
Note
|
We use Neo4j’s pattern comprehensions to represent nested graph patterns in Cypher. This will be updated to subqueries from 4.1 |
Handle first, offset Arguments
To support pagination first
is translated to LIMIT
in Cypher and offset
into SKIP
For nested queries these are converted into slices for arrays.
person(offset: 5, first:10) {
name
}
MATCH (person:Person) RETURN person { .name } AS person SKIP 5 LIMIT 10
Argument Types: string, int, float, array
The default Neo4j Cypher types are handled both as argument types as well as field types.
Note
|
Spatial is not yet covered. |
Usage of ID
Each type is expected to have exactly one filed of type ID
defined.
If the field is named _id
, it is interpreted as the database internal graph ID.
So there are 3 cases:
type User {
email: ID!
name: String!
}
type User {
_id: ID!
email: String!
name: String!
}
type User {
_id: Int!
email: ID!
name: String!
}
Important
|
For the auto generated queries and mutations the ID field is used as primary key.
|
Tip
|
You should create a unique constraint on the ID fields
|
Parameter Support
GraphQL parameters are passed onto Cypher, these are resolved correctly when used within the GraphQL query.
Parametrization
For query injection prevention and caching purposes, literal values are translated into parameters.
person(name:"Joe", age:42, first:10) {
name
}
to
MATCH (person:Person)
WHERE person.name = $personName AND person.age = $personAge
RETURN person { .name } AS person
LIMIT $first
Those parameters are returned as part of the Cypher
type that’s returned from the translate()
method.
Aliases
We support query aliases, they are used as Cypher aliases too, so you get them back as keys in your result records.
For example:
query {
jane: person(name:"Jane") { name, age }
joe: person(name:"Joe") { name, age }
}
Inline and Named Fragments
This is more of a technical feature, both types of fragments are resolved internally.
Sorting (top-level)
We support sorting via an orderBy
argument, which takes an Enum or String value of fieldName_asc
or fieldName_desc
.
query {
person(orderBy:[name_asc, age_desc]) {
name
age
}
}
MATCH (person:Person)
RETURN person { .name, .age } AS person
ORDER BY person.name ASC, person.age DESC
Note
|
We don’t yet support ordering on nested relationship fields. |
Handle Rich Relationships via @relation Directive on Schema Types
To represent rich relationship types with properties, a @relation
directive is supported on an object type.
In our example it would be the Role
type.
type Role @relation(name:"ACTED_IN", from:"actor", to:"movie") {
actor: Person
movie: Movie
roles: [String]
}
type Person {
name: String
born: Int
roles: [Role]
}
type Movie {
title: String
released: Int
characters: [Role]
}
person(name:"Keanu Reeves") {
roles {
roles
movie {
title
}
}
}
Filters
Filters are a powerful way of selecting a subset of data. Inspired by the graph.cool/Prisma filter approach, our filters work the same way.
These filters are documented in detail in the https://grandstack.io/docs/graphql-filtering [GRANDstack docs^].
We use nested input types for arbitrary filtering on query types and fields.
{ Company(filter: { AND: { name_contains: "Ne", country_in ["SE"]}}) { name } }
You can also apply nested filter on relations, which use suffixes like ("",not,some, none, single, every)
{ Company(filter: {
employees_none { name_contains: "Jan"},
employees_some: { gender_in : [female]},
company_not: null })
{
name
}
}
Optimized Filters
If you encounter performance problems with the cypher queries generated for the filter, you can activate an alternative algorithm using:
var query
try {
val ctx = QueryContext(optimizedQuery = setOf(QueryContext.OptimizationStrategy.FILTER_AS_MATCH))
query = translator.translate(query, params, ctx)
} catch (e: OptimizedQueryException) {
query = translator.translate(query, params)
}
If no query can be generated by the alternative algorithm, an OptimizedQueryException
is thrown,
so that a fallback to the actual algorithm can used.
Examples of the alternative algorithm can be seen in the tests.
Inline and Named Fragments
We support inline and named fragments according to the GraphQL spec. Most of this is resolved on the parser/query side.
fragment details on Person { name, email, dob }
query {
person {
...details
}
}
query {
person {
... on Person { name, email, dob }
}
}
@cypher Directives
With @cypher
directives you can add the power of Cypher to your GraphQL API.
It allows you, without code to compute field values using complex queries.
You can also write your own, custom top-level queries and mutations using Cypher.
Arguments on the field are passed to the Cypher statement as parameters.
Input types are supported, they appear as Map
type in your Cypher statement.
Note
|
Those Cypher directive queries are only included in the generated Cypher statement if the field or query is included in the GraphQL query. |
On Fields
type Movie {
title: String
released: Int
similar(limit:Int=10): [Movie] @cypher(statement:
"""
MATCH (this)-->(:Genre)<--(sim)
WITH sim, count(*) as c ORDER BY c DESC LIMIT $limit
RETURN sim
""")
}
Here the this
variable is bound to the current movie.
You can use it to navigate the graph and collect data.
The limit
variable is passed to the query as parameter.
On Queries
Similarly, you can use the @cypher
directive with a top-level query.
type Query {
person(name:String) Person @cypher("MATCH (p:Person) WHERE p.name = $name RETURN p")
}
You can also return arrays from your query, the statements on query fields should be read-only queries.
On Mutations
You can do the same for mutations, just with updating Cypher statements.
type Mutation {
createPerson(name:String, age:Int) Person @cypher("CREATE (p:Person) SET p.name = $name, p.age = $age RETURN p")
}
You can use more complex statements for creating these entities or even subgraphs.
Note
|
The common CRUD mutations and queries are auto-generated, see below. |
Auto Generated Queries and Mutations
To reduce the amount of boilerplate code you have to write, we auto-generate generate top-level CRUD queries and mutations for all types.
This is configurable via the API, you can:
-
disable auto-generation (for mutations/queries)
-
disable per type
-
disable mutations per operation (create,delete,update)
-
configure capitalization of top level generated fields
For a schema like this:
type Person {
id:ID!
name: String
age: Int
movies: [Movie]
}
It would auto-generate quite a lot of things:
-
a query:
person(id:ID, name:String , age: Int, _id: Int, filter:_PersonFilter, orderBy:_PersonOrdering, first:Int, offset:Int) : [Person]
-
a
_PersonOrdering
enum, for theorderBy
argument with all fields for_asc
and_desc
sort order -
a
_PersonInput
for creating Person objects -
a
_PersonFilter
for thefilter
argument, which is a deeply nested input object (see Filters) -
mutations for:
-
createPerson:
createPerson(id:ID!, name:String, age: Int) : Person
-
mergePerson:
mergePerson(id:ID!, name:String, age:Int) : Person
-
updatePerson:
updatePerson(id:ID!, name:String, age:Int) : Person
-
deletePerson:
deletePerson(id:ID!) : Person
-
addPersonMovies:
addPersonMovies(id:ID!,movies:[ID!]!) : Person
-
deletePersonMovies:
deletePersonMovies(id:ID!,movies:[ID!]!) : Person
-
You can then use those in your GraphQL queries like this:
query { person(age:42, orderBy:name_asc) {
id
name
age
}
or
mutation {
createPerson(id: "34920n9qw0", name:"Jane Doe", age:42) {
id
name
age
}
}
You find more examples in the Augmentation Tests and the Custom queries and mutations Tests
Build time schema augmentation
Sometimes you need the possibility to generate the augmented schema at compile time. To achieve this, we provide a maven plugin which can be used as follows:
<plugin>
<groupId>org.neo4j</groupId>
<artifactId>neo4j-graphql-augmented-schema-generator-maven-plugin</artifactId>
<version>1.4.0</version>
<executions>
<execution>
<goals>
<goal>generate-schema</goal>
</goals>
<configuration>
<schemaConfig> <!--(1)-->
<pluralizeFields>true</pluralizeFields>
<useWhereFilter>true</useWhereFilter>
<queryOptionStyle>INPUT_TYPE</queryOptionStyle>
<mutation>
<enabled>false</enabled>
</mutation>
</schemaConfig>
<outputDirectory>${project.build.directory}/augmented-schema</outputDirectory>
<fileset> <!--(2)-->
<directory>${project.basedir}/src/main/resources</directory>
<include>*.graphql</include>
</fileset>
</configuration>
</execution>
</executions>
</plugin>
-
Use the same configuration as for your SchemaBuilder
-
Define the source schema for which you want to have an augmented schema generated
Take a look at the spring boot dsg example for a use case of this plugin, where it is used in combination with a code generator to have a type save graphql API