The Es Kotlin Client provides a friendly Kotlin API on top of the official Elastic Java client.
Elastic's HighLevelRestClient
is written in Java and provides access to essentially everything in the REST API that Elasticsearch exposes. This API provides access to all of the oss and x-pack features. However, it is not the easiest thing to work with directly.
The Es Kotlin Client takes away none of that power but adds a lot of power and convenience.
The underlying java functionality is always there ready to be used if you need it. However, for most commonly used things, this client provides more Kotlin appropriate ways to access the functionality.
- Extensible Kotlin DSLs for Querying, Mappings, Bulk Indexing, and Object CRUD. These kotlin DSLs provide type safe support for commonly used things such as match and bool queries as well as defining mappings and settings for your indices. At this point most commonly used queries are supported including all full-text queries, compound queries, and term-level queries.
- Things that are not supported are easy to configure by modifying the underlying data model directly using Kotlin's syntactic sugar for working with
Map
. - To enable this our DSL classes delegate to a
MapBackedProperties
class that backs normal type safe kotlin properties with aMap
. Anything that's not supported, you can just add yourself. Additionally, it is easy to extend the DSLs with your own type safe constructions (pull requests welcome) if you are using some query or mapping construction that is not yet supported.
- Things that are not supported are easy to configure by modifying the underlying data model directly using Kotlin's syntactic sugar for working with
- Kotlin Extension functions, default argument values, delegate properties, and many other kotlin features add convenience and get rid of Java specific boilerplate. The Java client is designed for Java users and comes with a lot of things that are a bit awkward / non idiomatic in Kotlin. This client cuts down on the boiler plate and uses Kotlin's DSL features, extension functions, etc. to layer a friendly API over the underlying client functionality.
- A repository abstraction that allows you represent an Index with a data class:
- Manage indices with a flexible DSL for mappings.
- Serialize/deserialize JSON objects using your favorite serialization framework. A Jackson implementation comes with the client but you can trivially add support for other frameworks. Deserialization support is also available on search results and the bulk API.
- Do CRUD on json documents with safe updates that retry in case of a version conflict.
- Bulk indexing DSL to do bulk operations without boiler plate and with fine-grained error handling (via callbacks)
- Search & count objects in the index using a Kotlin Query DSL or simply use raw json from either a file or a templated multiline kotlin string. Or if you really want, you can use the Java builders that come with the RestHighLevelClient.
- Much more
- Co-routine friendly & ready for reactive usage. We use generated extension functions that we add with a source generation plugin to add cancellable suspend functions for almost all client functions. Additionally, the before mentioned
IndexRepository
has anAsyncIndexRepository
variant with suspending variants of the same functionality. Where appropriate, Kotlin's newFlow
API is used. - This means this Kotlin library is currently the most convenient way to use Elasticsearch from e.g. Ktor or Spring Boot if you want to use asynchronous IO. Using the Java client like this library does is of course possible but will end up being very boiler plate heavy.
As JFrog is shutting down JCenter, the latest releases are once more available via Jitpack.
Add this to your build.gradke.kts
implementation("com.github.jillesvangurp:es-kotlin-client:<version>")
You will also need to add the Jitpack repository:
repositories {
mavenCentral()
maven { url = uri("https://jitpack.io") }
}
See release notes with each github release tag for an overview what changed.
- manual A growing collection of executable examples. This manual is generated from kotlin code and all the examples in it are run as part of the test suite. This is the best place to get started.
- The same manual as an epub. Very much a work in progress. Please give me feedback on this. I may end up self publishing this at some point.
- dokka api docs - API documentation - this gets regenerated for each release and should usually be up to date. But you can always
gradle dokka
yourself. - Some stand alone examples are included in the examples source folder.
- The tests test most of the important features and should be fairly readable and provide a good overview of how to use things.
- Elasticsearch java client documentation - All of the functionality provided by the java client is supported. All this kotlin wrapper does is add stuff. Elasticsearch has awesome documentation for this.
- demo project - Note, this is outdated and has largely been replaced by the manual mentioned above.
- Within reason, I'm always happy to support users with issues. Feel free to approach me via the issue tracker, twitter (jillesvangurp), or email (jilles AT jillesvangurp.com). I'm also available for consulting on Elasticsearch projects and am happy to help you with architecture reviews, implementation work, query & indexing strategy, etc. Check my homepage for more details.
This example is a bit more complicated than a typical hello world but more instructive than just putting some objects in a schema less index. Which of course is something you should not do. The idea here is to touch on most topics a software engineer would need to deal with when creating a new project using Elasticsearch:
- figuring out how to create an index and define a mapping
- populating the index with content using the bulk API
- querying data
// given some model class with two fields
data class Thing(
val name: String,
val amount: Long = 42
)
// create a Repository
// with the default jackson model reader and writer
// (you can use something else by overriding default values of the args)
val thingRepository = esClient.indexRepository<Thing>(
index = "things",
// you have to opt in to refreshes, bad idea to refresh in production code
refreshAllowed = true
)
// let the Repository create the index with the specified mappings & settings
thingRepository.createIndex {
// we use our settings DSL here
// you can also choose to use a source block with e.g. multiline strings
// containing json
configure {
settings {
replicas = 0
shards = 2
}
mappings {
// mappings DSL, most common field types are supported
text("name")
// floats, longs, doubles, etc. should just work
number<Int>("amount")
}
}
}
// lets create a few Things
thingRepository.index("1", Thing("foo", 42))
thingRepository.index("2", Thing("bar", 42))
thingRepository.index("3", Thing("foobar", 42))
// make sure ES commits the changes so we can search
thingRepository.refresh()
val results = thingRepository.search {
configure {
// added names to the args for clarity here, but optional of course
query = match(field = "name", query = "bar")
}
}
// results know hot deserialize Things
results.mappedHits.forEach {
println(it.name)
}
// but you can also access the raw hits of course
results.searchHits.forEach {
println("hit with id ${it.id} and score ${it.score}")
}
// putting things into an index 1 by 1 is not scalable
// lets do some bulk inserts with the Bulk DSL
thingRepository.bulk {
// we are passed a BulkIndexingSession<Thing> in the block as 'this'
// we will bulk re-index the objects we already added with
// a scrolling search. Scrolling searches work just
// like normal searches (except they are not ranked)
// all you do is set scrolling to true and you can
// scroll through billions of results.
val sequence = thingRepository.search(scrolling = true) {
configure {
from = 0
// when scrolling, this is the scroll page size
resultSize = 10
query = bool {
should(
match("name", "foo"),
match("name", "bar"),
match("name", "foobar")
)
}
}
}.hits
// hits is a Sequence<Pair<SearchHit,Thing?>> so we get both the hit and
// the deserialized value. Sequences are of course lazy and we fetch
// more results as you process them.
// Thing is nullable because Elasticsearch allows source to be
// disabled on indices.
sequence.forEach { (esResult, deserialized) ->
if (deserialized != null) {
// Use the BulkIndexingSession to index a transformed version
// of the original thing
index(
esResult.id, deserialized.copy(amount = deserialized.amount + 1),
// allow updates of existing things
create = false
)
}
}
}
Using co-routines is easy in Kotlin. Mostly things work almost the same way. Except everything is non blocking and asynchronous, which is nice. In other languages this creates all sorts of complications that Kotlin largely avoids.
The Java client in Elasticsearch has support for non blocking IO. We leverage this to add our own suspending calls using extension functions via our gradle code generation plugin. This runs as part of the build process for this library so there should be no need for you to use this plugin.
The added functions have the same signatures as their blocking variants. Except of course they have the word async in their names and the suspend keyword in front of them.
We added suspending versions of the Repository
and BulkSession
as well, so either blocking or non
blocking. It's up to you.
// we reuse the index we created already to create an ayncy index repo
val repo = esClient.asyncIndexRepository<Thing>(
index = "things",
refreshAllowed = true
)
// co routines require a CoroutineScope, so let use one
runBlocking {
// lets create some more things; this works the same as before
repo.bulk {
// but we now get an AsyncBulkIndexingSession<Thing>
(1..1000).forEach {
index(
"$it",
Thing("thing #$it", it.toLong())
)
}
}
// refresh so we can search
repo.refresh()
// if you are wondering, yes this is almost identical
// to the synchronous version above.
// However, we now get an AsyncSearchResults back
val results = repo.search {
configure {
query = term("name.keyword", "thing #666")
}
}
// However, mappedHits is now a Flow instead of a Sequence
results.mappedHits.collect {
// collect is part of the kotlin Flow API
// this is one of the few parts where the API is different
println("we've found a thing with: ${it.amount}")
}
}
Captured Output:
we've found a thing with: 666
For more examples, check the manual or look in the examples source directory.
This library makes use of code generated by our code generation gradle plugin. This is mainly used to generate co-routine friendly suspendable extension functions for all asynchronous methods in the RestHighLevelClient. We may add more features in the future.
This client requires Java 8 or higher (same JVM requirements as Elasticsearch). Some of the Kotlin functionality is also usable by Java developers (with some restrictions). However, you will probably want to use this from Kotlin. Android is currently not supported as the minimum requirements for Elastic's highlevel client are Java 8. Besides, embedding a fat library like that on Android is probably a bad idea and you should probably not be talking to Elasticsearch directly from a mobile phone in any case.
This library is tested with Elasticsearch 7.x. Usually we update to the latest version within days after Elasticsearch releases. Barring REST API changes, most of this client should work with any recent 7.x releases, any future releases and probably also 6.x.
For version 6.x, check the es-6.7.x branch.
You need java >= 8 and docker + docker compose to run elasticsearch and the tests.
Simply use the gradle wrapper to build things:
./gradlew build
Look inside the build file for comments and documentation.
If you create a pull request, please also regenerate the documentation with build.sh
script. It regenerates
the documentation. We host and version the documentation in this repository along with the source code.
Gradle will spin up Elasticsearch using the docker compose plugin for gradle and then run the tests against that.
If you want to run the tests from your IDE, just use docker-compose up -d
to start ES. The tests expect to find
that on a non standard port of 9999
. This is to avoid accidentally running tests against a real cluster and making
a mess there (I learned that lesson a long time ago).
This library should be perfectly fine for general use at this point. We released 1.0 beginning of 2021 an the plan is to continue to add functionality and continue to maintain this.
If you want to contribute, please file tickets, create PRs, etc. For bigger work, please communicate before hand before committing a lot of your time. I'm just inventing this as I go. Let me know what you think.
This project is licensed under the MIT license. This maximizes everybody's freedom to do what needs doing. Please exercise your rights under this license in any way you feel is appropriate. Forking is allowed and encouraged. I do appreciate attribution and pull requests ...
For small things, use the issue tracker; I generally try to be helpful and try to resolve valid issues quickly. However, I'm also available for consultancy and specialize in a few things, including of course Elasticsearch. I've supported small and large companies with search related product features, complex migrations, query optimizations, plugin development, and more. I have over 15 years of experience with Lucene, Solr, and of course Elasticsearch. So, if you need help with your Elastic setup, want to improve your queries or ingest architecture, or otherwise could use a second pair of eyes, please reach out.