Fork me on GitHub
#xtdb
<
2023-09-13
>
jarohen07:09:51

@otfrom if you're using Arrow, have you seen the upcoming XT 2? https://xtdb.com/v2

otfrom08:09:24

yeah, I've got my eye on that. Is the goal that I'd just need the files and the jar or would I need other external things (kafka other) as well?

jarohen08:09:05

at the moment it's coupled in the artifact but we could likely release the Arrow query engine separately if there was demand for it

otfrom10:09:59

I think I'd be looking for something that I could use in process from a single jar. I've not properly looked at things since the very early days of needing to set up things like kafka (and postgres? been so long I can't even remember)

jarohen10:09:50

ah, sorry, my misunderstanding - it'll still be possible to create in-memory/local-disk nodes straight from the core JAR, Kafka is an optional extra in that regard

jarohen10:09:50

I thought you'd just be interested in the standalone query engine running queries over existing, provided Arrow files, at which point things like transaction processing etc also aren't necessary

jarohen10:09:29

but yes, in-process nodes are certainly still a core feature

otfrom10:09:06

the arrow stuff would be interesting, tho the competition there is DuckDB in terms of performance and I'm more performance than expressively constrained in my analysis

otfrom10:09:18

https://github.com/techascent/tmducken and general tmd ways of using larger than memory arrow files

otfrom10:09:27

sounds like my understanding of what is needed for xtdb was a bit out of date. I had it in the "very capable, but high cost (of effort and admin)" category of tools and it sounds like it might work for some embedded single node purposes

otfrom10:09:11

getting used to using it in small personal things and then being able to scale up if needed but being familiar with data modelling and the rest using to tool is a real sweet spot for me