Fork me on GitHub
#datomic
<
2018-10-12
>
stijn09:10:19

to be clear, there's no update of the storage stack required, right?

oz14:10:48

In CFT 441 you switched to YAML, now in 441-8505 you are using json again. It's not a big deal this time, but in the storage stack we have a couple of modification that get around the issue with running in a AWS account that has EC2 classic support. So we have to cherry pick those changes by hand into any storge CF template upgrades. I did this just recently from 297 to 409 and it wasn't too bad, but the version 441 was a bit harder due to the change from json to yaml. Again it's a non-issue for 441-8505 since it's only a compute stack change, but going forward can you please distribute one format or both yaml and json?

marshall14:10:02

the YAML change was a marketplace artifact; we did not choose that and we intend to use json

oz14:10:21

đź‘Ś

eoliphant20:10:00

I posted this in the main thread as well. I'm applying this to a solo setup. the compute upgrade worked fine, but I'm getting a Error creating change set: The submitted information didn't contain changes. Submit different information to create a change set. back from CF when I try to apply the update for storage

csm00:10:33

does upgrading really require deleting the stack and creating it again?

stuarthalloway00:10:41

Hi @U1WMPA45U, it depends on what you mean by "the stack".

csm00:10:31

I launched the CF template and got three CF stacks in total — so update the stack named “compute”, yes?

stuarthalloway00:10:22

Unfortunately, no. AWS's marketplace rules are in direct conflict with AWS's CloudFormation best practice guidelines. The "deleting" path takes you from Marketplace-land to CF-best-practice-land.

stuarthalloway00:10:54

After you do this once, you will be in CF-best-practice-land and never have to do it again.

csm00:10:10

got it, thanks!

jocrau01:10:22

Just a quick note on the update: I had to re-adjust the capacity settings of the compute stack autoscaling group. The update seems to reset this to “desired 2, min 2, and max 3”.

stijn09:10:24

same issue here

marshall13:10:45

those are the default values; had you changed them to something else?

marshall13:10:33

you should not be using autoscaling on the primary compute group

jocrau13:10:28

I have two use-cases to change autoscaling of the primary compute group: First, to save money while experimenting with the prod topology (I set them to 0 during times I am not working on it), and second to try to increase transaction performance for large imports (that might be a brute force approach, see also https://forum.datomic.com/t/tuning-transactor-memory-in-datomic-cloud/643).

jocrau13:10:20

The documentation is a bit confusing. The two sentences “If you are writing to a large number of different databases, you can increase the size of the primary compute group by explicitly expanding its Auto Scaling Group.” and “You should not enable AWS Auto Scaling on the primary compute group.” seem to contradict each other. Am I missing something?

marshall13:10:21

Those should not be autoscaling events; Scaling the group will not affect throughput for a single DB

marshall13:10:39

you can adjust the size of the group explicitly

marshall13:10:45

you shouldn’t use AutoScaling

marshall13:10:10

i.e. Autoscaling events == things that AWS does for you triggered based on some metric/event

marshall13:10:33

Changing the “min” “max” and “desired” explicitly is OK, but should be a fairly infrequent human-required action

jocrau13:10:17

You are right that it does not make sense to change the “desired” setting of the autoscaling group to adapt to spikes (that’s what the “auto” in autoscaling is for). But to increase the “max” and “desired” seems to be the best option currently to increase transaction (and indexing) performance in case of a known, large import.

marshall14:10:45

if the import runs against a single database, changing the size of the compute group will not affect throughput

marshall14:10:16

you can change to using an i3.xlarge (instead of the default i3.large)

marshall14:10:21

in your compute group

marshall14:10:26

that will improve import perf

jocrau14:10:25

Ok. I will give that a try.

jocrau15:10:15

Does the number of nodes influence indexing performance (on a single database)?

stuarthalloway15:10:44

@jocrau @marshall as long as you have at least two, no

stijn07:10:09

our use case is to have a prod setup for our staging environment, but without HA. so we set the 3 values (min, max, desired) to 1.

jocrau14:10:20

A deployment to the compute group seems to be performed sequentially (see attached graph which shows the incoming network traffic to 5 nodes; mostly the JAR files I assume). Can this be done in parallel to speed up the deployment?

marshall14:10:15

@jocrau No, that is specifically the way that rolling deployments work to maintain uptime and enable rollback

jocrau14:10:10

@marshall Makes sense. Thanks.

Lone Ranger16:10:20

Does anyone happen to know where I could find some good marketing materials on the Datomic value prop for non-technical executives?

val_waeselynck16:10:51

Note that the value prop of Datomic Cloud is a bit different

Lone Ranger16:10:59

Ha! How serendipitous!

Lone Ranger16:10:45

that's fine, due to the nature of the data/work it would have to be self-hosted anyway.

csm17:10:49

I don’t need to recreate my VPC endpoint again when I perform my first upgrade, do I?

csm17:10:19

I think I have an answer to that: I can’t delete the datomic cloud root stack since the VPC endpoint stack depends on resources

csm17:10:52

the “compute” stack just failed to delete for me, with reason The following resource(s) failed to delete: [HostedZone].

csm17:10:00

…which was because I had a record set for my VPC endpoint…

grzm17:10:53

I'm setting up a new query group. The stack created without error. When doing the first deploy to the query group, it's failing ValidateService with ScriptFailed. The events details show [stdout]Received 503 a number of times, and then finally [stdout]WARN: validation did not succeed after two minutes. Ideas on where to start debugging this?

kenny18:10:59

@grzm Check the CloudWatch logs - there was probably an exception.

grzm18:10:37

@kenny cheers. thanks for the kick in the right direction.