slackbot6<dr.max> Ok, thx @keviny, about week after next?
slackbot6<keviny> We should be available by then
slackbot6<dr.max> For me two things to highlight: 1. Various acceptance 2. I got a proposal from @hsliev for two commiters from his team (one a replacement As previous left SAP). I am doing reviewing and might run a vote later today unless I have questions
slackbot6<dr.max> Please chime in when you have a minute and send questions to me about ^^^
slackbot6<dr.max> Finally, IBM team looking forward to chatting on July 12th @ 11a. Can someone schedule the normal room we've used in the past in Foster City. Thx
slackbot6<hsiliev> !here|@here Team was busy fixing: 1. negative usage caused by bridge due to a) restart using stop&start events with the same timeout, b) events reported with seconds accuracy by CC, therefore we might get events with same timestamp 2. reporting timeout due to a) reporting killed since heartbeat & healthcheck fail b) request timeout. Both are caused by event loop occupied by summary/charge for > 3000 consumers
slackbot6<hsiliev> Both bugs imho are prerequisites for 1.0 release
slackbot6<dr.max> Hi, <here>, remember our call at 10a PDT today. Just joined by clicking on this zoom URL: https://zoom.us/j/788691850
slackbot6<dr.max> See you soon. Cheers
chihyi88For example, I shall send usage # 1, usage # 2, usage #3 to abacus, but for some reason, I failed to send usage #2, If it is a discrete type, I only lost usage #2
chihyi88But if it is a time-based, the charge at report would go crazy. My question is, for this problem, is there a way that I can recovery from this, and avoid the charge go crazy ?
slackbot6<hsiliev> You need to wait a bit for the doc to go through the pipeline and end up processed in the final database. This usually takes up to 2 minutes
slackbot6<hsiliev> Then the doc would be accumulated/aggregated right away
slackbot6<david.wu> Got it. Thank you very much.
slackbot6<dr.max> Hi, <here>, remember our call at 10a PDT today. I have a small conflict, would it be possible to delay 30 mins? Sorry about the late notice. Please chime here. If not possible let's do status/update here and next week we can do live since I'll be at Foster City. Thx!!
slackbot6<dr.max> I'll assume we do status here. My status is twofold: 1. Doing some acceptance and clean up and adding the new committers from SAP 2. Using abacus as test for new cf-Extensions project so you will see a few commits from me adding metadata to repo. I plan to discuss this at the cf-Extensions PMC call Monday 7/31 if you can join
slackbot6<dr.max> Let me know if you are waiting on me for anything. For the IBM folks I am looking forward to chatting in person in foster city. However, don't have invite. Can someone send. We agreed Th or Friday 11a
slackbot6<hsiliev> Status update from the team: doing some profiling of reporting. Seems most of the time is spent in vm/vm2 sandbox. We applied caching and managed to reduce the time for generating a report from 2:40 minutes to 0:25 seconds. started productizing our internal Service Bridge MVP prepared a demo recording for Abacus Broker. Should be available next week hidden in youtube. Will send a link to the team for comments g
slackbot6describing different aspects of Abacus for Resource Providers
slackbot1<dr.max> <here> anyone else joining? I am on
slackbot1<hsiliev> !here|@here I'm on vacation the whole month. Sorry for the late notice.
slackbot1<hsiliev> Not sure the plan really has previous and current. @keviny Is this normal in your pov?
slackbot1<hsiliev> It may happen if we have retries after the collector
slackbot1<hsiliev> Actually after the meter
slackbot1<carrolp> retrying now....
slackbot1<carrolp> Did not recreate with a second attempt either. @keviny I didn't quite follow your example of "failing at the duplicate doc". Would this have happened if I had emptied the databases but not deleted them and then tried to send in the same data again? The CouchDB I'm using would still have the old documents marked as _deleted=true, and potentially leading to a document conflict.
slackbot1<carrolp> Could this explain the problem I saw originally perhaps?
slackbot1<hsiliev> Did you restart Abacus after clearing the DB?
slackbot1<carrolp> @hsiliev It was a couple days ago, I can't recall for sure. If I do not restart the DB after clearing the databases, I think I just recreated the problem -- I sent in two records, deleted them, sent in two more records, and the consumption report returned '40' instead of the correct '20'
slackbot1<keviny> Yeah, if you cleaned the db, but does not restart the app
slackbot1<keviny> Abacus has the cache on the accumulated usage doc (which keeps the 20)
slackbot1<keviny> And since the db is cleaned, when it checks if the record was submitted previously (by fetching the duplicate doc from db) it will says not a duplicate.
slackbot1<keviny> so it will go ahead and accumulate it to 40
slackbot1<carrolp> I can confirm now that the 10+10=30 problem was purely due to my experimentation -- clearing the databases and retrying without restarting abacus. I wasn't deliberately trying to keep abacus running originally, but could easily have done it by accident when resetting for another test. Thanks for the help!