New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Major bug: GroupBy aggregates incorrect ranges #717
Comments
A possible workaround is to specify a granularity of type "period" with a period of "PT3M", with the same origin. PeriodGranularity should not be affected by this but, since it uses a different algorithm to truncate timestamps. |
As far as the timestamps are concerned, the granularity timestamp truncation is always relative to time zero (a.k.a. Jan 1 1970 UTC), and not relative to the query interval. |
So this is indeed a bug then. On Saturday, August 30, 2014, xvrl notifications@github.com wrote:
Prashant |
Yes, there is a bug if you specify an origin. If you do not specify an origin, the behavior is correct. |
ok so using period granularity on 0.6.121, I still get a 500, Details: Query sent:
This time i see exception only on broker, not realtime node. There are multiple:
I guess just using an origin at all screws up. in a completely different way when using period granularity though. |
also while i am doing this, I just want to know if i am on the right track. |
You are specifying your granularity period incorrectly, you need to replace the
To answer your second question, yes if you specify 17:46 as the origin, your timestamps will go in increments of 3 minutes starting at 17:46. |
ok, yes with the correction you mentioned, using period granularity indeed works as needed. |
Hi @pdeva, were you able to check with a recent version of Druid? |
not yet Prashant On Mon, Sep 8, 2014 at 11:31 AM, xvrl notifications@github.com wrote:
|
I am closing this. Please reopen if error reproducible. |
Assume your data begins at time 17:48 and is stored every minute.
You give GroupBy an interval begining at 17:46, with a granularity of 3 mins.
You should get data with timestamps:
etc..
However, you get:
This is not the data that was asked for. The interval returned should be the one in the former list.
There is currently no workaround for this bug either. If you want data aggregated and looking like the former list, there is no way to ask druid for it
Update
While there is a origin option in granularity, it doesnt actually work since it always makes the realtime node throw an exception (thereby the broker returns a 500):
I am using druid 0.6.121
The text was updated successfully, but these errors were encountered: