New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
avro-extensions -- feature to specify avro reader schema inline #3249
Conversation
This is something we need in our set up. Thanks Himanshu. Which release is this expected to be a part of ? |
|
||
private final Schema schemaObj; | ||
|
||
private Map<String, Object> schema; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
final?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
added
This should be reasonable to be in 0.9.2, and if there's a community need even more so |
@skoppar do you have single schema for all the events in the kafka topic or multiple ? is it possible for you to test this patch? are you already trying to use avro extension with schema_repo and facing some error? |
LGTM, 👍 |
updated the docs. |
Hi Himanshu, I am testing your patch now. Will keep you posted |
…e task json for all events
Hi Himanshu, I am not able to get the avro stream to work with the new jar. I have druid.extensions.loadList=["druid-avro-extensions", "druid-histogram", I've modified kafka.json in imply-1.3.0/conf-quickstart/tranquility to the Upon starting imply: I do not see any topics being picked in the tranquility logs Doesn't it mean that no topics have been registered? Only relevant entries in the log are: I am new to Druid, so please pardon my ignorance. Is the regards On Fri, Aug 5, 2016 at 7:35 AM, Himanshu notifications@github.com wrote:
|
Thanks for quick response Himanshu. Provides the topic which is per datasource level and below properties provide overall cluster connection. Adding group_id to data source specs might be more advisable, but I haven't tried that yet. I am using all default ports. "kafka.group.id":"cust1-tranquility", regards On Mon, Aug 8, 2016 at 11:47 AM, Himanshu notifications@github.com wrote:
|
Another fact is, Zookeeper is able to recognize the consumer group and the |
@skoppar seems like you have specified the properties at wrong place in json spec file, |
Thanks Nishant for bringing it to my notice. Currently getting some class loading issues with the new jar. Probably I need to double check the extension loadlist. Will update here. |
Hi Himanshu, Looks like there is a dependency on io.druid.druid-api-0.9.1.jar based on the errors I see. imply-1.3.0/dist/tranquility/lib has a huge list of jars. I hope I have to replace fewer jars. Are there any documented dependencies? Attaching the exception I get now. |
@skoppar what is the tranquility config you are using now? |
@gianm attaching the current tranquility-kafka json file. |
@himanshug doesnt look like loading the jar is the issue. I could be wrong though. Attaching the indexing-log in imply-1.3.0/var/druid/indexing-logs which shows the druid-avro-extensions-0.9.2-SNAPSHOT.jar being added. Please let me know if I am looking at a wrong place. |
@skoppar one of the problems in your json is that "parseSpec" is inside "avroBytesDecoder" while it should be a level up inside "parser" . |
oops ! let me relook at the complete json On Tue, Aug 9, 2016 at 1:14 PM, Himanshu notifications@github.com wrote:
|
Its working now. 2 more changes I had to do was:
With that done to the attached conf, I am able to process records with schema inline. However, I do not see the data source in druid console or pivot. I can see metrics-kafka (had run the example), so I was expecting this should show up as well. Using topic name - raw-avro for this test. Below is what the tranquility-kafka log shows - the avro schema and then the below message |
@skoppar glad to know that it worked out for you and thanks for testing the patch. |
👍 |
Well not completely. Trying to figure out why segments are not getting On Aug 10, 2016, at 10:47 AM, Himanshu notifications@github.com wrote: @skoppar https://github.com/skoppar glad to know that it worked out for — |
Well not completely. Trying to figure out why segments are not getting |
@skoppar enable debug logging and that would show you more info. |
Figured it out with the help of a kafka consumer class. The console consumer showed data just fine, so couldn't spot the error. The rootcause is, I was sending the avro message using StringEncoder. Using DefaultEncoder on producer side fixed the issue. Thanks @himanshug and @gianm. Awaiting 9.2 release :) 👍 |
in the task json for all events.
See https://groups.google.com/forum/?utm_medium=email&utm_source=footer#!msg/druid-development/tg2q3VHN5rQ/5kUuiTouEQAJ