{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":32607095,"defaultBranch":"master","name":"wikidata-query-rdf","ownerLogin":"wikimedia","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2015-03-20T21:06:14.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/56668?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1714676537.0","currentOid":""},"activityList":{"items":[{"before":"a33c6ea65c49b0d32b7f4e7a2da80e0ca9747a9b","after":"d5491828dba0a29a8024af0d3358af99c3daaef3","ref":"refs/heads/master","pushedAt":"2024-05-03T01:50:05.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"SubgraphDefinitions: embed prefixes in the declaration\n\nPrefixes should be part of the model definition, this might be less\nerror-prone.\n\nChange-Id: Ic0377f77af5d7a33368606d66f948bff4bf02276","shortMessageHtmlLink":"SubgraphDefinitions: embed prefixes in the declaration"}},{"before":"4a8e8c3883a5d1837e6c3485c25d90e7a6eae6bc","after":"a33c6ea65c49b0d32b7f4e7a2da80e0ca9747a9b","ref":"refs/heads/master","pushedAt":"2024-05-02T20:45:17.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"ScholarlyArticleSplit: add support for stubs\n\nSubgraphRuleMapper is able to identify entities requiring stubs.\nAdded a new ontology entry to add custum query service vocabulary.\nGeneralize ScholarlyArticleSplit to allow passing other strategies.\n\nBug: T362060\nChange-Id: I5881e3314f002822469f41672b2e4c0ac07940fc","shortMessageHtmlLink":"ScholarlyArticleSplit: add support for stubs"}},{"before":"d1e6bd47b1d5d64fd3ba0cf6871332b685882504","after":"4a8e8c3883a5d1837e6c3485c25d90e7a6eae6bc","ref":"refs/heads/master","pushedAt":"2024-05-02T20:05:05.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Split based on subgraph rules\n\nApply rules based on a set of subgraph definitions.\n\nSubgraphRuleMapper generate a dataset per subgraph with the entity URIs.\n\nBug: T362060\nChange-Id: I761d3df6b2cd795e3e0e4d8915bf3671e6360927","shortMessageHtmlLink":"Split based on subgraph rules"}},{"before":"59d200526d6f76f78d9030d499c91216c788edc6","after":"d1e6bd47b1d5d64fd3ba0cf6871332b685882504","ref":"refs/heads/master","pushedAt":"2024-05-02T19:02:17.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"[maven-release-plugin] prepare for next development iteration","shortMessageHtmlLink":"[maven-release-plugin] prepare for next development iteration"}},{"before":"375c5bf1aeb3886fbaeea30b8a7e28f37bf13529","after":"59d200526d6f76f78d9030d499c91216c788edc6","ref":"refs/heads/master","pushedAt":"2024-04-30T09:12:47.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"[maven-release-plugin] prepare for next development iteration","shortMessageHtmlLink":"[maven-release-plugin] prepare for next development iteration"}},{"before":"5c9affbf53cd9163337375810a630e1f42e24591","after":"375c5bf1aeb3886fbaeea30b8a7e28f37bf13529","ref":"refs/heads/master","pushedAt":"2024-04-30T01:12:57.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"TurtleImporter: Parse header and first wikibase:Dump stmt\n\nThe first one was ignored previously but seems useful to keep.\n\nChange-Id: I978473e283bd9591b1dc6a91903068cf8084548f","shortMessageHtmlLink":"TurtleImporter: Parse header and first wikibase:Dump stmt"}},{"before":"4a8d89b595a6f126544b57bfb8243e3d8260db5b","after":"5c9affbf53cd9163337375810a630e1f42e24591","ref":"refs/heads/master","pushedAt":"2024-04-26T18:14:32.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Deduplicate side output events and send them with a recent event-time\n\nSide output events may be duplicated in case the flink job is having\nissues, deduplicate them based on revision id and other criteria to\nreduce their impact.\nAlso re-assign a new event-time right before sending them to event-gate\nso that it's less likely to have them being considered late again.\n\nBug: T362508\nChange-Id: I5d7cb9a58cbf261786ec65bbb86ebf154438d3ef","shortMessageHtmlLink":"Deduplicate side output events and send them with a recent event-time"}},{"before":"fa9bf4e160b954f5bc8b0fe432c40086f63a6172","after":"4a8d89b595a6f126544b57bfb8243e3d8260db5b","ref":"refs/heads/master","pushedAt":"2024-04-25T19:05:03.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"ScholarlyArticleSplitter: add more complex values in test\n\nTest values having normalized to themselves and references having\nnormalized values.\n\nChange-Id: Ifdd99534119d38bc1037c00f35a7743ffabe3691","shortMessageHtmlLink":"ScholarlyArticleSplitter: add more complex values in test"}},{"before":"8c3b3cdd10a707858ea07d63d6fc95bf35ed031b","after":"fa9bf4e160b954f5bc8b0fe432c40086f63a6172","ref":"refs/heads/master","pushedAt":"2024-04-25T19:04:40.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Generalize ScholarlyArticleSplitter\n\nrefactor the splitting process to help the generalization of the\nsplitting rules.\nRules should be responsible for identifying the entity uris that will\nbelong to each subgraph.\nThen based on this set of entities collect the corresponding triples\nusing the same initial logic used to extract scholarly articles triples.\nSimplify the code a bit by removing the need to alias dataframes using\nDataFrame.apply(colName: String) function to reference individual\ncolumns.\n\nBug: T362060\nChange-Id: I030035076cd984f30d822f0b1ef25fc6ed787b09","shortMessageHtmlLink":"Generalize ScholarlyArticleSplitter"}},{"before":"667e3cb2d1c6406239ef52f5026d103ccae0871f","after":"8c3b3cdd10a707858ea07d63d6fc95bf35ed031b","ref":"refs/heads/master","pushedAt":"2024-04-23T16:48:56.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Add a basic model for defining subgraphs\n\nIt supports defining only very simple rules, prefixes are supported\nto ease the description of the various URIs.\nIt does only allow to encode simple triple patterns.\nIt is placed in tools since it will be required by both the flink &\nspark jobs.\n\nBug: T361935\nChange-Id: Ia17726673c890600259dd6924867e482dadab0a8","shortMessageHtmlLink":"Add a basic model for defining subgraphs"}},{"before":"64f6d70329f2b42b610c22476c18102492f51047","after":"667e3cb2d1c6406239ef52f5026d103ccae0871f","ref":"refs/heads/master","pushedAt":"2024-04-19T14:45:37.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Fix forbiddenapi usage\n\nAlso remove obsolete eclipse profile.\n\nChange-Id: Ie1111e740547657566fa86cb1139f694966eafe8","shortMessageHtmlLink":"Fix forbiddenapi usage"}},{"before":"870ba34301d1688a6f795344702e6fe5ebad0b36","after":"64f6d70329f2b42b610c22476c18102492f51047","ref":"refs/heads/master","pushedAt":"2024-03-27T21:33:45.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"[maven-release-plugin] prepare for next development iteration","shortMessageHtmlLink":"[maven-release-plugin] prepare for next development iteration"}},{"before":"f6e895c3780447c0155138b6f5ac88762a919f9c","after":"870ba34301d1688a6f795344702e6fe5ebad0b36","ref":"refs/heads/master","pushedAt":"2024-03-27T14:50:41.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Add support for x-monitoring-query header\n\nwhen this header is set we do not log the query.\n\nBug: T360993\nChange-Id: Iead1cdbf0b7b15836755bb5b0ad6f20f421480cd","shortMessageHtmlLink":"Add support for x-monitoring-query header"}},{"before":"ca747b79858e30d17ddd774d2791e85731f1e7e5","after":"f6e895c3780447c0155138b6f5ac88762a919f9c","ref":"refs/heads/master","pushedAt":"2024-01-30T11:51:06.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"QueryResultRecorder: add support for blank nodes\n\nBlank nodes can be requested from the query, unfortunately due to their\nnature the id that is assigned by blazegraph is highly dependent on the\nresults output order and thus makes the reordered comparison useless.\nComparison and hash will treat all blank nodes as equal to workaround\nthis problem.\n\nBug: T351819\nChange-Id: If0b3b696fa81b4f6ba08a43e76f8bde15fa8182a","shortMessageHtmlLink":"QueryResultRecorder: add support for blank nodes"}},{"before":"a266e83774204a803e58cb2cb939fae8b5242080","after":"ca747b79858e30d17ddd774d2791e85731f1e7e5","ref":"refs/heads/master","pushedAt":"2024-01-22T16:24:10.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"QueryResultRecorder: add support for ask, describe and construct\n\nQueries seen in the logs might be of various types and sadly their\nresponse format is not necessarily compatible with the binary tuple\nformat we expect.\nFor ASK queries we construct an artificial result set with a single\nstatic binding and a single line holding the boolean.\nFor construct and describe we rely on the fact that blazegraph does\nproduce a json response that is parseable as a tuple query response to\nobtain the tuples we expect (hence the new sesame rio dependency\nrequired).\n\nSome code has been ported from blazegraph itself to properly detect the\ntype of the query, main \"difficulty\" is to ignore the query prolog\n(prefix definitions/comments) to obtain the first token that identifies\nthe query type.\n\nBug: T351819\nChange-Id: If7f390efdbf3a41c710ae2b73cd63082c4cc8719","shortMessageHtmlLink":"QueryResultRecorder: add support for ask, describe and construct"}},{"before":"3469ec4acca88fcfb1e857ff83bcd0897a571e04","after":"a266e83774204a803e58cb2cb939fae8b5242080","ref":"refs/heads/master","pushedAt":"2024-01-19T17:19:13.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"QueryResultRecorder: bump the max response size to 16M\n\nWe still see query responses failing to be parsed because the result is\ntoo large. Make another attempt at 16M.\n\nBug: T351819\nChange-Id: Ib2a6c8943f96acc27e33f9f0b6f8e335e5ef84dd","shortMessageHtmlLink":"QueryResultRecorder: bump the max response size to 16M"}},{"before":"02afafa57041aa7e14e78af62551ffc6f3484596","after":"3469ec4acca88fcfb1e857ff83bcd0897a571e04","ref":"refs/heads/master","pushedAt":"2024-01-16T16:37:09.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"QueryResultRecorder: add support for blazegraph default collation key\n\nBlazegraph uses tertiary collation of an old ICU version. Use the same\ntechnique to sort and build result hashes so that we avoid few false\npositives.\n\nBug: T351819\nChange-Id: I900e8529e32344b3acfbf2bee0131da7313165ea","shortMessageHtmlLink":"QueryResultRecorder: add support for blazegraph default collation key"}},{"before":"bc83f126ba0ebb1b63727d7719d24dd295fddbe4","after":"02afafa57041aa7e14e78af62551ffc6f3484596","ref":"refs/heads/master","pushedAt":"2024-01-16T16:32:37.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Allow setting the max response size of the jetty httpclient\n\nThe 2mb default prevents loading larger payload.\n\nChange-Id: I1f781284b3c8b446c6ec7eaaaa0a727831e39ae6","shortMessageHtmlLink":"Allow setting the max response size of the jetty httpclient"}},{"before":"54ec4f3702bd624e347ca7c885b71d4da113e9ef","after":"bc83f126ba0ebb1b63727d7719d24dd295fddbe4","ref":"refs/heads/master","pushedAt":"2024-01-11T21:52:53.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Add a tool to execute SPARQL queries and record results\n\nQueryResultRecorder can be used in spark context as a UDF on a string\ncolumn to output a struct containing some metrics that might help\nto do some analysis on the results.\n\nIt does capture:\n- the result size\n- the response as an array of maps (if useful for fine-grained\n comparison)\n- a hash of the response\n- a hash that should be insensitive to the result order\n- whether the query failed or not\n- an error message in case it failed\n\nBug: T351819\nChange-Id: I27c48fed68bc481661d423928963400e154e2277","shortMessageHtmlLink":"Add a tool to execute SPARQL queries and record results"}},{"before":"3319b9c468c3baab68e7c8b464270ca449aa90a4","after":"54ec4f3702bd624e347ca7c885b71d4da113e9ef","ref":"refs/heads/master","pushedAt":"2024-01-11T15:41:50.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Update wikimedia-eventutilities to latest\n\nPin the guava version to 11.0.2 in rdf-spark-tools, this seems to be the\nversion we use in hadoop. Prefer the shaded artifact of eventutilities\nin rdf-spark-tools since it's unlikely to work with the old hadoop guava\nversion.\n\nChange-Id: I72087c541afd96eeb253f823ff8ef396c0ba9389","shortMessageHtmlLink":"Update wikimedia-eventutilities to latest"}},{"before":"0922c4c4dead1e4937f3c3b4afeb89eb79bd808d","after":"3319b9c468c3baab68e7c8b464270ca449aa90a4","ref":"refs/heads/master","pushedAt":"2024-01-09T16:27:48.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"HDFS to .ttl statement generator\n\nBug: T350106\nChange-Id: Iabda2033dd4512fe0b4325ff8e251218aebedfae","shortMessageHtmlLink":"HDFS to .ttl statement generator"}},{"before":"6ae30fa489822d46570f6c16be53840ef8b13903","after":"0922c4c4dead1e4937f3c3b4afeb89eb79bd808d","ref":"refs/heads/master","pushedAt":"2023-11-16T18:25:11.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Add support for emitter_id in error streams\n\nThe flink job can tag with an optional emitter_id its events sent to the\nerror streams.\nThe reconcile batch job can allow constructing the reconcile source tag\nusing the emitter_id.\n\nBug: T347515\nDepends-On: I321e8ad911d9f040fca12e213de931adfa47ec95\nChange-Id: I5a1fb71ecf712fa95215a0bf18ef618c157ff14c","shortMessageHtmlLink":"Add support for emitter_id in error streams"}},{"before":"bec5c44d9f06f337052a12d66cace1751a7a2b7f","after":"6ae30fa489822d46570f6c16be53840ef8b13903","ref":"refs/heads/master","pushedAt":"2023-11-16T17:37:10.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"EntityRevisionMapGenerator: reduce to 8 the number of ouput partition\n\nTotal dataset is around 800M.\n\nChange-Id: Ibc710d50ece84828acf1f1cb3931b832922dc357","shortMessageHtmlLink":"EntityRevisionMapGenerator: reduce to 8 the number of ouput partition"}},{"before":"cb1b60b55c093ff6d692477f3717ab7e1039bbf5","after":"bec5c44d9f06f337052a12d66cace1751a7a2b7f","ref":"refs/heads/master","pushedAt":"2023-11-16T14:57:22.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"[maven-release-plugin] prepare for next development iteration","shortMessageHtmlLink":"[maven-release-plugin] prepare for next development iteration"}},{"before":"f5c73e1fb6376f2898db0a76805682ceedaacb06","after":"cb1b60b55c093ff6d692477f3717ab7e1039bbf5","ref":"refs/heads/master","pushedAt":"2023-11-16T09:04:32.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Split the graph - more partitioning of wikibase_rdf\n\nBug: T347989\n\nGiven an HDFS Parquet partition with String cols:\n - subject\n - predicate\n - object\n - context\n\nAnd if those match the Wikidata dump ontology:\n - Creates new partitions 'scholarly_articles' and 'wikidata_main'\n - Keeps pertinent Value and Reference in both partitions\n\nChange-Id: Ia4c438ddf5b758e269b04e63619da05129ead4de","shortMessageHtmlLink":"Split the graph - more partitioning of wikibase_rdf"}},{"before":"c1956f1fa2e198a25a6acdd87295adf65748ef49","after":"f5c73e1fb6376f2898db0a76805682ceedaacb06","ref":"refs/heads/master","pushedAt":"2023-10-20T15:27:07.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Reflect the fact that kafka-jumbo1001.eqiad.wmnet has been decommissioned\n\nBug: T336044\nChange-Id: I72e528b0486c0dce808f6736a31f7711f6f20306","shortMessageHtmlLink":"Reflect the fact that kafka-jumbo1001.eqiad.wmnet has been decommissi…"}},{"before":"03ba3066f382cf7ca30a05a0e03c3a63d9fac4ad","after":"c1956f1fa2e198a25a6acdd87295adf65748ef49","ref":"refs/heads/master","pushedAt":"2023-10-20T14:29:44.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"[maven-release-plugin] prepare for next development iteration","shortMessageHtmlLink":"[maven-release-plugin] prepare for next development iteration"}},{"before":"c6974943384abcfc764fa5eaac73cc80193e90d1","after":"03ba3066f382cf7ca30a05a0e03c3a63d9fac4ad","ref":"refs/heads/master","pushedAt":"2023-10-19T15:07:22.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"Use LATEST offset reset strategy\n\nInitial run is apparently not the sole condition in which the offsets\nmight have to be reset.\nA problem occured few days after the switch to the new KafkaSource\nimplementation. The application failed to restart with\nNoOffsetForPartitionException: Undefined offset with no reset policy for\npartitions. So it appears that even if a consumer group is bound to a\ntopic kafka might forget its offset if there are no activity on it:\nhttps://issues.apache.org/jira/browse/KAFKA-4682.\nUse LATEST because it was the default used by the previous\nFlinkKafkaConsumer (that relies on default kafka auto.offset.reset).\nThis is not ideal as it's unclear if some messages might get lost on\nsuch idle stream when events are starting to flow again.\nUsing EARLIEST sounds dangerous as well esp if the offset retention is\nshorter than the topic retention, the stream after going idle might be\nreset to EARLIEST after the offsets are gone causing many events to be\nreprocessed.\nI'm not clear if there are ways to solve this problem without either:\n- have almost infinite retention of offsets (offset.retention.minutes)\n- upgrade to kafka >= 2.1\n- do not have idle streams for long periods (longer than\n offset.retention.minutes). Canary events everywhere?\n\nBug: T349147\nChange-Id: I89f81a8eb0ff2f8c97cd81a46a6f0c718c347f63","shortMessageHtmlLink":"Use LATEST offset reset strategy"}},{"before":"ee8d52bff0d0e3b2f67dbee6f91811f5971329bf","after":"c6974943384abcfc764fa5eaac73cc80193e90d1","ref":"refs/heads/master","pushedAt":"2023-10-16T17:11:05.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"reconcile_items.py: fix type annotation\n\nShould typing.Type not type.\n\nChange-Id: Idb9741d55b27b441d75c96df8512981152860865","shortMessageHtmlLink":"reconcile_items.py: fix type annotation"}},{"before":"cb1024f28a694dc234a472ca8067d7cdc599178a","after":"ee8d52bff0d0e3b2f67dbee6f91811f5971329bf","ref":"refs/heads/master","pushedAt":"2023-10-04T14:44:57.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"wmfgerrit","name":"Wikimedia Gerrit","path":"/wmfgerrit","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2539292?s=80&v=4"},"commit":{"message":"[maven-release-plugin] prepare for next development iteration","shortMessageHtmlLink":"[maven-release-plugin] prepare for next development iteration"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEQFkNKwA","startCursor":null,"endCursor":null}},"title":"Activity · wikimedia/wikidata-query-rdf"}