From aaa1ceba8dc8d2d30ff9a3db144327aa8720fb2f Mon Sep 17 00:00:00 2001 From: brijoobopanna Date: Tue, 18 Dec 2018 16:15:13 +0530 Subject: [PATCH] [HOTFIX] replace apache common log with carbondata log4j --- .../org/apache/carbondata/core/datamap/TableDataMap.java | 7 ++++--- .../indexstore/blockletindex/BlockletDataMapFactory.java | 3 --- .../apache/carbondata/core/util/BlockletDataMapUtil.java | 7 ++++--- .../carbondata/core/util/ObjectSerializationUtil.java | 8 +++++--- .../apache/carbondata/hadoop/api/CarbonInputFormat.java | 7 ++++--- .../carbondata/hadoop/api/CarbonTableInputFormat.java | 7 ++++--- .../carbondata/hadoop/api/CarbonTableOutputFormat.java | 7 ++++--- .../carbondata/hive/server/HiveEmbeddedServer2.java | 7 ++++--- .../loading/iterator/CarbonOutputIteratorWrapper.java | 7 ++++--- 9 files changed, 33 insertions(+), 27 deletions(-) diff --git a/core/src/main/java/org/apache/carbondata/core/datamap/TableDataMap.java b/core/src/main/java/org/apache/carbondata/core/datamap/TableDataMap.java index 4de7449cdba..86390e80926 100644 --- a/core/src/main/java/org/apache/carbondata/core/datamap/TableDataMap.java +++ b/core/src/main/java/org/apache/carbondata/core/datamap/TableDataMap.java @@ -29,6 +29,7 @@ import java.util.concurrent.TimeUnit; import org.apache.carbondata.common.annotations.InterfaceAudience; +import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datamap.dev.BlockletSerializer; import org.apache.carbondata.core.datamap.dev.DataMap; @@ -50,8 +51,7 @@ import org.apache.carbondata.events.OperationContext; import org.apache.carbondata.events.OperationEventListener; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.apache.log4j.Logger; /** * Index at the table level, user can add any number of DataMap for one table, by @@ -74,7 +74,8 @@ public final class TableDataMap extends OperationEventListener { private SegmentPropertiesFetcher segmentPropertiesFetcher; - private static final Log LOG = LogFactory.getLog(TableDataMap.class); + private static final Logger LOG = + LogServiceFactory.getLogService(TableDataMap.class.getName()); /** * It is called to initialize and load the required table datamap metadata. diff --git a/core/src/main/java/org/apache/carbondata/core/indexstore/blockletindex/BlockletDataMapFactory.java b/core/src/main/java/org/apache/carbondata/core/indexstore/blockletindex/BlockletDataMapFactory.java index 096a5e36c74..5892f78a6bd 100644 --- a/core/src/main/java/org/apache/carbondata/core/indexstore/blockletindex/BlockletDataMapFactory.java +++ b/core/src/main/java/org/apache/carbondata/core/indexstore/blockletindex/BlockletDataMapFactory.java @@ -54,8 +54,6 @@ import org.apache.carbondata.core.util.path.CarbonTablePath; import org.apache.carbondata.events.Event; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.LocatedFileStatus; import org.apache.hadoop.fs.Path; @@ -67,7 +65,6 @@ public class BlockletDataMapFactory extends CoarseGrainDataMapFactory implements BlockletDetailsFetcher, SegmentPropertiesFetcher, CacheableDataMap { - private static final Log LOG = LogFactory.getLog(BlockletDataMapFactory.class); private static final String NAME = "clustered.btree.blocklet"; /** * variable for cache level BLOCKLET diff --git a/core/src/main/java/org/apache/carbondata/core/util/BlockletDataMapUtil.java b/core/src/main/java/org/apache/carbondata/core/util/BlockletDataMapUtil.java index fbdeff87db4..b81bc75d3aa 100644 --- a/core/src/main/java/org/apache/carbondata/core/util/BlockletDataMapUtil.java +++ b/core/src/main/java/org/apache/carbondata/core/util/BlockletDataMapUtil.java @@ -34,6 +34,7 @@ import java.util.Set; import java.util.TreeMap; +import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datamap.Segment; import org.apache.carbondata.core.datastore.block.SegmentProperties; @@ -60,15 +61,15 @@ import org.apache.carbondata.core.scan.filter.resolver.FilterResolverIntf; import org.apache.carbondata.core.util.path.CarbonTablePath; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.PathFilter; +import org.apache.log4j.Logger; public class BlockletDataMapUtil { - private static final Log LOG = LogFactory.getLog(BlockletDataMapUtil.class); + private static final Logger LOG = + LogServiceFactory.getLogService(BlockletDataMapUtil.class.getName()); public static Map getBlockMetaInfoMap( TableBlockIndexUniqueIdentifierWrapper identifierWrapper, diff --git a/core/src/main/java/org/apache/carbondata/core/util/ObjectSerializationUtil.java b/core/src/main/java/org/apache/carbondata/core/util/ObjectSerializationUtil.java index 020787de681..9a4c02dae02 100644 --- a/core/src/main/java/org/apache/carbondata/core/util/ObjectSerializationUtil.java +++ b/core/src/main/java/org/apache/carbondata/core/util/ObjectSerializationUtil.java @@ -24,15 +24,17 @@ import java.util.zip.GZIPInputStream; import java.util.zip.GZIPOutputStream; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; +import org.apache.carbondata.common.logging.LogServiceFactory; + +import org.apache.log4j.Logger; /** * It provides methods to convert object to Base64 string and vice versa. */ public class ObjectSerializationUtil { - private static final Log LOG = LogFactory.getLog(ObjectSerializationUtil.class); + private static final Logger LOG = + LogServiceFactory.getLogService(ObjectSerializationUtil.class.getName()); /** * Convert object to Base64 String diff --git a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonInputFormat.java b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonInputFormat.java index 843f5643bd9..c35862a3e8f 100644 --- a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonInputFormat.java +++ b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonInputFormat.java @@ -23,6 +23,7 @@ import java.lang.reflect.Constructor; import java.util.*; +import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.constants.CarbonCommonConstantsInternal; import org.apache.carbondata.core.datamap.DataMapChooser; @@ -69,8 +70,6 @@ import org.apache.carbondata.hadoop.readsupport.impl.DictionaryDecodeReadSupport; import org.apache.commons.collections.CollectionUtils; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.LocalFileSystem; @@ -82,6 +81,7 @@ import org.apache.hadoop.mapreduce.lib.input.FileInputFormat; import org.apache.hadoop.mapreduce.lib.input.FileSplit; import org.apache.hadoop.mapreduce.security.TokenCache; +import org.apache.log4j.Logger; /** * Base class for carbondata input format, there are two input format implementations: @@ -100,7 +100,8 @@ public abstract class CarbonInputFormat extends FileInputFormat { "mapreduce.input.carboninputformat.validsegments"; // comma separated list of input files private static final String ALTER_PARTITION_ID = "mapreduce.input.carboninputformat.partitionid"; - private static final Log LOG = LogFactory.getLog(CarbonInputFormat.class); + private static final Logger LOG = + LogServiceFactory.getLogService(CarbonInputFormat.class.getName()); private static final String FILTER_PREDICATE = "mapreduce.input.carboninputformat.filter.predicate"; private static final String COLUMN_PROJECTION = "mapreduce.input.carboninputformat.projection"; diff --git a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableInputFormat.java b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableInputFormat.java index 86cbfec9686..ea1bcd3f1b3 100644 --- a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableInputFormat.java +++ b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableInputFormat.java @@ -27,6 +27,7 @@ import java.util.List; import java.util.Map; +import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.datamap.DataMapStoreManager; import org.apache.carbondata.core.datamap.Segment; import org.apache.carbondata.core.datamap.TableDataMap; @@ -60,8 +61,6 @@ import org.apache.carbondata.core.util.path.CarbonTablePath; import org.apache.carbondata.hadoop.CarbonInputSplit; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.BlockLocation; import org.apache.hadoop.fs.FileStatus; @@ -71,6 +70,7 @@ import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.mapreduce.JobContext; import org.apache.hadoop.mapreduce.lib.input.FileSplit; +import org.apache.log4j.Logger; /** * InputFormat for reading carbondata files with table level metadata support, @@ -86,7 +86,8 @@ public class CarbonTableInputFormat extends CarbonInputFormat { // comma separated list of input files public static final String INPUT_FILES = "mapreduce.input.carboninputformat.files"; private static final String ALTER_PARTITION_ID = "mapreduce.input.carboninputformat.partitionid"; - private static final Log LOG = LogFactory.getLog(CarbonTableInputFormat.class); + private static final Logger LOG = + LogServiceFactory.getLogService(CarbonTableInputFormat.class.getName()); private static final String CARBON_READ_SUPPORT = "mapreduce.input.carboninputformat.readsupport"; private static final String CARBON_CONVERTER = "mapreduce.input.carboninputformat.converter"; private static final String CARBON_TRANSACTIONAL_TABLE = diff --git a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableOutputFormat.java b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableOutputFormat.java index 16486d0c51b..88ba16c9896 100644 --- a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableOutputFormat.java +++ b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableOutputFormat.java @@ -25,6 +25,7 @@ import java.util.concurrent.Future; import java.util.concurrent.atomic.AtomicLong; +import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.constants.CarbonLoadOptionConstants; import org.apache.carbondata.core.datastore.compression.CompressorFactory; @@ -45,8 +46,6 @@ import org.apache.carbondata.processing.loading.model.CarbonLoadModel; import org.apache.commons.lang3.StringUtils; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.io.NullWritable; @@ -54,6 +53,7 @@ import org.apache.hadoop.mapreduce.RecordWriter; import org.apache.hadoop.mapreduce.TaskAttemptContext; import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat; +import org.apache.log4j.Logger; /** * This is table level output format which writes the data to store in new segment. Each load @@ -115,7 +115,8 @@ public class CarbonTableOutputFormat extends FileOutputFormat { - private static final Log LOG = LogFactory.getLog(CarbonOutputIteratorWrapper.class); + private static final Logger LOG = + LogServiceFactory.getLogService(CarbonOutputIteratorWrapper.class.getName()); private boolean close;