From bd166b3bd655a287e2408c827e178a07bd6e711d Mon Sep 17 00:00:00 2001 From: Radhakrishnan Venkataramani Date: Mon, 15 Jun 2020 18:25:38 -0700 Subject: [PATCH] Move quantized embedding ops to OSS. Summary: Move quantized embedding ops to OSS. This is the first step in supporting Graph Mode Quantization for EmbeddingBag. The next step would be to look into packing for embeddings as well as graph code rewrite to support post training dynamic quantization. Differential Revision: D21949828 fbshipit-source-id: 112ace54f570611e8634092da43a63baa7651427 --- torch_glow/src/PyTorchModelLoader.cpp | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/torch_glow/src/PyTorchModelLoader.cpp b/torch_glow/src/PyTorchModelLoader.cpp index 9e46466a65..498191250a 100644 --- a/torch_glow/src/PyTorchModelLoader.cpp +++ b/torch_glow/src/PyTorchModelLoader.cpp @@ -713,8 +713,8 @@ struct EmbeddingBagInputs { }; }; -/// Indexes used for fb::embedding_bag_byte_rowwise_offsets and -/// fb::embedding_bag_4bit_rowwise_offsets inputs. +/// Indexes used for quantized::embedding_bag_byte_rowwise_offsets and +/// quantized::embedding_bag_4bit_rowwise_offsets inputs. struct EmbeddingBagByteRowwiseOffsetsInputs { enum { weight, @@ -805,9 +805,9 @@ PyTorchModelLoader::buildSymbolsMapping() { {{"aten::topk"}, &PyTorchModelLoader::loadTopK}, {{"prim::ConstantChunk"}, &PyTorchModelLoader::loadConstantChunk}, {{"aten::embedding_bag"}, &PyTorchModelLoader::loadEmbeddingBag}, - {{"fb::embedding_bag_byte_rowwise_offsets"}, + {{"quantized::embedding_bag_byte_rowwise_offsets"}, &PyTorchModelLoader::loadEmbeddingBagByteRowwiseOffsets}, - {{"fb::embedding_bag_4bit_rowwise_offsets"}, + {{"quantized::embedding_bag_4bit_rowwise_offsets"}, &PyTorchModelLoader::loadEmbeddingBag4BitRowwiseOffsets}, });