From d36df793ffd1f1982803e4c1455d60377859dd44 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Thu, 23 Feb 2023 16:13:21 -0800 Subject: [PATCH 01/24] Bump to TF 2.12.0 rc0 Signed-off-by: Yong Tang --- WORKSPACE | 6 +++--- tensorflow_io/python/ops/version_ops.py | 2 +- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/WORKSPACE b/WORKSPACE index 66edd1c19..f27922bb0 100644 --- a/WORKSPACE +++ b/WORKSPACE @@ -115,10 +115,10 @@ pip_install() http_archive( name = "org_tensorflow", - sha256 = "99c732b92b1b37fc243a559e02f9aef5671771e272758aa4aec7f34dc92dac48", - strip_prefix = "tensorflow-2.11.0", + sha256 = "0090d8c11a346f078acdf16ee83a64adf96c43eb1e34b971b90bb9bc5303a60b", + strip_prefix = "tensorflow-2.12.0-rc0", urls = [ - "https://github.com/tensorflow/tensorflow/archive/refs/tags/v2.11.0.tar.gz", + "https://github.com/tensorflow/tensorflow/archive/refs/tags/v2.12.0-rc0.tar.gz", ], ) diff --git a/tensorflow_io/python/ops/version_ops.py b/tensorflow_io/python/ops/version_ops.py index 9580a4bbc..c50c2e8d5 100644 --- a/tensorflow_io/python/ops/version_ops.py +++ b/tensorflow_io/python/ops/version_ops.py @@ -15,4 +15,4 @@ """version_ops""" version = "0.31.0" -require = "tensorflow>=2.11.0,<2.12.0" +require = "tensorflow>=2.12.0rc0,<2.13.0" From 0d699070080c2e12682161b01f9156cfa69b0e44 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Thu, 23 Feb 2023 16:13:45 -0800 Subject: [PATCH 02/24] Try Signed-off-by: Yong Tang --- .github/workflows/build.yml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/.github/workflows/build.yml b/.github/workflows/build.yml index 016cdeb98..c2ad22240 100644 --- a/.github/workflows/build.yml +++ b/.github/workflows/build.yml @@ -2,8 +2,8 @@ name: GitHub CI on: push: - branches: - - master + # branches: + # - master pull_request: branches: - master From 29410a58d2e91d6c34859cab5184f56c641bb40c Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sat, 25 Feb 2023 12:58:15 -0800 Subject: [PATCH 03/24] Change to OkStatus Signed-off-by: Yong Tang --- .../core/filesystems/az/az_filesystem.cc | 8 +-- .../core/filesystems/http/http_filesystem.cc | 2 +- .../core/filesystems/oss/oss_filesystem.cc | 54 +++++++++---------- .../core/kernels/arrow/arrow_dataset_ops.cc | 48 ++++++++--------- .../core/kernels/arrow/arrow_kernels.cc | 22 ++++---- .../kernels/arrow/arrow_stream_client_unix.cc | 4 +- .../arrow/arrow_stream_client_windows.cc | 4 +- .../core/kernels/arrow/arrow_util.cc | 42 +++++++-------- tensorflow_io/core/kernels/audio_kernels.cc | 6 +-- .../core/kernels/audio_video_flac_kernels.cc | 8 +-- .../core/kernels/audio_video_mp3_kernels.cc | 8 +-- .../core/kernels/audio_video_mp4_kernels.cc | 20 +++---- .../core/kernels/audio_video_ogg_kernels.cc | 10 ++-- .../core/kernels/audio_video_wav_kernels.cc | 12 ++--- .../avro/avro_record_dataset_kernels.cc | 16 +++--- .../core/kernels/avro/parse_avro_kernels.cc | 6 +-- .../core/kernels/avro/utils/avro_parser.cc | 28 +++++----- .../kernels/avro/utils/avro_parser_tree.cc | 30 +++++------ .../kernels/avro/utils/avro_record_reader.cc | 2 +- .../kernels/avro/utils/parse_avro_attrs.cc | 2 +- .../core/kernels/avro/utils/value_buffer.cc | 8 +-- tensorflow_io/core/kernels/avro_kernels.cc | 14 ++--- .../kernels/bigquery/bigquery_dataset_op.cc | 4 +- .../core/kernels/bigquery/bigquery_kernels.cc | 2 +- .../core/kernels/bigquery/bigquery_lib.cc | 4 +- .../bigtable/bigtable_dataset_kernel.cc | 8 +-- .../core/kernels/bigtable/serialization.cc | 2 +- tensorflow_io/core/kernels/csv_kernels.cc | 12 ++--- .../core/kernels/elasticsearch_kernels.cc | 14 ++--- tensorflow_io/core/kernels/ffmpeg_kernels.cc | 46 ++++++++-------- .../core/kernels/ffmpeg_kernels_deprecated.cc | 54 +++++++++---------- tensorflow_io/core/kernels/file_kernels.cc | 10 ++-- tensorflow_io/core/kernels/grpc_kernels.cc | 10 ++-- .../gce_memcached_server_list_provider.cc | 4 +- .../memcached_file_block_cache.cc | 16 +++--- tensorflow_io/core/kernels/hdf5_kernels.cc | 14 ++--- .../core/kernels/image_dicom_kernels.cc | 4 +- .../core/kernels/image_font_kernels.cc | 2 +- tensorflow_io/core/kernels/io_optimization.cc | 2 +- tensorflow_io/core/kernels/json_kernels.cc | 16 +++--- tensorflow_io/core/kernels/kafka_kernels.cc | 34 ++++++------ .../core/kernels/kafka_kernels_deprecated.cc | 24 ++++----- tensorflow_io/core/kernels/kinesis_kernels.cc | 10 ++-- tensorflow_io/core/kernels/lmdb_kernels.cc | 12 ++--- tensorflow_io/core/kernels/mongodb_kernels.cc | 20 +++---- tensorflow_io/core/kernels/numpy_kernels.cc | 8 +-- tensorflow_io/core/kernels/orc/orc_kernels.cc | 12 ++--- tensorflow_io/core/kernels/parquet_kernels.cc | 14 ++--- tensorflow_io/core/kernels/pcap_kernels.cc | 12 ++--- .../core/kernels/prometheus_kernels.cc | 12 ++--- tensorflow_io/core/kernels/pubsub_kernels.cc | 12 ++--- tensorflow_io/core/kernels/pulsar_kernel.cc | 16 +++--- .../core/kernels/serialization_kernels.cc | 14 ++--- tensorflow_io/core/kernels/sql_kernels.cc | 10 ++-- .../kernels/tests/bigquery_test_client_op.cc | 2 +- tensorflow_io/core/kernels/text_kernels.cc | 4 +- tensorflow_io/core/kernels/text_sequence.cc | 4 +- tensorflow_io/core/kernels/video_kernels.cc | 8 +-- tensorflow_io/core/ops/arrow_ops.cc | 16 +++--- tensorflow_io/core/ops/audio_ops.cc | 38 ++++++------- tensorflow_io/core/ops/avro_ops.cc | 16 +++--- tensorflow_io/core/ops/bigquery_ops.cc | 2 +- tensorflow_io/core/ops/bigtable_ops.cc | 2 +- tensorflow_io/core/ops/core_ops.cc | 4 +- tensorflow_io/core/ops/elasticsearch_ops.cc | 4 +- tensorflow_io/core/ops/ffmpeg_ops.cc | 16 +++--- tensorflow_io/core/ops/file_ops.cc | 4 +- tensorflow_io/core/ops/genome_ops.cc | 2 +- tensorflow_io/core/ops/grpc_ops.cc | 4 +- tensorflow_io/core/ops/hdf5_ops.cc | 6 +-- tensorflow_io/core/ops/image_ops.cc | 32 +++++------ tensorflow_io/core/ops/json_ops.cc | 6 +-- tensorflow_io/core/ops/kafka_ops.cc | 14 ++--- .../core/ops/kafka_ops_deprecated.cc | 10 ++-- tensorflow_io/core/ops/kinesis_ops.cc | 4 +- tensorflow_io/core/ops/libsvm_ops.cc | 2 +- tensorflow_io/core/ops/lmdb_ops.cc | 12 ++--- tensorflow_io/core/ops/mongodb_ops.cc | 2 +- tensorflow_io/core/ops/numpy_ops.cc | 6 +-- tensorflow_io/core/ops/obj_ops.cc | 2 +- tensorflow_io/core/ops/operation_ops.cc | 4 +- tensorflow_io/core/ops/orc_ops.cc | 6 +-- tensorflow_io/core/ops/parquet_ops.cc | 6 +-- tensorflow_io/core/ops/pcap_ops.cc | 4 +- tensorflow_io/core/ops/prometheus_ops.cc | 8 +-- tensorflow_io/core/ops/pubsub_ops.cc | 4 +- tensorflow_io/core/ops/pulsar_ops.cc | 6 +-- tensorflow_io/core/ops/serialization_ops.cc | 6 +-- tensorflow_io/core/ops/sql_ops.cc | 4 +- tensorflow_io/core/ops/text_ops.cc | 12 ++--- tensorflow_io/core/ops/video_ops.cc | 8 +-- 91 files changed, 539 insertions(+), 539 deletions(-) diff --git a/tensorflow_io/core/filesystems/az/az_filesystem.cc b/tensorflow_io/core/filesystems/az/az_filesystem.cc index daacf5842..b2609e3e3 100644 --- a/tensorflow_io/core/filesystems/az/az_filesystem.cc +++ b/tensorflow_io/core/filesystems/az/az_filesystem.cc @@ -59,7 +59,7 @@ bool GetTmpFilename(std::string* filename) { } #endif *filename = buffer; - // return Status::OK(); + // return OkStatus(); return true; } @@ -271,7 +271,7 @@ class AzBlobRandomAccessFile { TF_Status* status) const { TF_VLog(1, "ReadFileFromAz az://%s/%s/%s from %u for n: %u\n", account_.c_str(), container_.c_str(), object_.c_str(), offset, n); - // If n == 0, then return Status::OK() + // If n == 0, then return OkStatus() // otherwise, if bytes_read < n then return OutofRange if (n == 0) { TF_SetStatus(status, TF_OK, ""); @@ -457,7 +457,7 @@ Status GetMatchingPaths(const std::string& pattern, std::vector* re return Env::Default()->MatchPath(full_path, pattern); }); - return Status::OK(); + return OkStatus(); } #endif @@ -876,7 +876,7 @@ static bool IsDirectory(const TF_Filesystem* filesystem, const char* path, // bool is_account; // TF_RETURN_IF_ERROR(AccountExists(account, &is_account, blob_client)); - // return is_account ? Status::OK() + // return is_account ? OkStatus() // : errors::NotFound("The specified account az://", // account, " was not found."); } diff --git a/tensorflow_io/core/filesystems/http/http_filesystem.cc b/tensorflow_io/core/filesystems/http/http_filesystem.cc index 26a652a15..7c1a0108c 100644 --- a/tensorflow_io/core/filesystems/http/http_filesystem.cc +++ b/tensorflow_io/core/filesystems/http/http_filesystem.cc @@ -549,7 +549,7 @@ class HTTPRandomAccessFile { ~HTTPRandomAccessFile() {} int64_t Read(uint64_t offset, size_t n, char* buffer, TF_Status* status) const { - // If n == 0, then return Status::OK() + // If n == 0, then return OkStatus() // otherwise, if bytes_read < n then return OutofRange if (n == 0) { TF_SetStatus(status, TF_OK, ""); diff --git a/tensorflow_io/core/filesystems/oss/oss_filesystem.cc b/tensorflow_io/core/filesystems/oss/oss_filesystem.cc index ce59dc305..e682e2471 100644 --- a/tensorflow_io/core/filesystems/oss/oss_filesystem.cc +++ b/tensorflow_io/core/filesystems/oss/oss_filesystem.cc @@ -70,7 +70,7 @@ Status oss_initialize() { return errors::Internal("can not init OSS connection"); } - return Status::OK(); + return OkStatus(); } void oss_error_message(aos_status_s* status, std::string* msg) { @@ -195,7 +195,7 @@ class OSSRandomAccessFile : public RandomAccessFile { " bytes were read out of ", n, " bytes requested."); } - return Status::OK(); + return OkStatus(); } private: @@ -251,7 +251,7 @@ class OSSRandomAccessFile : public RandomAccessFile { pos += size; } buffer_size_ = pos; - return Status::OK(); + return OkStatus(); } else { string msg; oss_error_message(s, &msg); @@ -316,7 +316,7 @@ class OSSWritableFile : public WritableFile { aos_buf_t* tmp_buf = aos_create_buf(pool_, data.size() + 1); aos_buf_append_string(pool_, tmp_buf, data.data(), data.size()); aos_list_add_tail(&tmp_buf->node, &buffer_); - return Status::OK(); + return OkStatus(); } Status Close() override { @@ -369,7 +369,7 @@ class OSSWritableFile : public WritableFile { } is_closed_ = true; - return Status::OK(); + return OkStatus(); } Status Flush() override { @@ -380,7 +380,7 @@ class OSSWritableFile : public WritableFile { TF_RETURN_IF_ERROR(_FlushInternal()); } - return Status::OK(); + return OkStatus(); } Status Sync() override { return Flush(); } @@ -434,7 +434,7 @@ class OSSWritableFile : public WritableFile { upload_id_ = uploadId.data; } - return Status::OK(); + return OkStatus(); } Status _FlushInternal() { @@ -463,7 +463,7 @@ class OSSWritableFile : public WritableFile { ReleaseAprPool(); InitAprPool(); } - return Status::OK(); + return OkStatus(); } const size_t CurrentBufferLength() { return aos_buf_list_len(&buffer_); } @@ -473,7 +473,7 @@ class OSSWritableFile : public WritableFile { return errors::Internal("Already closed."); } - return Status::OK(); + return OkStatus(); } std::string shost; @@ -564,7 +564,7 @@ Status OSSFileSystem::_ParseOSSURIPath(const StringPiece fname, VLOG(1) << "bucket: " << bucket << ",access_id: " << access_id << ",access_key: " << access_key << ",host: " << host; - return Status::OK(); + return OkStatus(); } Status OSSFileSystem::NewRandomAccessFile( @@ -581,7 +581,7 @@ Status OSSFileSystem::NewRandomAccessFile( result->reset(new OSSRandomAccessFile(host, access_id, access_key, bucket, object, read_ahead_bytes_, stat.length)); - return Status::OK(); + return OkStatus(); } Status OSSFileSystem::NewWritableFile(const std::string& fname, @@ -594,7 +594,7 @@ Status OSSFileSystem::NewWritableFile(const std::string& fname, result->reset(new OSSWritableFile(host, access_id, access_key, bucket, object, upload_part_bytes_)); - return Status::OK(); + return OkStatus(); } Status OSSFileSystem::NewAppendableFile(const std::string& fname, @@ -617,13 +617,13 @@ Status OSSFileSystem::NewReadOnlyMemoryRegionFromFile( TF_RETURN_IF_ERROR(file->Read(0, size, &piece, data.get())); result->reset(new OSSReadOnlyMemoryRegion(std::move(data), size)); - return Status::OK(); + return OkStatus(); } Status OSSFileSystem::FileExists(const std::string& fname) { TF_FileStatistics stat; if (Stat(fname, &stat).ok()) { - return Status::OK(); + return OkStatus(); } else { return errors::NotFound(fname, " does not exists"); } @@ -713,7 +713,7 @@ Status OSSFileSystem::_ListObjects( aos_list_init(¶ms->common_prefix_list); } while (params->truncated == AOS_TRUE && return_all); - return Status::OK(); + return OkStatus(); } Status OSSFileSystem::_StatInternal(aos_pool_t* pool, @@ -743,7 +743,7 @@ Status OSSFileSystem::_StatInternal(aos_pool_t* pool, s = _ListObjects(pool, options, bucket, object, &listing, true, false, false, true, 10); - if (s == Status::OK() && !listing.empty()) { + if (s == OkStatus() && !listing.empty()) { if (str_util::EndsWith(object, "/")) { stat->is_directory = true; } @@ -773,7 +773,7 @@ Status OSSFileSystem::_RetrieveObjectMetadata( if (object.empty()) { // root always exists stat->is_directory = true; stat->length = 0; - return Status::OK(); + return OkStatus(); } aos_str_set(&oss_bucket, bucket.c_str()); @@ -811,7 +811,7 @@ Status OSSFileSystem::_RetrieveObjectMetadata( stat->is_directory = false; } - return Status::OK(); + return OkStatus(); } else { string msg; oss_error_message(status, &msg); @@ -869,7 +869,7 @@ Status OSSFileSystem::_DeleteObjectInternal( return errors::Internal("delete failed: ", object, " errMsg: ", msg); } - return Status::OK(); + return OkStatus(); } Status OSSFileSystem::DeleteFile(const std::string& fname) { @@ -915,7 +915,7 @@ Status OSSFileSystem::CreateDir(const std::string& dirname) { } TF_RETURN_IF_ERROR(_CreateDirInternal(pool, ossOptions, bucket, object)); - return Status::OK(); + return OkStatus(); } Status OSSFileSystem::RecursivelyCreateDir(const string& dirname) { @@ -946,7 +946,7 @@ Status OSSFileSystem::RecursivelyCreateDir(const string& dirname) { } } - return Status::OK(); + return OkStatus(); } Status OSSFileSystem::_CreateDirInternal(aos_pool_t* pool, @@ -960,7 +960,7 @@ Status OSSFileSystem::_CreateDirInternal(aos_pool_t* pool, return errors::AlreadyExists("object already exists as a file: ", dirname); } else { - return Status::OK(); + return OkStatus(); } } std::string object = dirname; @@ -988,7 +988,7 @@ Status OSSFileSystem::_CreateDirInternal(aos_pool_t* pool, &resp_headers); if (aos_status_is_ok(s)) { - return Status::OK(); + return OkStatus(); } else { string msg; oss_error_message(s, &msg); @@ -1029,7 +1029,7 @@ Status OSSFileSystem::GetFileSize(const std::string& fname, uint64* file_size) { TF_FileStatistics stat; TF_RETURN_IF_ERROR(Stat(fname, &stat)); *file_size = stat.length; - return Status::OK(); + return OkStatus(); } Status OSSFileSystem::RenameFile(const std::string& src, @@ -1237,7 +1237,7 @@ Status OSSFileSystem::IsDirectory(const std::string& fname) { TF_RETURN_IF_ERROR(Stat(fname, &stat)); return stat.is_directory - ? Status::OK() + ? OkStatus() : errors::FailedPrecondition(fname + " is not a directory"); } @@ -1298,7 +1298,7 @@ Status OSSFileSystem::DeleteRecursively(const std::string& dirname, return _DeleteObjectInternal(oss_options, bucket, object.append(kDelim)); } } - return Status::OK(); + return OkStatus(); } Status OSSFileSystem::CopyFile(const string& src, const string& target) { @@ -1345,7 +1345,7 @@ Status OSSFileSystem::CopyFile(const string& src, const string& target) { return errors::Internal("copy ", src, " to ", target, " failed, errMsg: ", msg); } - return Status::OK(); + return OkStatus(); } void ToTF_Status(const ::tensorflow::Status& s, TF_Status* status) { diff --git a/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc b/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc index 7716391a9..8e6fdf278 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc @@ -49,7 +49,7 @@ Status GetBatchModeStr(ArrowBatchMode batch_mode, tstring* batch_mode_str) { return errors::Internal("Unsupported batch mode: " + std::to_string(batch_mode)); } - return Status::OK(); + return OkStatus(); } Status GetBatchMode(string batch_mode_str, ArrowBatchMode* batch_mode) { @@ -62,7 +62,7 @@ Status GetBatchMode(string batch_mode_str, ArrowBatchMode* batch_mode) { } else { return errors::Internal("Unsupported batch mode: " + batch_mode_str); } - return Status::OK(); + return OkStatus(); } // Base class for defining a Dataset over Arrow record batches with an @@ -200,7 +200,7 @@ class ArrowDatasetBase : public DatasetBase { } while (!(have_result || *end_of_sequence)); - return Status::OK(); + return OkStatus(); } private: @@ -213,7 +213,7 @@ class ArrowDatasetBase : public DatasetBase { // If only one partial batch, can just move to output if (partials.size() == 1) { *out_tensors = std::move(*partials.at(0).get()); - return Status::OK(); + return OkStatus(); } // Copy all partial tensors to a single output tensor @@ -238,7 +238,7 @@ class ArrowDatasetBase : public DatasetBase { } batch_index += partial_batch_size; } - return Status::OK(); + return OkStatus(); } template @@ -249,7 +249,7 @@ class ArrowDatasetBase : public DatasetBase { parent->flat_outer_dims().chip(index + i, 0) = element.flat_outer_dims().chip(i, 0); } - return Status::OK(); + return OkStatus(); } Status CopyElementsToParent(const Tensor& element, Tensor* parent, @@ -294,7 +294,7 @@ class ArrowDatasetBase : public DatasetBase { virtual Status NextStreamLocked(Env* env) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) { current_batch_ = nullptr; current_row_idx_ = 0; - return Status::OK(); + return OkStatus(); } // Reset the Arrow record batch consumer when done with batches. @@ -312,7 +312,7 @@ class ArrowDatasetBase : public DatasetBase { std::shared_ptr arr = batch->column(col); TF_RETURN_IF_ERROR(ArrowUtil::CheckArrayType(arr->type(), dt)); } - return Status::OK(); + return OkStatus(); } mutex mu_; @@ -430,7 +430,7 @@ class ArrowZeroCopyDatasetOp : public ArrowOpKernelBase { buffer_ptr_(buffer_ptr), buffer_size_(buffer_size) {} - Status CheckExternalState() const override { return Status::OK(); } + Status CheckExternalState() const override { return OkStatus(); } string DebugString() const override { return "ArrowZeroCopyDatasetOp::Dataset"; @@ -456,7 +456,7 @@ class ArrowZeroCopyDatasetOp : public ArrowOpKernelBase { TF_RETURN_IF_ERROR(b->AddScalar(batch_mode_str, &batch_mode)); TF_RETURN_IF_ERROR(b->AddDataset( this, {buffer, size, columns, batch_size, batch_mode}, output)); - return Status::OK(); + return OkStatus(); } std::unique_ptr MakeIteratorInternal( @@ -490,7 +490,7 @@ class ArrowZeroCopyDatasetOp : public ArrowOpKernelBase { current_batch_ = std::move(result).ValueUnsafe(); TF_RETURN_IF_ERROR(CheckBatchColumnTypes(current_batch_)); } - return Status::OK(); + return OkStatus(); } Status NextStreamLocked(Env* env) @@ -502,7 +502,7 @@ class ArrowZeroCopyDatasetOp : public ArrowOpKernelBase { CHECK_ARROW(result.status()); current_batch_ = std::move(result).ValueUnsafe(); } - return Status::OK(); + return OkStatus(); } void ResetStreamsLocked() TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { @@ -564,7 +564,7 @@ class ArrowSerializedDatasetOp : public ArrowOpKernelBase { return "ArrowSerializedDatasetOp::Dataset"; } - Status CheckExternalState() const override { return Status::OK(); } + Status CheckExternalState() const override { return OkStatus(); } protected: Status AsGraphDefInternal(SerializationContext* ctx, @@ -591,7 +591,7 @@ class ArrowSerializedDatasetOp : public ArrowOpKernelBase { TF_RETURN_IF_ERROR(b->AddScalar(batch_mode_str, &batch_mode)); TF_RETURN_IF_ERROR(b->AddDataset( this, {batches, columns, batch_size, batch_mode}, output)); - return Status::OK(); + return OkStatus(); } std::unique_ptr MakeIteratorInternal( @@ -622,7 +622,7 @@ class ArrowSerializedDatasetOp : public ArrowOpKernelBase { current_batch_ = std::move(result).ValueUnsafe(); TF_RETURN_IF_ERROR(CheckBatchColumnTypes(current_batch_)); } - return Status::OK(); + return OkStatus(); } Status NextStreamLocked(Env* env) @@ -633,7 +633,7 @@ class ArrowSerializedDatasetOp : public ArrowOpKernelBase { CHECK_ARROW(result.status()); current_batch_ = std::move(result).ValueUnsafe(); } - return Status::OK(); + return OkStatus(); } void ResetStreamsLocked() TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { @@ -697,7 +697,7 @@ class ArrowFeatherDatasetOp : public ArrowOpKernelBase { return "ArrowFeatherDatasetOp::Dataset"; } - Status CheckExternalState() const override { return Status::OK(); } + Status CheckExternalState() const override { return OkStatus(); } protected: Status AsGraphDefInternal(SerializationContext* ctx, @@ -715,7 +715,7 @@ class ArrowFeatherDatasetOp : public ArrowOpKernelBase { TF_RETURN_IF_ERROR(b->AddScalar(batch_mode_str, &batch_mode)); TF_RETURN_IF_ERROR(b->AddDataset( this, {filenames, columns, batch_size, batch_mode}, output)); - return Status::OK(); + return OkStatus(); } std::unique_ptr MakeIteratorInternal( @@ -769,7 +769,7 @@ class ArrowFeatherDatasetOp : public ArrowOpKernelBase { record_batches_.push_back(batch); CHECK_ARROW(tr.ReadNext(&batch)); } - return Status::OK(); + return OkStatus(); } Status NextStreamLocked(Env* env) @@ -782,7 +782,7 @@ class ArrowFeatherDatasetOp : public ArrowOpKernelBase { record_batches_.clear(); return SetupStreamsLocked(env); } - return Status::OK(); + return OkStatus(); } void ResetStreamsLocked() TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { @@ -847,7 +847,7 @@ class ArrowStreamDatasetOp : public ArrowOpKernelBase { return "ArrowStreamDatasetOp::Dataset"; } - Status CheckExternalState() const override { return Status::OK(); } + Status CheckExternalState() const override { return OkStatus(); } protected: Status AsGraphDefInternal(SerializationContext* ctx, @@ -865,7 +865,7 @@ class ArrowStreamDatasetOp : public ArrowOpKernelBase { TF_RETURN_IF_ERROR(b->AddScalar(batch_mode_str, &batch_mode)); TF_RETURN_IF_ERROR(b->AddDataset( this, {endpoints, columns, batch_size, batch_mode}, output)); - return Status::OK(); + return OkStatus(); } std::unique_ptr MakeIteratorInternal( @@ -906,7 +906,7 @@ class ArrowStreamDatasetOp : public ArrowOpKernelBase { reader_ = std::move(result).ValueUnsafe(); CHECK_ARROW(reader_->ReadNext(¤t_batch_)); TF_RETURN_IF_ERROR(CheckBatchColumnTypes(current_batch_)); - return Status::OK(); + return OkStatus(); } Status NextStreamLocked(Env* env) @@ -918,7 +918,7 @@ class ArrowStreamDatasetOp : public ArrowOpKernelBase { reader_.reset(); SetupStreamsLocked(env); } - return Status::OK(); + return OkStatus(); } void ResetStreamsLocked() TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { diff --git a/tensorflow_io/core/kernels/arrow/arrow_kernels.cc b/tensorflow_io/core/kernels/arrow/arrow_kernels.cc index 6641f06a0..1ed8f3bc5 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_kernels.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_kernels.cc @@ -49,7 +49,7 @@ class ArrowReadableResource : public ArrowReadableResourceBase { Status Init(const std::shared_ptr& table) override { mutex_lock l(mu_); table_ = table; - return Status::OK(); + return OkStatus(); } int32 GetColumnIndex(const string& column_name) override { @@ -76,7 +76,7 @@ class ArrowReadableResource : public ArrowReadableResourceBase { dims[0] = table_->num_rows(); *shape = TensorShape(dims); - return Status::OK(); + return OkStatus(); } Status Read(int64 start, int64 stop, int32 column_index, @@ -96,7 +96,7 @@ class ArrowReadableResource : public ArrowReadableResourceBase { // Column is empty if (chunked_arr->num_chunks() == 0) { - return Status::OK(); + return OkStatus(); } // Convert the array else if (chunked_arr->num_chunks() == 1) { @@ -118,7 +118,7 @@ class ArrowReadableResource : public ArrowReadableResourceBase { } } - return Status::OK(); + return OkStatus(); } string DebugString() const override { @@ -252,7 +252,7 @@ class ArrowReadableFromMemoryInitOp Status CreateResource(ArrowReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new ArrowReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -608,14 +608,14 @@ class FeatherReadable : public IOReadableInterface { columns_index_[schema->field(i)->name()] = i; } - return Status::OK(); + return OkStatus(); } Status Components(std::vector* components) override { components->clear(); for (size_t i = 0; i < columns_.size(); i++) { components->push_back(columns_[i]); } - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, PartialTensorShape* shape, DataType* dtype, bool label) override { @@ -625,7 +625,7 @@ class FeatherReadable : public IOReadableInterface { int64 column_index = columns_index_[component]; *shape = shapes_[column_index]; *dtype = dtypes_[column_index]; - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, const string& component, @@ -637,7 +637,7 @@ class FeatherReadable : public IOReadableInterface { (*record_read) = 0; if (start >= shapes_[column_index].dim_size(0)) { - return Status::OK(); + return OkStatus(); } int64 element_start = start < shapes_[column_index].dim_size(0) ? start @@ -650,7 +650,7 @@ class FeatherReadable : public IOReadableInterface { return errors::InvalidArgument("dataset selection is out of boundary"); } if (element_start == element_stop) { - return Status::OK(); + return OkStatus(); } if (feather_file_.get() == nullptr) { @@ -738,7 +738,7 @@ class FeatherReadable : public IOReadableInterface { DataTypeString(value->dtype())); } (*record_read) = element_stop - element_start; - return Status::OK(); + return OkStatus(); } string DebugString() const override { diff --git a/tensorflow_io/core/kernels/arrow/arrow_stream_client_unix.cc b/tensorflow_io/core/kernels/arrow/arrow_stream_client_unix.cc index bc171a6cb..78cd85465 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_stream_client_unix.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_stream_client_unix.cc @@ -95,7 +95,7 @@ arrow::Status ArrowStreamClient::Connect() { socket_family); } - return arrow::Status::OK(); + return arrow::OkStatus(); } arrow::Status ArrowStreamClient::Close() { @@ -106,7 +106,7 @@ arrow::Status ArrowStreamClient::Close() { return arrow::Status::IOError("Failed to correctly close connection"); } - return arrow::Status::OK(); + return arrow::OkStatus(); } bool ArrowStreamClient::closed() const { return sock_ == -1; } diff --git a/tensorflow_io/core/kernels/arrow/arrow_stream_client_windows.cc b/tensorflow_io/core/kernels/arrow/arrow_stream_client_windows.cc index 64c465205..6af6c91b6 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_stream_client_windows.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_stream_client_windows.cc @@ -115,7 +115,7 @@ arrow::Status ArrowStreamClient::Connect() { return arrow::Status::IOError("Unable to connect to server"); } - return arrow::Status::OK(); + return arrow::OkStatus(); } arrow::Status ArrowStreamClient::Close() { @@ -128,7 +128,7 @@ arrow::Status ArrowStreamClient::Close() { std::to_string(WSAGetLastError())); } - return arrow::Status::OK(); + return arrow::OkStatus(); } bool ArrowStreamClient::closed() const { return sock_ == -1; } diff --git a/tensorflow_io/core/kernels/arrow/arrow_util.cc b/tensorflow_io/core/kernels/arrow/arrow_util.cc index 462d0642a..ba24f2a73 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_util.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_util.cc @@ -31,11 +31,11 @@ Status GetTensorFlowType(std::shared_ptr<::arrow::DataType> dtype, ::tensorflow::DataType* out) { if (dtype->id() == ::arrow::Type::STRING) { *out = ::tensorflow::DT_STRING; - return Status::OK(); + return OkStatus(); } if (dtype->id() == ::arrow::Type::BINARY) { *out = ::tensorflow::DT_STRING; - return Status::OK(); + return OkStatus(); } ::arrow::Status status = ::arrow::adapters::tensorflow::GetTensorFlowType(dtype, out); @@ -43,14 +43,14 @@ Status GetTensorFlowType(std::shared_ptr<::arrow::DataType> dtype, return errors::InvalidArgument("arrow data type ", dtype->name(), " is not supported: ", status); } - return Status::OK(); + return OkStatus(); } Status GetArrowType(::tensorflow::DataType dtype, std::shared_ptr<::arrow::DataType>* out) { if (dtype == ::tensorflow::DT_STRING) { *out = ::arrow::utf8(); - return Status::OK(); + return OkStatus(); } ::arrow::Status status = ::arrow::adapters::tensorflow::GetArrowType(dtype, out); @@ -58,7 +58,7 @@ Status GetArrowType(::tensorflow::DataType dtype, return errors::InvalidArgument("tensorflow data type ", dtype, " is not supported: ", status); } - return Status::OK(); + return OkStatus(); } class ArrowAssignSpecImpl : public arrow::ArrayVisitor { @@ -91,7 +91,7 @@ class ArrowAssignSpecImpl : public arrow::ArrayVisitor { } CHECK_ARROW(array->Accept(this)); - return Status::OK(); + return OkStatus(); } protected: @@ -101,7 +101,7 @@ class ArrowAssignSpecImpl : public arrow::ArrayVisitor { return ::arrow::adapters::tensorflow::GetTensorFlowType(array.type(), out_dtype_); } - return arrow::Status::OK(); + return arrow::OkStatus(); } #define VISIT_PRIMITIVE(TYPE) \ @@ -187,7 +187,7 @@ class ArrowAssignTensorImpl : public arrow::ArrayVisitor { "Arrow arrays with null values not currently supported"); } CHECK_ARROW(array->Accept(this)); - return Status::OK(); + return OkStatus(); } protected: @@ -202,7 +202,7 @@ class ArrowAssignTensorImpl : public arrow::ArrayVisitor { memcpy(dst, &value, sizeof(value)); } - return arrow::Status::OK(); + return arrow::OkStatus(); } template @@ -227,7 +227,7 @@ class ArrowAssignTensorImpl : public arrow::ArrayVisitor { void* dst = const_cast(out_tensor_->tensor_data().data()); std::memcpy(dst, src, out_tensor_->NumElements() * type_width); - return arrow::Status::OK(); + return arrow::OkStatus(); } #define VISIT_FIXED_WIDTH(TYPE) \ @@ -288,7 +288,7 @@ class ArrowAssignTensorImpl : public arrow::ArrayVisitor { output_flat(j) = array.GetString(i_ + j); } - return arrow::Status::OK(); + return arrow::OkStatus(); } virtual arrow::Status Visit(const arrow::BinaryArray& array) override { @@ -299,7 +299,7 @@ class ArrowAssignTensorImpl : public arrow::ArrayVisitor { output_flat(j) = array.GetString(i_ + j); } - return arrow::Status::OK(); + return arrow::OkStatus(); } private: @@ -324,12 +324,12 @@ class ArrowArrayTypeCheckerImpl : public arrow::TypeVisitor { // First see if complex type handled by visitor arrow::Status visit_status = type->Accept(this); if (visit_status.ok()) { - return Status::OK(); + return OkStatus(); } // Check type as a scalar type CHECK_ARROW(CheckScalarType(type)); - return Status::OK(); + return OkStatus(); } protected: @@ -351,7 +351,7 @@ class ArrowArrayTypeCheckerImpl : public arrow::TypeVisitor { std::to_string(expected_type_) + ", but got dtype=" + std::to_string(converted_type)); } - return arrow::Status::OK(); + return arrow::OkStatus(); } private: @@ -375,7 +375,7 @@ class ArrowMakeArrayDataImpl : public arrow::TypeVisitor { buffers_ = buffers; out_data_ = out_data; CHECK_ARROW(type->Accept(this)); - return Status::OK(); + return OkStatus(); } protected: @@ -384,7 +384,7 @@ class ArrowMakeArrayDataImpl : public arrow::TypeVisitor { // TODO null count == 0 *out_data_ = arrow::ArrayData::Make(type_, lengths_[0], std::move(buffers_), 0); - return arrow::Status::OK(); + return arrow::OkStatus(); } #define VISIT_PRIMITIVE(TYPE) \ @@ -428,7 +428,7 @@ class ArrowMakeArrayDataImpl : public arrow::TypeVisitor { *out_data_ = arrow::ArrayData::Make(list_type, list_length, std::move(list_bufs), {child_data}, 0); - return arrow::Status::OK(); + return arrow::OkStatus(); } private: @@ -462,14 +462,14 @@ Status ParseEndpoint(std::string endpoint, std::string* endpoint_type, if (endpoint.substr(sep_pos + 1, 2) != "//") { *endpoint_type = ""; *endpoint_value = endpoint; - return Status::OK(); + return OkStatus(); } // Parse string as :// *endpoint_type = endpoint.substr(0, sep_pos); *endpoint_value = endpoint.substr(sep_pos + 3); - return Status::OK(); + return OkStatus(); } Status ParseHost(std::string host, std::string* host_address, @@ -483,7 +483,7 @@ Status ParseHost(std::string host, std::string* host_address, *host_address = host.substr(0, sep_pos); *host_port = host.substr(sep_pos + 1); - return Status::OK(); + return OkStatus(); } } // namespace ArrowUtil diff --git a/tensorflow_io/core/kernels/audio_kernels.cc b/tensorflow_io/core/kernels/audio_kernels.cc index 338d53a71..0abfe9150 100644 --- a/tensorflow_io/core/kernels/audio_kernels.cc +++ b/tensorflow_io/core/kernels/audio_kernels.cc @@ -48,7 +48,7 @@ Status PartitionsLookup(const std::vector& partitions, const int64 start, } else { *extra = start - partitions[(*lower) - 1]; } - return Status::OK(); + return OkStatus(); } namespace { @@ -133,7 +133,7 @@ class AudioReadableInitOp : public ResourceOpKernel { Status CreateResource(AudioReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new AudioReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -208,7 +208,7 @@ class AudioReadableReadOp : public OpKernel { [&](const TensorShape& shape, Tensor** value) -> Status { TF_RETURN_IF_ERROR( context->allocate_output(0, shape, value)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/audio_video_flac_kernels.cc b/tensorflow_io/core/kernels/audio_video_flac_kernels.cc index ce04842f2..8c12d9e85 100644 --- a/tensorflow_io/core/kernels/audio_video_flac_kernels.cc +++ b/tensorflow_io/core/kernels/audio_video_flac_kernels.cc @@ -283,7 +283,7 @@ class FlacReadableResource : public AudioReadableResourceBase { shape_ = TensorShape({samples, channels}); dtype_ = dtype; rate_ = rate; - return Status::OK(); + return OkStatus(); } Status Spec(TensorShape* shape, DataType* dtype, int32* rate) override { @@ -291,7 +291,7 @@ class FlacReadableResource : public AudioReadableResourceBase { *shape = shape_; *dtype = dtype_; *rate = rate_; - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, @@ -318,7 +318,7 @@ class FlacReadableResource : public AudioReadableResourceBase { stream_decoder_->sample_index); } } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "FlacReadableResource"; } @@ -381,7 +381,7 @@ class AudioDecodeFlacOp : public OpKernel { [&](const TensorShape& shape, Tensor** value) -> Status { TF_RETURN_IF_ERROR( context->allocate_output(0, shape, value)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/audio_video_mp3_kernels.cc b/tensorflow_io/core/kernels/audio_video_mp3_kernels.cc index 7b31500f6..20933c35b 100644 --- a/tensorflow_io/core/kernels/audio_video_mp3_kernels.cc +++ b/tensorflow_io/core/kernels/audio_video_mp3_kernels.cc @@ -134,7 +134,7 @@ class MP3ReadableResource : public AudioReadableResourceBase { dtype_ = DT_FLOAT; rate_ = rate; - return Status::OK(); + return OkStatus(); } Status Spec(TensorShape* shape, DataType* dtype, int32* rate) override { @@ -142,7 +142,7 @@ class MP3ReadableResource : public AudioReadableResourceBase { *shape = shape_; *dtype = dtype_; *rate = rate_; - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, @@ -170,7 +170,7 @@ class MP3ReadableResource : public AudioReadableResourceBase { sample_start, " failed: ", mp3dec_ex_.last_error); } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "MP3ReadableResource"; } @@ -234,7 +234,7 @@ class AudioDecodeMP3Op : public OpKernel { [&](const TensorShape& shape, Tensor** value) -> Status { TF_RETURN_IF_ERROR( context->allocate_output(0, shape, value)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/audio_video_mp4_kernels.cc b/tensorflow_io/core/kernels/audio_video_mp4_kernels.cc index 414595716..481d3d7ec 100644 --- a/tensorflow_io/core/kernels/audio_video_mp4_kernels.cc +++ b/tensorflow_io/core/kernels/audio_video_mp4_kernels.cc @@ -320,7 +320,7 @@ class MP4AACReadableResource : public AudioReadableResourceBase { shape_ = TensorShape({samples, channels}); dtype_ = DT_FLOAT; rate_ = rate; - return Status::OK(); + return OkStatus(); } } @@ -332,7 +332,7 @@ class MP4AACReadableResource : public AudioReadableResourceBase { *shape = shape_; *dtype = dtype_; *rate = rate_; - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, @@ -350,7 +350,7 @@ class MP4AACReadableResource : public AudioReadableResourceBase { TensorShape({sample_stop - sample_start, shape_.dim_size(1)}), &value)); if (sample_stop == sample_start) { - return Status::OK(); + return OkStatus(); } void* state = state_.get(); @@ -428,7 +428,7 @@ class MP4AACReadableResource : public AudioReadableResourceBase { char* base = (char*)(value->flat().data()); char* data = (char*)&data_out[0] + extra * channels * sizeof(float); memcpy(base, data, value->NumElements() * sizeof(float)); - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "MP4AACReadableResource"; } @@ -507,7 +507,7 @@ class AudioDecodeAACOp : public OpKernel { [&](const TensorShape& shape, Tensor** value) -> Status { TF_RETURN_IF_ERROR( context->allocate_output(0, shape, value)); - return Status::OK(); + return OkStatus(); })); } @@ -767,7 +767,7 @@ class VideoReadableResource : public ResourceBase { width_ = width; height_ = height; bytes_ = bytes; - return Status::OK(); + return OkStatus(); } Status Read( const int64 index, @@ -781,7 +781,7 @@ class VideoReadableResource : public ResourceBase { Tensor* value_tensor; if (sample_index_ >= mp4d_demux_.track[track_index_].sample_count) { TF_RETURN_IF_ERROR(allocate_func(TensorShape({0}), &value_tensor)); - return Status::OK(); + return OkStatus(); } unsigned frame_bytes, timestamp, duration; @@ -808,7 +808,7 @@ class VideoReadableResource : public ResourceBase { return errors::InvalidArgument("error to decode: ", status); } sample_index_++; - return Status::OK(); + return OkStatus(); } string DebugString() const override { mutex_lock l(mu_); @@ -854,7 +854,7 @@ class VideoReadableInitOp : public ResourceOpKernel { Status CreateResource(VideoReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new VideoReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -885,7 +885,7 @@ class VideoReadableReadOp : public OpKernel { Tensor** value_tensor) -> Status { TF_RETURN_IF_ERROR(context->allocate_output( 0, shape, value_tensor)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/audio_video_ogg_kernels.cc b/tensorflow_io/core/kernels/audio_video_ogg_kernels.cc index 2d2d9350f..b0dc3f956 100644 --- a/tensorflow_io/core/kernels/audio_video_ogg_kernels.cc +++ b/tensorflow_io/core/kernels/audio_video_ogg_kernels.cc @@ -112,7 +112,7 @@ class OggVorbisReadableResource : public AudioReadableResourceBase { dtype_ = DT_FLOAT; rate_ = rate; - return Status::OK(); + return OkStatus(); } Status Spec(TensorShape* shape, DataType* dtype, int32* rate) override { @@ -120,7 +120,7 @@ class OggVorbisReadableResource : public AudioReadableResourceBase { *shape = shape_; *dtype = dtype_; *rate = rate_; - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, @@ -164,7 +164,7 @@ class OggVorbisReadableResource : public AudioReadableResourceBase { samples_read += chunk; } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "OggVorbisReadableResource"; } @@ -227,7 +227,7 @@ class AudioDecodeVorbisOp : public OpKernel { [&](const TensorShape& shape, Tensor** value) -> Status { TF_RETURN_IF_ERROR( context->allocate_output(0, shape, value)); - return Status::OK(); + return OkStatus(); })); } @@ -258,7 +258,7 @@ Status OggVorbisEncodeStreamProcess(vorbis_dsp_state& vd, vorbis_block& vb, } } } - return Status::OK(); + return OkStatus(); } class AudioEncodeVorbisOp : public OpKernel { diff --git a/tensorflow_io/core/kernels/audio_video_wav_kernels.cc b/tensorflow_io/core/kernels/audio_video_wav_kernels.cc index af9993f4e..3f0f760d4 100644 --- a/tensorflow_io/core/kernels/audio_video_wav_kernels.cc +++ b/tensorflow_io/core/kernels/audio_video_wav_kernels.cc @@ -72,7 +72,7 @@ Status ValidateWAVHeader(struct WAVHeader* header) { return errors::InvalidArgument("WAV file have invalide channels: ", header->nChannels); } - return Status::OK(); + return OkStatus(); } class WAVReadableResource : public AudioReadableResourceBase { @@ -171,7 +171,7 @@ class WAVReadableResource : public AudioReadableResourceBase { rate_ = header_.nSamplesPerSec; - return Status::OK(); + return OkStatus(); } Status Spec(TensorShape* shape, DataType* dtype, int32* rate) override { @@ -179,7 +179,7 @@ class WAVReadableResource : public AudioReadableResourceBase { *shape = shape_; *dtype = dtype_; *rate = rate_; - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, @@ -195,7 +195,7 @@ class WAVReadableResource : public AudioReadableResourceBase { TF_RETURN_IF_ERROR(allocate_func( TensorShape({sample_stop - sample_start, shape_.dim_size(1)}), &value)); if (sample_stop == start) { - return Status::OK(); + return OkStatus(); } const int64 channels = shape_.dim_size(1); @@ -274,7 +274,7 @@ class WAVReadableResource : public AudioReadableResourceBase { base_offset += chunk_length; } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "WAVReadableResource"; } @@ -339,7 +339,7 @@ class AudioDecodeWAVOp : public OpKernel { [&](const TensorShape& shape, Tensor** value) -> Status { TF_RETURN_IF_ERROR( context->allocate_output(0, shape, value)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.cc b/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.cc index 10ad74899..c1ab2e62f 100644 --- a/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.cc +++ b/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.cc @@ -57,7 +57,7 @@ class AvroRecordDatasetOp::Dataset : public DatasetBase { } Status InputDatasets(std::vector* inputs) const { - return Status::OK(); + return OkStatus(); } std::unique_ptr MakeIteratorInternal( @@ -81,7 +81,7 @@ class AvroRecordDatasetOp::Dataset : public DatasetBase { return name_utils::DatasetDebugString(kDatasetType); } - Status CheckExternalState() const override { return Status::OK(); } + Status CheckExternalState() const override { return OkStatus(); } protected: Status AsGraphDefInternal(SerializationContext* ctx, @@ -91,7 +91,7 @@ class AvroRecordDatasetOp::Dataset : public DatasetBase { TF_RETURN_IF_ERROR(b->AddVector(filenames_, &filenames)); Node* buffer_size = nullptr; TF_RETURN_IF_ERROR(b->AddScalar(options_.buffer_size, &buffer_size)); - return Status::OK(); + return OkStatus(); } private: @@ -113,7 +113,7 @@ class AvroRecordDatasetOp::Dataset : public DatasetBase { reader_->ReadRecord(&out_tensors->back().scalar()()); if (s.ok()) { *end_of_sequence = false; - return Status::OK(); + return OkStatus(); } out_tensors->pop_back(); if (!errors::IsOutOfRange(s)) { @@ -134,7 +134,7 @@ class AvroRecordDatasetOp::Dataset : public DatasetBase { // Iteration ends when there are no more files to process. if (current_file_index_ == dataset()->filenames_.size()) { *end_of_sequence = true; - return Status::OK(); + return OkStatus(); } TF_RETURN_IF_ERROR(SetupStreamsLocked(ctx->env())); @@ -162,7 +162,7 @@ class AvroRecordDatasetOp::Dataset : public DatasetBase { TF_RETURN_IF_ERROR( writer->WriteScalar(full_name(kOffset), reader_->TellOffset())); } - return Status::OK(); + return OkStatus(); } */ @@ -187,7 +187,7 @@ class AvroRecordDatasetOp::Dataset : public DatasetBase { TF_RETURN_IF_ERROR(SetupStreamsLocked(ctx->env())); TF_RETURN_IF_ERROR(reader_->SeekOffset(offset)); } - return Status::OK(); + return OkStatus(); } */ @@ -205,7 +205,7 @@ class AvroRecordDatasetOp::Dataset : public DatasetBase { TF_RETURN_IF_ERROR(env->NewRandomAccessFile(next_filename, &file_)); reader_ = absl::make_unique( file_.get(), dataset()->options_); - return Status::OK(); + return OkStatus(); } // Resets all reader streams. diff --git a/tensorflow_io/core/kernels/avro/parse_avro_kernels.cc b/tensorflow_io/core/kernels/avro/parse_avro_kernels.cc index e0e02fdc6..17ed0549b 100644 --- a/tensorflow_io/core/kernels/avro/parse_avro_kernels.cc +++ b/tensorflow_io/core/kernels/avro/parse_avro_kernels.cc @@ -318,7 +318,7 @@ Status ParseAvro(const AvroParserConfig& config, VLOG(5) << "Sparse dense shapes shape: " << result->sparse_shapes[i_sparse].shape(); - return Status::OK(); + return OkStatus(); }; auto MergeDenseMinibatches = [&](size_t i_dense) -> Status { @@ -379,7 +379,7 @@ Status ParseAvro(const AvroParserConfig& config, VLOG(5) << "Dense tensor " << dense_tensor->SummarizeValue(9); - return Status::OK(); + return OkStatus(); }; const auto before_sparse_merge = clock::now(); for (size_t d = 0; d < config.sparse.size(); ++d) { @@ -397,7 +397,7 @@ Status ParseAvro(const AvroParserConfig& config, VLOG(5) << "PARSER_TIMING: Dense merge duration" << d_merge_duration.count() << " ms "; - return Status::OK(); + return OkStatus(); } // Inspired from here diff --git a/tensorflow_io/core/kernels/avro/utils/avro_parser.cc b/tensorflow_io/core/kernels/avro/utils/avro_parser.cc index 097569d97..1b9c21e7b 100644 --- a/tensorflow_io/core/kernels/avro/utils/avro_parser.cc +++ b/tensorflow_io/core/kernels/avro/utils/avro_parser.cc @@ -94,7 +94,7 @@ Status CheckValidDefault(const string& key, expected, "' but got data type '", default_value.dtype(), "'."); } - return Status::OK(); + return OkStatus(); } // This implementation assumes there is at least one expected type @@ -126,7 +126,7 @@ Status BoolValueParser::Parse(std::map* values, } (*reinterpret_cast((*values).at(key_).get())).Add(value); - return Status::OK(); + return OkStatus(); } string BoolValueParser::ToString(size_t level) const { return LevelToString(level) + "|---BoolValue(" + key_ + ")\n"; @@ -150,7 +150,7 @@ Status LongValueParser::Parse(std::map* values, // Assume the key exists and cast is possible (*reinterpret_cast((*values).at(key_).get())).Add(value); - return Status::OK(); + return OkStatus(); } string LongValueParser::ToString(size_t level) const { @@ -174,7 +174,7 @@ Status IntValueParser::Parse(std::map* values, // Assume the key exists and cast is possible (*reinterpret_cast((*values).at(key_).get())).Add(value); - return Status::OK(); + return OkStatus(); } string IntValueParser::ToString(size_t level) const { @@ -199,7 +199,7 @@ Status DoubleValueParser::Parse( // Assume the key exists and cast is possible (*reinterpret_cast((*values)[key_].get())).Add(value); - return Status::OK(); + return OkStatus(); } string DoubleValueParser::ToString(size_t level) const { @@ -223,7 +223,7 @@ Status FloatValueParser::Parse(std::map* values, // Assume the key exists and cast is possible (*reinterpret_cast((*values)[key_].get())).Add(value); - return Status::OK(); + return OkStatus(); } string FloatValueParser::ToString(size_t level) const { @@ -271,7 +271,7 @@ Status StringBytesEnumFixedValueParser::Parse( (*reinterpret_cast((*values)[key_].get())) .AddByRef(value); - return Status::OK(); + return OkStatus(); } string StringBytesEnumFixedValueParser::ToString(size_t level) const { return LevelToString(level) + "|---StringBytesEnumFixedValue(" + key_ + ")\n"; @@ -316,7 +316,7 @@ Status ArrayAllParser::Parse(std::map* values, (*(*values)[(*value_parser).GetKey()]).FinishMark(); } - return Status::OK(); + return OkStatus(); } string ArrayAllParser::ToString(size_t level) const { std::stringstream ss; @@ -356,7 +356,7 @@ Status ArrayIndexParser::Parse(std::map* values, TF_RETURN_IF_ERROR((*child).Parse(values, d, defaults)); } - return Status::OK(); + return OkStatus(); } string ArrayIndexParser::ToString(size_t level) const { std::stringstream ss; @@ -440,7 +440,7 @@ Status ArrayFilterParser::Parse( (*(*values)[(*value_parser).GetKey()]).FinishMark(); } - return Status::OK(); + return OkStatus(); } string ArrayFilterParser::ToString(size_t level) const { std::stringstream ss; @@ -481,7 +481,7 @@ Status MapKeyParser::Parse(std::map* values, return errors::InvalidArgument("Unable to find key '", key_, "'."); } - return Status::OK(); + return OkStatus(); } string MapKeyParser::ToString(size_t level) const { std::stringstream ss; @@ -516,7 +516,7 @@ Status RecordParser::Parse(std::map* values, TF_RETURN_IF_ERROR((*child).Parse(values, d, defaults)); } - return Status::OK(); + return OkStatus(); } string RecordParser::ToString(size_t level) const { std::stringstream ss; @@ -554,7 +554,7 @@ Status UnionParser::Parse(std::map* values, } // We could not find the right child, log a warning since this datum is lost // LOG(WARN) << "Branch value in data is not consumed"; - return Status::OK(); + return OkStatus(); } string UnionParser::ToString(size_t level) const { std::stringstream ss; @@ -572,7 +572,7 @@ Status RootParser::Parse(std::map* values, for (const AvroParserSharedPtr& child : children) { TF_RETURN_IF_ERROR((*child).Parse(values, datum, defaults)); } - return Status::OK(); + return OkStatus(); } string RootParser::ToString(size_t level) const { std::stringstream ss; diff --git a/tensorflow_io/core/kernels/avro/utils/avro_parser_tree.cc b/tensorflow_io/core/kernels/avro/utils/avro_parser_tree.cc index b8d5616d5..a1dd86573 100644 --- a/tensorflow_io/core/kernels/avro/utils/avro_parser_tree.cc +++ b/tensorflow_io/core/kernels/avro/utils/avro_parser_tree.cc @@ -73,7 +73,7 @@ Status AvroParserTree::ParseValues( // add end marks to all buffers for batch TF_RETURN_IF_ERROR(AddFinishMarks(key_to_value)); - return Status::OK(); + return OkStatus(); } Status AvroParserTree::ParseValues( @@ -114,7 +114,7 @@ Status AvroParserTree::ParseValues( // add end marks to all buffers for batch TF_RETURN_IF_ERROR(AddFinishMarks(key_to_value)); - return Status::OK(); + return OkStatus(); } Status AvroParserTree::Build(AvroParserTree* parser_tree, @@ -171,7 +171,7 @@ Status AvroParserTree::Build(AvroParserTree* parser_tree, VLOG(7) << "Parser tree \n" << (*parser_tree).ToString(); - return Status::OK(); + return OkStatus(); } Status AvroParserTree::Build( @@ -209,7 +209,7 @@ Status AvroParserTree::Build( (*parent).AddChild(std::move(avro_parser)); } - return Status::OK(); + return OkStatus(); } void AvroParserTree::Initialize() { @@ -239,7 +239,7 @@ Status AvroParserTree::ValidateUniqueKeys( } } - return Status::OK(); + return OkStatus(); } struct KeyWithTypeHash { @@ -321,13 +321,13 @@ Status AvroParserTree::CreateValueParser(AvroParserUniquePtr& avro_parser, const string& user_name) const { if (IsArrayAll(infix)) { avro_parser.reset(new ArrayAllParser()); - return Status::OK(); + return OkStatus(); } int index; if (IsArrayIndex(&index, infix)) { avro_parser.reset(new ArrayIndexParser(index)); - return Status::OK(); + return OkStatus(); } string lhs_name; @@ -361,24 +361,24 @@ Status AvroParserTree::CreateValueParser(AvroParserUniquePtr& avro_parser, avro_parser.reset(new ArrayFilterParser( lhs_resolved_name, rhs_resolved_name, array_filter_type)); - return Status::OK(); + return OkStatus(); } string key; if (IsMapKey(&key, infix)) { avro_parser.reset(new MapKeyParser(key)); - return Status::OK(); + return OkStatus(); } // TODO(fraudies): Check that the name appears in the prefix tree if (IsAttribute(infix)) { avro_parser.reset(new RecordParser(infix)); - return Status::OK(); + return OkStatus(); } if (IsBranch(infix)) { avro_parser.reset(new UnionParser(infix)); - return Status::OK(); + return OkStatus(); } return Status(errors::InvalidArgument("Unable to match ", infix, @@ -414,7 +414,7 @@ Status AvroParserTree::CreateFinalValueParser(AvroParserUniquePtr& value_parser, "', because data type '", DataTypeString(data_type), "' is not supported!"); } - return Status::OK(); + return OkStatus(); } bool AvroParserTree::ContainsFilter(string* lhs_name, string* rhs_name, @@ -463,7 +463,7 @@ Status AvroParserTree::AddBeginMarks( for (auto const& key_value : *key_to_value) { (*key_value.second).BeginMark(); } - return Status::OK(); + return OkStatus(); } Status AvroParserTree::AddFinishMarks( @@ -471,7 +471,7 @@ Status AvroParserTree::AddFinishMarks( for (auto const& key_value : *key_to_value) { (*key_value.second).FinishMark(); } - return Status::OK(); + return OkStatus(); } Status AvroParserTree::InitializeValueBuffers( @@ -521,7 +521,7 @@ Status AvroParserTree::InitializeValueBuffers( "' is not supported!")); } } - return Status::OK(); + return OkStatus(); } // TODO(fraudies): Replace by regex once tensorflow is compiled with GCC > 4.8 diff --git a/tensorflow_io/core/kernels/avro/utils/avro_record_reader.cc b/tensorflow_io/core/kernels/avro/utils/avro_record_reader.cc index 649be3c69..57a11608e 100644 --- a/tensorflow_io/core/kernels/avro/utils/avro_record_reader.cc +++ b/tensorflow_io/core/kernels/avro/utils/avro_record_reader.cc @@ -117,7 +117,7 @@ Status AvroRecordReader::ReadRecord(uint64* offset, tstring* record) { while (reader_stream->next(&data, &n_data)) { record->append((const char*)data, n_data); } - return record->empty() ? errors::OutOfRange("eof") : Status::OK(); + return record->empty() ? errors::OutOfRange("eof") : OkStatus(); } SequentialAvroRecordReader::SequentialAvroRecordReader( diff --git a/tensorflow_io/core/kernels/avro/utils/parse_avro_attrs.cc b/tensorflow_io/core/kernels/avro/utils/parse_avro_attrs.cc index 78bd9dad4..4d5785288 100644 --- a/tensorflow_io/core/kernels/avro/utils/parse_avro_attrs.cc +++ b/tensorflow_io/core/kernels/avro/utils/parse_avro_attrs.cc @@ -32,7 +32,7 @@ Status CheckValidType(const DataType& dtype) { case DT_FLOAT: case DT_DOUBLE: case DT_STRING: - return Status::OK(); + return OkStatus(); default: return errors::InvalidArgument("Received input dtype: ", DataTypeString(dtype)); diff --git a/tensorflow_io/core/kernels/avro/utils/value_buffer.cc b/tensorflow_io/core/kernels/avro/utils/value_buffer.cc index 722983a40..4bbbc561e 100644 --- a/tensorflow_io/core/kernels/avro/utils/value_buffer.cc +++ b/tensorflow_io/core/kernels/avro/utils/value_buffer.cc @@ -155,7 +155,7 @@ Status ShapeBuilder::GetCopyInfo( } } - return Status::OK(); + return OkStatus(); } Status ShapeBuilder::GetFillInfo( @@ -215,7 +215,7 @@ Status ShapeBuilder::GetFillInfo( } } - return Status::OK(); + return OkStatus(); } Status ShapeBuilder::GetIndices(Tensor* indices) const { @@ -249,7 +249,7 @@ Status ShapeBuilder::GetIndices(Tensor* indices) const { } } - return Status::OK(); + return OkStatus(); } string ShapeBuilder::ToString() const { @@ -334,7 +334,7 @@ Status MergeAs(ValueStoreUniquePtr& merged, DataTypeString(dtype)); } - return Status::OK(); + return OkStatus(); } } // namespace data diff --git a/tensorflow_io/core/kernels/avro_kernels.cc b/tensorflow_io/core/kernels/avro_kernels.cc index a6e15cb0c..c48ccdbb9 100644 --- a/tensorflow_io/core/kernels/avro_kernels.cc +++ b/tensorflow_io/core/kernels/avro_kernels.cc @@ -406,7 +406,7 @@ class AvroReadable : public IOReadableInterface { for (size_t i = 0; i < columns_.size(); i++) { shapes_.emplace_back(TensorShape({total})); } - return Status::OK(); + return OkStatus(); } Status Partitions(std::vector* partitions) override { @@ -415,7 +415,7 @@ class AvroReadable : public IOReadableInterface { for (size_t i = 0; i < positions_.size(); i++) { partitions->emplace_back(positions_[i].first); } - return Status::OK(); + return OkStatus(); } Status Components(std::vector* components) override { @@ -423,7 +423,7 @@ class AvroReadable : public IOReadableInterface { for (size_t i = 0; i < columns_.size(); i++) { components->push_back(columns_[i]); } - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, PartialTensorShape* shape, DataType* dtype, bool label) override { @@ -433,7 +433,7 @@ class AvroReadable : public IOReadableInterface { int64 column_index = columns_index_[component]; *shape = shapes_[column_index]; *dtype = dtypes_[column_index]; - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, const string& component, @@ -444,7 +444,7 @@ class AvroReadable : public IOReadableInterface { int64 column_index = columns_index_[component]; (*record_read) = 0; if (start >= shapes_[column_index].dim_size(0)) { - return Status::OK(); + return OkStatus(); } const string& column = component; int64 element_start = start < shapes_[column_index].dim_size(0) @@ -459,7 +459,7 @@ class AvroReadable : public IOReadableInterface { " selection is out of boundary"); } if (element_start == element_stop) { - return Status::OK(); + return OkStatus(); } avro::GenericDatum datum(reader_schema_); @@ -543,7 +543,7 @@ class AvroReadable : public IOReadableInterface { } } (*record_read) = element_stop - element_start; - return Status::OK(); + return OkStatus(); } string DebugString() const override { diff --git a/tensorflow_io/core/kernels/bigquery/bigquery_dataset_op.cc b/tensorflow_io/core/kernels/bigquery/bigquery_dataset_op.cc index 1784ed9d5..b724a93e4 100644 --- a/tensorflow_io/core/kernels/bigquery/bigquery_dataset_op.cc +++ b/tensorflow_io/core/kernels/bigquery/bigquery_dataset_op.cc @@ -40,7 +40,7 @@ class BigQueryDatasetOp : public DatasetOpKernel { using DatasetOpKernel::DatasetOpKernel; Status InputDatasets(std::vector *inputs) const { - return Status::OK(); + return OkStatus(); } void MakeDataset(OpKernelContext *ctx, DatasetBase **output) override { @@ -204,7 +204,7 @@ class BigQueryDatasetOp : public DatasetOpKernel { string DebugString() const override { return "BigQueryDatasetOp::Dataset"; } - Status CheckExternalState() const override { return Status::OK(); } + Status CheckExternalState() const override { return OkStatus(); } tensorflow::BigQueryClientResource *client_resource() const { return client_resource_; diff --git a/tensorflow_io/core/kernels/bigquery/bigquery_kernels.cc b/tensorflow_io/core/kernels/bigquery/bigquery_kernels.cc index b1247191a..ea6043e6c 100644 --- a/tensorflow_io/core/kernels/bigquery/bigquery_kernels.cc +++ b/tensorflow_io/core/kernels/bigquery/bigquery_kernels.cc @@ -47,7 +47,7 @@ class BigQueryClientOp : public OpKernel { [this, ctx](BigQueryClientResource** ret) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) { *ret = new BigQueryClientResource(); - return Status::OK(); + return OkStatus(); })); core::ScopedUnref resource_cleanup(resource); initialized_ = true; diff --git a/tensorflow_io/core/kernels/bigquery/bigquery_lib.cc b/tensorflow_io/core/kernels/bigquery/bigquery_lib.cc index d6782b5a9..534c1b223 100644 --- a/tensorflow_io/core/kernels/bigquery/bigquery_lib.cc +++ b/tensorflow_io/core/kernels/bigquery/bigquery_lib.cc @@ -60,7 +60,7 @@ ::tensorflow::error::Code GcpErrorCodeToTfErrorCode(::grpc::StatusCode code) { Status GrpcStatusToTfStatus(const ::grpc::Status& status) { if (status.ok()) { - return Status::OK(); + return OkStatus(); } return Status(GcpErrorCodeToTfErrorCode(status.error_code()), strings::StrCat("Error reading from Cloud BigQuery: ", @@ -87,7 +87,7 @@ Status GetDataFormat(string data_format_str, } else { return errors::Internal("Unsupported data format: " + data_format_str); } - return Status::OK(); + return OkStatus(); } } // namespace tensorflow diff --git a/tensorflow_io/core/kernels/bigtable/bigtable_dataset_kernel.cc b/tensorflow_io/core/kernels/bigtable/bigtable_dataset_kernel.cc index 66db5e7d7..b1585895f 100644 --- a/tensorflow_io/core/kernels/bigtable/bigtable_dataset_kernel.cc +++ b/tensorflow_io/core/kernels/bigtable/bigtable_dataset_kernel.cc @@ -74,7 +74,7 @@ tensorflow::error::Code GoogleCloudErrorCodeToTfErrorCode( Status GoogleCloudStatusToTfStatus(const ::google::cloud::Status& status) { if (status.ok()) { - return Status::OK(); + return OkStatus(); } return Status( GoogleCloudErrorCodeToTfErrorCode(status.code()), @@ -164,7 +164,7 @@ class Iterator : public DatasetIterator { if (it_ == reader_.end()) { VLOG(1) << "End of sequence"; *end_of_sequence = true; - return Status::OK(); + return OkStatus(); } *end_of_sequence = false; @@ -199,7 +199,7 @@ class Iterator : public DatasetIterator { VLOG(1) << "incrementing iterator"; it_ = std::next(it_); - return Status::OK(); + return OkStatus(); } protected: @@ -339,7 +339,7 @@ class Dataset : public DatasetBase { DebugString()); } - Status CheckExternalState() const override { return Status::OK(); } + Status CheckExternalState() const override { return OkStatus(); } private: std::shared_ptr const& data_client_; diff --git a/tensorflow_io/core/kernels/bigtable/serialization.cc b/tensorflow_io/core/kernels/bigtable/serialization.cc index 43d5b448e..3beb9e81f 100644 --- a/tensorflow_io/core/kernels/bigtable/serialization.cc +++ b/tensorflow_io/core/kernels/bigtable/serialization.cc @@ -187,7 +187,7 @@ Status PutCellValueInTensor(Tensor& tensor, size_t index, DataType cell_type, default: return errors::Unimplemented("Data type not supported."); } - return Status::OK(); + return OkStatus(); } } // namespace io diff --git a/tensorflow_io/core/kernels/csv_kernels.cc b/tensorflow_io/core/kernels/csv_kernels.cc index f41d20b19..897430f2b 100644 --- a/tensorflow_io/core/kernels/csv_kernels.cc +++ b/tensorflow_io/core/kernels/csv_kernels.cc @@ -128,14 +128,14 @@ class CSVReadable : public IOReadableInterface { columns_index_[table_->ColumnNames()[i]] = i; } - return Status::OK(); + return OkStatus(); } Status Components(std::vector* components) override { components->clear(); for (size_t i = 0; i < columns_.size(); i++) { components->push_back(columns_[i]); } - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, PartialTensorShape* shape, DataType* dtype, bool label) override { @@ -149,7 +149,7 @@ class CSVReadable : public IOReadableInterface { } else { *dtype = dtypes_[column_index]; } - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, const string& component, @@ -161,7 +161,7 @@ class CSVReadable : public IOReadableInterface { (*record_read) = 0; if (start >= shapes_[column_index].dim_size(0)) { - return Status::OK(); + return OkStatus(); } const string& column = component; int64 element_start = start < shapes_[column_index].dim_size(0) @@ -176,7 +176,7 @@ class CSVReadable : public IOReadableInterface { " selection is out of boundary"); } if (element_start == element_stop) { - return Status::OK(); + return OkStatus(); } std::shared_ptr<::arrow::ChunkedArray> slice = @@ -261,7 +261,7 @@ class CSVReadable : public IOReadableInterface { } (*record_read) = element_stop - element_start; - return Status::OK(); + return OkStatus(); } string DebugString() const override { diff --git a/tensorflow_io/core/kernels/elasticsearch_kernels.cc b/tensorflow_io/core/kernels/elasticsearch_kernels.cc index 345ca580b..6862a9114 100644 --- a/tensorflow_io/core/kernels/elasticsearch_kernels.cc +++ b/tensorflow_io/core/kernels/elasticsearch_kernels.cc @@ -117,7 +117,7 @@ class ElasticsearchReadableResource : public ResourceBase { return errors::FailedPrecondition("Corrupted response from the server"); } - return Status::OK(); + return OkStatus(); } Status Next( @@ -170,7 +170,7 @@ class ElasticsearchReadableResource : public ResourceBase { error_response); } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "ElasticsearchBaseResource"; } @@ -189,7 +189,7 @@ class ElasticsearchReadableResource : public ResourceBase { } else return errors::FailedPrecondition("healthcheck failed"); - return Status::OK(); + return OkStatus(); } Status MakeAPICall(const std::string& url, rapidjson::Document* response_json, @@ -248,7 +248,7 @@ class ElasticsearchReadableResource : public ResourceBase { // Store the default headers if the response is valid headers_ = headers; - return Status::OK(); + return OkStatus(); } mutable mutex mu_; @@ -305,14 +305,14 @@ class ElasticsearchReadableInitOp context->allocate_output(1, columns_shape, columns)); TF_RETURN_IF_ERROR( context->allocate_output(2, columns_shape, dtypes)); - return Status::OK(); + return OkStatus(); })); } Status CreateResource(ElasticsearchReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new ElasticsearchReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -349,7 +349,7 @@ class ElasticsearchReadableNextOp : public OpKernel { Tensor** items) -> Status { TF_RETURN_IF_ERROR(context->allocate_output( 0, tensor_shape, items)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/ffmpeg_kernels.cc b/tensorflow_io/core/kernels/ffmpeg_kernels.cc index 1ab37f830..9089c731d 100644 --- a/tensorflow_io/core/kernels/ffmpeg_kernels.cc +++ b/tensorflow_io/core/kernels/ffmpeg_kernels.cc @@ -178,7 +178,7 @@ class FFmpegStream { return errors::InvalidArgument( "unable to find specified stream: media=", media, ", index=", index); } - return Status::OK(); + return OkStatus(); } Status OpenCodec(int64 thread_count, int64 thread_type) { int64 stream_index = stream_index_; @@ -224,7 +224,7 @@ class FFmpegStream { nb_frames_ = format_context_->streams[stream_index]->nb_frames; - return Status::OK(); + return OkStatus(); } static int ReadPacket(void* opaque, uint8_t* buf, int buf_size) { @@ -371,7 +371,7 @@ class FFmpegAudioStream : public FFmpegStream { } av_packet_unref(&packet_); - return Status::OK(); + return OkStatus(); } Status Peek(int64* samples) { *samples = 0; @@ -379,7 +379,7 @@ class FFmpegAudioStream : public FFmpegStream { for (size_t i = 0; i < frames_.size(); i++) { (*samples) += frames_[i]->nb_samples; } - return Status::OK(); + return OkStatus(); } Status Read(Tensor* value) { int64 datasize = DataTypeSize(dtype_); @@ -400,7 +400,7 @@ class FFmpegAudioStream : public FFmpegStream { base += datasize * frames_[i]->nb_samples; } frames_.clear(); - return Status::OK(); + return OkStatus(); } Status DecodePacket() { @@ -422,7 +422,7 @@ class FFmpegAudioStream : public FFmpegStream { TF_RETURN_IF_ERROR(DecodeFrame(&got_frame)); } av_packet_unref(&packet_); - return Status::OK(); + return OkStatus(); } // final cache clean up do { @@ -430,7 +430,7 @@ class FFmpegAudioStream : public FFmpegStream { } while (got_frame); packet_scope_.reset(nullptr); - return Status::OK(); + return OkStatus(); } Status DecodeFrame(int* got_frame) { std::unique_ptr frame(av_frame_alloc(), @@ -451,7 +451,7 @@ class FFmpegAudioStream : public FFmpegStream { if (*got_frame) { frames_.push_back(std::move(frame)); } - return Status::OK(); + return OkStatus(); } DataType dtype() { return dtype_; } int64 channels() { return channels_; } @@ -482,7 +482,7 @@ class FFmpegAudioReadableResource : public ResourceBase { sample_index_ = 0; - return Status::OK(); + return OkStatus(); } Status Seek(const int64 index) { if (index != 0) { @@ -493,13 +493,13 @@ class FFmpegAudioReadableResource : public ResourceBase { TF_RETURN_IF_ERROR(ffmpeg_audio_stream_->OpenAudio(audio_index_)); sample_index_ = 0; - return Status::OK(); + return OkStatus(); } Status Peek(TensorShape* shape) { int64 samples = 0; Status status = ffmpeg_audio_stream_->Peek(&samples); *shape = TensorShape({samples, ffmpeg_audio_stream_->channels()}); - return Status::OK(); + return OkStatus(); } Status Read(Tensor* value) { return ffmpeg_audio_stream_->Read(value); } string DebugString() const override { return "FFmpegAudioReadableResource"; } @@ -539,7 +539,7 @@ class FFmpegAudioReadableInitOp Status CreateResource(FFmpegAudioReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new FFmpegAudioReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -647,7 +647,7 @@ class FFmpegVideoStream : public FFmpegStream { } av_packet_unref(&packet_); - return Status::OK(); + return OkStatus(); } Status Peek(int64* frames) { *frames = 0; @@ -655,7 +655,7 @@ class FFmpegVideoStream : public FFmpegStream { TF_RETURN_IF_ERROR(DecodePacket()); (*frames) = frames_.size(); } - return Status::OK(); + return OkStatus(); } Status PeekAll(int64* frames) { Status status; @@ -663,7 +663,7 @@ class FFmpegVideoStream : public FFmpegStream { status = DecodePacket(); } while (status.ok()); (*frames) = frames_.size(); - return Status::OK(); + return OkStatus(); } Status Read(Tensor* value) { char* base = ((char*)(value->flat().data())); @@ -674,7 +674,7 @@ class FFmpegVideoStream : public FFmpegStream { } frames_.clear(); frames_buffer_.clear(); - return Status::OK(); + return OkStatus(); } Status DecodePacket() { @@ -696,7 +696,7 @@ class FFmpegVideoStream : public FFmpegStream { TF_RETURN_IF_ERROR(DecodeFrame(&got_frame)); } av_packet_unref(&packet_); - return Status::OK(); + return OkStatus(); } // final cache clean up do { @@ -704,7 +704,7 @@ class FFmpegVideoStream : public FFmpegStream { } while (got_frame); packet_scope_.reset(nullptr); - return Status::OK(); + return OkStatus(); } Status DecodeFrame(int* got_frame) { std::unique_ptr frame(av_frame_alloc(), @@ -746,7 +746,7 @@ class FFmpegVideoStream : public FFmpegStream { frames_.push_back(std::move(frame_rgb)); frames_buffer_.push_back(std::move(buffer_rgb)); } - return Status::OK(); + return OkStatus(); } DataType dtype() { return dtype_; } int64 channels() { return channels_; } @@ -782,7 +782,7 @@ class FFmpegVideoReadableResource : public ResourceBase { frame_index_ = 0; - return Status::OK(); + return OkStatus(); } Status Seek(const int64 index) { if (index != 0) { @@ -793,7 +793,7 @@ class FFmpegVideoReadableResource : public ResourceBase { TF_RETURN_IF_ERROR(ffmpeg_video_stream_->OpenVideo(video_index_)); frame_index_ = 0; - return Status::OK(); + return OkStatus(); } Status Peek(TensorShape* shape) { int64 frames = 0; @@ -801,7 +801,7 @@ class FFmpegVideoReadableResource : public ResourceBase { *shape = TensorShape({frames, ffmpeg_video_stream_->height(), ffmpeg_video_stream_->width(), ffmpeg_video_stream_->channels()}); - return Status::OK(); + return OkStatus(); } Status Read(Tensor* value) { return ffmpeg_video_stream_->Read(value); } string DebugString() const override { return "FFmpegVideoReadableResource"; } @@ -841,7 +841,7 @@ class FFmpegVideoReadableInitOp Status CreateResource(FFmpegVideoReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new FFmpegVideoReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/ffmpeg_kernels_deprecated.cc b/tensorflow_io/core/kernels/ffmpeg_kernels_deprecated.cc index c785a5ce7..d1fef8fc0 100644 --- a/tensorflow_io/core/kernels/ffmpeg_kernels_deprecated.cc +++ b/tensorflow_io/core/kernels/ffmpeg_kernels_deprecated.cc @@ -102,7 +102,7 @@ class FFmpegReadStream { stream_index_ = stream_index; io_context_.reset(io_context); format_context_.reset(format_context); - return Status::OK(); + return OkStatus(); } avformat_close_input(&format_context); } @@ -245,7 +245,7 @@ class FFmpegReadStreamMeta : public FFmpegReadStream { nb_frames_ = format_context_->streams[stream_index]->nb_frames; - return Status::OK(); + return OkStatus(); } Status InitializeDecoder() { // Initialize the decoders @@ -258,7 +258,7 @@ class FFmpegReadStreamMeta : public FFmpegReadStream { // TODO: reference after first? packet_scope_.reset(&packet_); - return Status::OK(); + return OkStatus(); } int64 Type() { return media_type_; } int64 RecordIndex() { return record_index_; } @@ -288,7 +288,7 @@ class FFmpegReadStreamMeta : public FFmpegReadStream { TF_RETURN_IF_ERROR(DecodeFrame(&got_frame)); } av_packet_unref(&packet_); - return Status::OK(); + return OkStatus(); } // final cache clean up do { @@ -296,7 +296,7 @@ class FFmpegReadStreamMeta : public FFmpegReadStream { } while (got_frame); packet_scope_.reset(nullptr); - return Status::OK(); + return OkStatus(); } virtual Status DecodeFrame(int* got_frame) = 0; virtual Status ReadDecoded(int64 record_to_read, int64* record_read, @@ -313,13 +313,13 @@ class FFmpegReadStreamMeta : public FFmpegReadStream { TF_RETURN_IF_ERROR(ReadDecoded(record_to_read, record_read, value)); if ((*record_read) >= record_to_read) { record_index_ += (*record_read); - return Status::OK(); + return OkStatus(); } status = DecodePacket(); } while (status.ok()); TF_RETURN_IF_ERROR(ReadDecoded(record_to_read, record_read, value)); record_index_ += (*record_read); - return Status::OK(); + return OkStatus(); } protected: @@ -373,13 +373,13 @@ class FFmpegVideoReadStreamMeta : public FFmpegReadStreamMeta { shape_ = PartialTensorShape({-1, height_, width_, 3}); dtype_ = DT_UINT8; - return Status::OK(); + return OkStatus(); } Status ReadDecoded(int64 record_to_read, int64* record_read, Tensor* value) override { while ((*record_read) < record_to_read) { if (frames_.empty()) { - return Status::OK(); + return OkStatus(); } int64 offset = (*record_read) * height_ * width_ * 3; memcpy(reinterpret_cast(&value->flat().data()[offset]), @@ -389,7 +389,7 @@ class FFmpegVideoReadStreamMeta : public FFmpegReadStreamMeta { frames_buffer_.pop_front(); (*record_read)++; } - return Status::OK(); + return OkStatus(); } Status DecodeFrame(int* got_frame) override { std::unique_ptr frame(av_frame_alloc(), @@ -429,7 +429,7 @@ class FFmpegVideoReadStreamMeta : public FFmpegReadStreamMeta { frames_.push_back(std::move(frame_rgb)); frames_buffer_.push_back(std::move(buffer_rgb)); } - return Status::OK(); + return OkStatus(); } virtual Status Peek(int64* record_to_read) { if (!initialized_) { @@ -442,7 +442,7 @@ class FFmpegVideoReadStreamMeta : public FFmpegReadStreamMeta { status = DecodePacket(); } while (status.ok()); *record_to_read = frames_.size(); - return Status::OK(); + return OkStatus(); } int64 Height() { return height_; } @@ -516,7 +516,7 @@ class FFmpegAudioReadStreamMeta : public FFmpegReadStreamMeta { ") format: ", format); } - return Status::OK(); + return OkStatus(); } Status ReadDecodedRecord(int64 record_to_read, int64* record_read, Tensor* value) { @@ -545,16 +545,16 @@ class FFmpegAudioReadStreamMeta : public FFmpegReadStreamMeta { (*record_read)++; samples_index_++; if ((*record_read) >= record_to_read) { - return Status::OK(); + return OkStatus(); } } - return Status::OK(); + return OkStatus(); } Status ReadDecoded(int64 record_to_read, int64* record_read, Tensor* value) override { while ((*record_read) < record_to_read) { if (frames_.empty()) { - return Status::OK(); + return OkStatus(); } if (samples_index_ < frames_.front()->nb_samples) { TF_RETURN_IF_ERROR( @@ -565,7 +565,7 @@ class FFmpegAudioReadStreamMeta : public FFmpegReadStreamMeta { samples_index_ = 0; } } - return Status::OK(); + return OkStatus(); } Status DecodeFrame(int* got_frame) override { std::unique_ptr frame(av_frame_alloc(), @@ -586,7 +586,7 @@ class FFmpegAudioReadStreamMeta : public FFmpegReadStreamMeta { if (*got_frame) { frames_.push_back(std::move(frame)); } - return Status::OK(); + return OkStatus(); } int64 Channels() { return channels_; } int64 Rate() { return rate_; } @@ -607,7 +607,7 @@ class FFmpegSubtitleReadStreamMeta : public FFmpegReadStreamMeta { TF_RETURN_IF_ERROR(FFmpegReadStreamMeta::Open(stream_index)); shape_ = PartialTensorShape({-1}); dtype_ = DT_STRING; - return Status::OK(); + return OkStatus(); } private: @@ -615,13 +615,13 @@ class FFmpegSubtitleReadStreamMeta : public FFmpegReadStreamMeta { Tensor* value) override { while ((*record_read) < record_to_read) { if (subtitles_.empty()) { - return Status::OK(); + return OkStatus(); } value->flat()((*record_read)) = subtitles_.front(); subtitles_.pop_front(); (*record_read)++; } - return Status::OK(); + return OkStatus(); } Status DecodeFrame(int* got_frame) override { AVSubtitle subtitle; @@ -667,7 +667,7 @@ class FFmpegSubtitleReadStreamMeta : public FFmpegReadStreamMeta { subtitle.rects[0]->type); } } - return Status::OK(); + return OkStatus(); } std::deque subtitles_; }; @@ -733,14 +733,14 @@ class FFmpegReadable : public IOReadableInterface { "invalid steam (", i, ") type: ", ffmpeg_file_->StreamType(i)); } } - return Status::OK(); + return OkStatus(); } Status Components(std::vector* components) override { components->clear(); for (size_t i = 0; i < columns_.size(); i++) { components->push_back(columns_[i]); } - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, PartialTensorShape* shape, DataType* dtype, bool label) override { @@ -750,7 +750,7 @@ class FFmpegReadable : public IOReadableInterface { int64 column_index = columns_index_[component]; *shape = shapes_[column_index]; *dtype = dtypes_[column_index]; - return Status::OK(); + return OkStatus(); } Status Extra(const string& component, std::vector* extra) override { @@ -764,7 +764,7 @@ class FFmpegReadable : public IOReadableInterface { Tensor rate(DT_INT64, TensorShape({})); rate.scalar()() = (meta != nullptr) ? meta->Rate() : 0; extra->push_back(rate); - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, const string& component, @@ -777,7 +777,7 @@ class FFmpegReadable : public IOReadableInterface { if (start != columns_meta_[column_index]->RecordIndex()) { // If we reach to the end, then just 0 if (start > columns_meta_[column_index]->RecordIndex()) { - return Status::OK(); + return OkStatus(); } if (start != 0) { return errors::InvalidArgument( diff --git a/tensorflow_io/core/kernels/file_kernels.cc b/tensorflow_io/core/kernels/file_kernels.cc index 08fdab1e3..d299c7b83 100644 --- a/tensorflow_io/core/kernels/file_kernels.cc +++ b/tensorflow_io/core/kernels/file_kernels.cc @@ -137,22 +137,22 @@ class FileResource : public ResourceBase { Status Init(const string& filename) { TF_RETURN_IF_ERROR(env_->NewWritableFile(filename, &file_)); - return Status::OK(); + return OkStatus(); } Status Write(const Tensor& content) { mutex_lock l(mu_); for (int64 i = 0; i < content.NumElements(); i++) { TF_RETURN_IF_ERROR(file_->Append(content.flat()(i))); } - return Status::OK(); + return OkStatus(); } Status Sync() { file_->Flush(); - return Status::OK(); + return OkStatus(); } Status Close() { file_.reset(nullptr); - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "FileResource"; } @@ -181,7 +181,7 @@ class FileInitOp : public ResourceOpKernel { Status CreateResource(FileResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new FileResource(env_); - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/grpc_kernels.cc b/tensorflow_io/core/kernels/grpc_kernels.cc index af21b9e41..ddf404a10 100644 --- a/tensorflow_io/core/kernels/grpc_kernels.cc +++ b/tensorflow_io/core/kernels/grpc_kernels.cc @@ -33,7 +33,7 @@ class GRPCReadableResource : public ResourceBase { endpoint_ = input; stub_ = GRPCEndpoint::NewStub( grpc::CreateChannel(endpoint_, grpc::InsecureChannelCredentials())); - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const TensorShape& shape, std::function @@ -43,7 +43,7 @@ class GRPCReadableResource : public ResourceBase { Tensor* value; TF_RETURN_IF_ERROR(allocate_func(shape, &value)); if (shape.dim_size(0) == 0) { - return Status::OK(); + return OkStatus(); } Request request; @@ -64,7 +64,7 @@ class GRPCReadableResource : public ResourceBase { return errors::InvalidArgument("unable to fill tensor"); } - return Status::OK(); + return OkStatus(); } string DebugString() const override { mutex_lock l(mu_); @@ -98,7 +98,7 @@ class GRPCReadableInitOp : public ResourceOpKernel { Status CreateResource(GRPCReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new GRPCReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -133,7 +133,7 @@ class GRPCReadableReadOp : public OpKernel { [&](const TensorShape& shape, Tensor** value) -> Status { TF_RETURN_IF_ERROR( context->allocate_output(0, shape, value)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.cc b/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.cc index b86a120b7..af1fc28ec 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.cc +++ b/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.cc @@ -15,7 +15,7 @@ Status GceMemcachedServerListProvider::GetServerList( std::vector* server_list) { if (!cached_list_.empty()) { *server_list = cached_list_; - return Status::OK(); + return OkStatus(); } std::vector response_buffer; @@ -41,7 +41,7 @@ Status GceMemcachedServerListProvider::GetServerList( } if (success) { - return Status::OK(); + return OkStatus(); } else { return Status(tensorflow::error::FAILED_PRECONDITION, "Unexpected server list format"); diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.cc b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.cc index 8c2033cd9..43a9a651d 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.cc +++ b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.cc @@ -214,7 +214,7 @@ Status block_get(MemcachedDaoInterface* memcached_dao, const string& key, StreamzRecordCacheHitBlockSize(value_length, cache_stats); // Any object returned by memcached must be released by the caller. free(retrieved_value); - return Status::OK(); + return OkStatus(); } if (rc == MEMCACHED_TIMEOUT) { @@ -243,7 +243,7 @@ Status block_multi_get(MemcachedDaoInterface* memcached_dao, rc = memcached_dao->MemcachedMget(key_ptrs.get(), key_lengths.get(), keys.size()); if (rc == MEMCACHED_SUCCESS) { - return Status::OK(); + return OkStatus(); } return errors::Internal("memcached multi_get failed ", @@ -257,7 +257,7 @@ Status block_set(MemcachedDaoInterface* memcached_dao, const string& key, value.size(), static_cast(0), static_cast(0)); if (rc == MEMCACHED_SUCCESS) { - return Status::OK(); + return OkStatus(); } return errors::Internal("memcached failed to store key ", key, @@ -329,7 +329,7 @@ Status read_with_multi_get( const auto after = absl::Now(); VLOG(2) << (to_claim - claim_checks->size()) << " multi-get fetches claimed in " << (after - before); - return Status::OK(); + return OkStatus(); } void cache_cleanup(void* tsd) { @@ -525,7 +525,7 @@ Status MemcachedFileBlockCache::ConfigureMemcachedServers( memcached_dao->MemcachedStrError(rc)); } - return Status::OK(); + return OkStatus(); } Status MemcachedFileBlockCache::Read(const string& filename, size_t offset, @@ -533,7 +533,7 @@ Status MemcachedFileBlockCache::Read(const string& filename, size_t offset, size_t* bytes_transferred) { *bytes_transferred = 0; if (n == 0) { - return Status::OK(); + return OkStatus(); } VLOG(2) << "original read: offset=" << offset << ", n=" << n << ", filename=" << filename; @@ -570,7 +570,7 @@ Status MemcachedFileBlockCache::Read(const string& filename, size_t offset, } if (local_cache_->Get(mini_read_key, offset_in_block, n, buffer, bytes_transferred)) { - return Status::OK(); + return OkStatus(); } } @@ -672,7 +672,7 @@ Status MemcachedFileBlockCache::Read(const string& filename, size_t offset, << (total_bytes_transferred / absl::ToDoubleSeconds(elapsed)) << " bytes / second"; *bytes_transferred = total_bytes_transferred; - return Status::OK(); + return OkStatus(); } string MemcachedFileBlockCache::MakeMemcachedKey(const Key& key) { diff --git a/tensorflow_io/core/kernels/hdf5_kernels.cc b/tensorflow_io/core/kernels/hdf5_kernels.cc index e3d53d929..0f2786c4e 100644 --- a/tensorflow_io/core/kernels/hdf5_kernels.cc +++ b/tensorflow_io/core/kernels/hdf5_kernels.cc @@ -308,7 +308,7 @@ class HDF5ReadableResource : public ResourceBase { columns_index_[data.datasets_[i]] = i; } - return Status::OK(); + return OkStatus(); } Status Components(std::vector* components) { @@ -319,7 +319,7 @@ class HDF5ReadableResource : public ResourceBase { for (const std::pair& e : columns_index_) { components->emplace_back(e.first); } - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, TensorShape* shape, DataType* dtype) { @@ -333,7 +333,7 @@ class HDF5ReadableResource : public ResourceBase { const int64 column_index = lookup->second; *shape = shapes_[column_index]; *dtype = dtypes_[column_index]; - return Status::OK(); + return OkStatus(); } Status Read(const string& component, @@ -573,7 +573,7 @@ class HDF5ReadableResource : public ResourceBase { filename_, ": ", e.getCDetailMsg()); } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "HDF5ReadableResource"; } @@ -641,7 +641,7 @@ class HDF5ReadableInfoOp : public IOResourceOpKernel { } dtype_tensor->flat()(i) = dtypes[i]; } - return Status::OK(); + return OkStatus(); } // HDF5 is not multi-threaded so use global mutext for protection @@ -704,9 +704,9 @@ class HDF5ReadableReadOp : public IOResourceOpKernel { component, start, shape, [&](const TensorShape& shape, Tensor** value) -> Status { TF_RETURN_IF_ERROR(context->allocate_output(0, shape, value)); - return Status::OK(); + return OkStatus(); })); - return Status::OK(); + return OkStatus(); } // HDF5 is not multi-threaded so use global mutext for protection diff --git a/tensorflow_io/core/kernels/image_dicom_kernels.cc b/tensorflow_io/core/kernels/image_dicom_kernels.cc index 86b692be1..550f6e62e 100644 --- a/tensorflow_io/core/kernels/image_dicom_kernels.cc +++ b/tensorflow_io/core/kernels/image_dicom_kernels.cc @@ -423,7 +423,7 @@ class DecodeDICOMDataOp : public OpKernel { uint16 tag_group_number = (uint16)((tag_value & 0xFFFF0000) >> 16); uint16 tag_element_number = (uint16)((tag_value & 0x0000FFFF) >> 0); *tag = DcmTag(tag_group_number, tag_element_number); - return Status::OK(); + return OkStatus(); } Status GetDcmTag(const absl::string_view tag_value, DcmTag *tag) { std::vector number_views = @@ -458,7 +458,7 @@ class DecodeDICOMDataOp : public OpKernel { uint16 tag_element_number = number; *tag = DcmTag(tag_group_number, tag_element_number); - return Status::OK(); + return OkStatus(); } }; diff --git a/tensorflow_io/core/kernels/image_font_kernels.cc b/tensorflow_io/core/kernels/image_font_kernels.cc index ef5eed71f..a8cfefc2a 100644 --- a/tensorflow_io/core/kernels/image_font_kernels.cc +++ b/tensorflow_io/core/kernels/image_font_kernels.cc @@ -83,7 +83,7 @@ Status InitializeFreeTypeLibrary() { } init = true; } - return Status::OK(); + return OkStatus(); } } // namespace diff --git a/tensorflow_io/core/kernels/io_optimization.cc b/tensorflow_io/core/kernels/io_optimization.cc index 239bdd1d2..e3de9f8e0 100644 --- a/tensorflow_io/core/kernels/io_optimization.cc +++ b/tensorflow_io/core/kernels/io_optimization.cc @@ -38,7 +38,7 @@ class IOGraphOptimizationPass : public GraphOptimizationPass { LOG(INFO) << "TFIO_GRAPH_DEBUG: [run]:" << graph->ToGraphDefDebug().DebugString(); } - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/json_kernels.cc b/tensorflow_io/core/kernels/json_kernels.cc index 3af30439c..7af956970 100644 --- a/tensorflow_io/core/kernels/json_kernels.cc +++ b/tensorflow_io/core/kernels/json_kernels.cc @@ -100,7 +100,7 @@ class JSONReadable : public IOReadableInterface { } } - return Status::OK(); + return OkStatus(); } json_file_.reset(new ArrowRandomAccessFile(file_.get(), file_size_)); @@ -136,14 +136,14 @@ class JSONReadable : public IOReadableInterface { columns_index_[table_->ColumnNames()[i]] = i; } - return Status::OK(); + return OkStatus(); } Status Components(std::vector* components) override { components->clear(); for (size_t i = 0; i < columns_.size(); i++) { components->push_back(columns_[i]); } - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, PartialTensorShape* shape, DataType* dtype, bool label) override { @@ -153,7 +153,7 @@ class JSONReadable : public IOReadableInterface { int64 column_index = columns_index_[component]; *shape = shapes_[column_index]; *dtype = dtypes_[column_index]; - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, const string& component, @@ -165,7 +165,7 @@ class JSONReadable : public IOReadableInterface { (*record_read) = 0; if (start >= shapes_[column_index].dim_size(0)) { - return Status::OK(); + return OkStatus(); } const string& column = component; int64 element_start = start < shapes_[column_index].dim_size(0) @@ -180,7 +180,7 @@ class JSONReadable : public IOReadableInterface { " selection is out of boundary"); } if (element_start == element_stop) { - return Status::OK(); + return OkStatus(); } if (mode_ == "records") { @@ -198,7 +198,7 @@ class JSONReadable : public IOReadableInterface { } (*record_read) = element_stop - element_start; - return Status::OK(); + return OkStatus(); } std::shared_ptr<::arrow::ChunkedArray> slice = @@ -254,7 +254,7 @@ class JSONReadable : public IOReadableInterface { DataTypeString(value->dtype())); } (*record_read) = element_stop - element_start; - return Status::OK(); + return OkStatus(); } string DebugString() const override { diff --git a/tensorflow_io/core/kernels/kafka_kernels.cc b/tensorflow_io/core/kernels/kafka_kernels.cc index 52f94972a..be448f1fa 100644 --- a/tensorflow_io/core/kernels/kafka_kernels.cc +++ b/tensorflow_io/core/kernels/kafka_kernels.cc @@ -162,7 +162,7 @@ class KafkaReadableResource : public ResourceBase { RdKafka::err2str(err)); } - return Status::OK(); + return OkStatus(); } Status Next(const int64 index, std::functionflat()(i) = message_value[i]; key_tensor->flat()(i) = key_value[i]; } - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, std::functionflat()(i) = message_value[i]; key_tensor->flat()(i) = key_value[i]; } - return Status::OK(); + return OkStatus(); } Status Spec(const int64 start, const int64 stop, int64* start_offset, int64* stop_offset) { @@ -322,7 +322,7 @@ class KafkaReadableResource : public ResourceBase { } } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "KafkaBaseResource"; } @@ -355,7 +355,7 @@ class KafkaReadableResource : public ResourceBase { RdKafka::err2str(err)); } - return Status::OK(); + return OkStatus(); } mutable mutex mu_; Env* env_ TF_GUARDED_BY(mu_); @@ -401,7 +401,7 @@ class KafkaReadableInitOp : public ResourceOpKernel { Status CreateResource(KafkaReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new KafkaReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -434,7 +434,7 @@ class KafkaReadableNextOp : public OpKernel { Tensor** key) -> Status { TF_RETURN_IF_ERROR(context->allocate_output(0, shape, message)); TF_RETURN_IF_ERROR(context->allocate_output(1, shape, key)); - return Status::OK(); + return OkStatus(); })); } @@ -472,7 +472,7 @@ class KafkaReadableReadOp : public OpKernel { Tensor** key) -> Status { TF_RETURN_IF_ERROR(context->allocate_output(0, shape, message)); TF_RETURN_IF_ERROR(context->allocate_output(1, shape, key)); - return Status::OK(); + return OkStatus(); })); } @@ -558,7 +558,7 @@ class KafkaIterableInitOp : public ResourceOpKernel { Status CreateResource(KafkaIterableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new KafkaIterableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -637,7 +637,7 @@ class LayerKafkaResource : public ResourceBase { } partition_ = partition; - return Status::OK(); + return OkStatus(); } Status Write(const Tensor& content) { mutex_lock l(mu_); @@ -651,7 +651,7 @@ class LayerKafkaResource : public ResourceBase { RdKafka::err2str(err)); } } - return Status::OK(); + return OkStatus(); } Status Sync() { if (producer_.get() != nullptr) { @@ -661,7 +661,7 @@ class LayerKafkaResource : public ResourceBase { RdKafka::err2str(err)); } } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "LayerKafkaResource"; } @@ -705,7 +705,7 @@ class LayerKafkaInitOp : public ResourceOpKernel { Status CreateResource(LayerKafkaResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new LayerKafkaResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -950,7 +950,7 @@ class KafkaGroupReadableResource : public ResourceBase { RdKafka::err2str(err)); } - return Status::OK(); + return OkStatus(); } Status Next(const int64 index, const int64 message_poll_timeout, const int64 stream_timeout, @@ -1022,7 +1022,7 @@ class KafkaGroupReadableResource : public ResourceBase { key_tensor->flat()(i) = key_value[i]; } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "KafkaBaseResource"; } @@ -1069,7 +1069,7 @@ class KafkaGroupReadableInitOp Status CreateResource(KafkaGroupReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new KafkaGroupReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -1115,7 +1115,7 @@ class KafkaGroupReadableNextOp : public OpKernel { TF_RETURN_IF_ERROR(context->allocate_output(1, shape, key)); TF_RETURN_IF_ERROR( context->allocate_output(2, TensorShape({}), continue_fetch)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/kafka_kernels_deprecated.cc b/tensorflow_io/core/kernels/kafka_kernels_deprecated.cc index 140eb42c9..3c0130cd6 100644 --- a/tensorflow_io/core/kernels/kafka_kernels_deprecated.cc +++ b/tensorflow_io/core/kernels/kafka_kernels_deprecated.cc @@ -148,7 +148,7 @@ class KafkaDatasetOp : public DatasetOpKernel { string DebugString() const override { return "KafkaDatasetOp::Dataset"; } - Status CheckExternalState() const override { return Status::OK(); } + Status CheckExternalState() const override { return OkStatus(); } protected: Status AsGraphDefInternal(SerializationContext* ctx, @@ -177,7 +177,7 @@ class KafkaDatasetOp : public DatasetOpKernel { {topics, servers, group, eof, timeout, config_global, config_topic, message_key, message_offset}, output)); - return Status::OK(); + return OkStatus(); } private: @@ -270,7 +270,7 @@ class KafkaDatasetOp : public DatasetOpKernel { *end_of_sequence = false; // Sync offset offset_ = message->offset(); - return Status::OK(); + return OkStatus(); } if (message->err() == RdKafka::ERR__PARTITION_EOF) { @@ -305,7 +305,7 @@ class KafkaDatasetOp : public DatasetOpKernel { // Iteration ends when there are no more topic to process. if (current_topic_index_ == dataset()->topics_.size()) { *end_of_sequence = true; - return Status::OK(); + return OkStatus(); } TF_RETURN_IF_ERROR(SetupStreamsLocked(ctx->env())); @@ -332,7 +332,7 @@ class KafkaDatasetOp : public DatasetOpKernel { << dataset()->topics_[current_topic_index_] << ", current offset: " << offset_; } - return Status::OK(); + return OkStatus(); */ } @@ -374,7 +374,7 @@ class KafkaDatasetOp : public DatasetOpKernel { << topic_partition_->partition() << ":" << topic_partition_->offset() << "]"; } - return Status::OK(); + return OkStatus(); } private: @@ -502,7 +502,7 @@ class KafkaDatasetOp : public DatasetOpKernel { LOG(INFO) << "Kafka stream starts with current offset: " << topic_partition_->offset(); - return Status::OK(); + return OkStatus(); } // Resets all Kafka streams. @@ -635,7 +635,7 @@ class KafkaOutputSequence : public OutputSequence { RdKafka::err2str(err)); } } - return Status::OK(); + return OkStatus(); } virtual Status Output() override { if (fifo_.front().get() != nullptr) { @@ -653,7 +653,7 @@ class KafkaOutputSequence : public OutputSequence { base_++; } } - return Status::OK(); + return OkStatus(); } virtual string DebugString() const { return strings::StrCat("KafkaOutputSequence[]"); @@ -728,7 +728,7 @@ class KafkaOutputSequence : public OutputSequence { errstr); } - return Status::OK(); + return OkStatus(); } private: @@ -832,7 +832,7 @@ class DecodeAvroResource : public ResourceBase { return errors::Unimplemented("Avro schema error: ", error); } - return Status::OK(); + return OkStatus(); } const avro::ValidSchema& avro_schema() { return avro_schema_; } string DebugString() const override { return "DecodeAvroResource"; } @@ -1108,7 +1108,7 @@ class DecodeAvroInitOp : public ResourceOpKernel { Status CreateResource(DecodeAvroResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new DecodeAvroResource(env_); - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/kinesis_kernels.cc b/tensorflow_io/core/kernels/kinesis_kernels.cc index 2149eb9e8..16053ed54 100644 --- a/tensorflow_io/core/kernels/kinesis_kernels.cc +++ b/tensorflow_io/core/kernels/kinesis_kernels.cc @@ -331,7 +331,7 @@ class KinesisReadableResource : public ResourceBase { ": ", iterator_outcome.GetError().GetMessage()); } iterator_ = iterator_outcome.GetResult().GetShardIterator(); - return Status::OK(); + return OkStatus(); } Status Read( std::functionflat()(0) = string(sequence.c_str(), sequence.size()); - return Status::OK(); + return OkStatus(); } while (true); - return Status::OK(); + return OkStatus(); } string DebugString() const override { mutex_lock l(mu_); @@ -431,7 +431,7 @@ class KinesisReadableInitOp : public ResourceOpKernel { Status CreateResource(KinesisReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new KinesisReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -464,7 +464,7 @@ class KinesisReadableReadOp : public OpKernel { context->allocate_output(2, shape, partition_tensor)); TF_RETURN_IF_ERROR( context->allocate_output(3, shape, sequence_tensor)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/lmdb_kernels.cc b/tensorflow_io/core/kernels/lmdb_kernels.cc index 9555b9d7a..9ddfdb78a 100644 --- a/tensorflow_io/core/kernels/lmdb_kernels.cc +++ b/tensorflow_io/core/kernels/lmdb_kernels.cc @@ -78,7 +78,7 @@ class LMDBReadable : public IOReadableInterface { if (status != MDB_SUCCESS) { return errors::InvalidArgument("error on mdb_cursor_open: ", status); } - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, const string& component, int64* record_read, Tensor* value, Tensor* label) override { @@ -94,13 +94,13 @@ class LMDBReadable : public IOReadableInterface { string(static_cast(mdb_key.mv_data), mdb_key.mv_size)); (*record_read)++; } - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, PartialTensorShape* shape, DataType* dtype, bool label) override { *shape = PartialTensorShape({-1}); *dtype = DT_STRING; - return Status::OK(); + return OkStatus(); } string DebugString() const override { @@ -166,14 +166,14 @@ class LMDBMapping : public IOMappingInterface { return errors::InvalidArgument("error on mdb_dbi_open: ", status); } - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, PartialTensorShape* shape, DataType* dtype, bool label) override { *shape = PartialTensorShape({-1}); *dtype = DT_STRING; - return Status::OK(); + return OkStatus(); } Status Read(const Tensor& key, Tensor* value) override { @@ -190,7 +190,7 @@ class LMDBMapping : public IOMappingInterface { value->flat()(i) = std::move( string(static_cast(mdb_data.mv_data), mdb_data.mv_size)); } - return Status::OK(); + return OkStatus(); } string DebugString() const override { diff --git a/tensorflow_io/core/kernels/mongodb_kernels.cc b/tensorflow_io/core/kernels/mongodb_kernels.cc index 826ec15ad..7c4e696ac 100644 --- a/tensorflow_io/core/kernels/mongodb_kernels.cc +++ b/tensorflow_io/core/kernels/mongodb_kernels.cc @@ -65,7 +65,7 @@ class MongoDBReadableResource : public ResourceBase { // Perform healthcheck before proceeding Healthcheck(); - return Status::OK(); + return OkStatus(); } Status Next(std::function @@ -107,7 +107,7 @@ class MongoDBReadableResource : public ResourceBase { records_tensor->flat()(i) = records[i]; } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "MongoDBReadableResource"; } @@ -127,7 +127,7 @@ class MongoDBReadableResource : public ResourceBase { } LOG(INFO) << "MongoDB connection ping successful"; - return Status::OK(); + return OkStatus(); } mutable mutex mu_; @@ -173,7 +173,7 @@ class MongoDBReadableInitOp : public ResourceOpKernel { Status CreateResource(MongoDBReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new MongoDBReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -195,7 +195,7 @@ class MongoDBReadableNextOp : public OpKernel { OP_REQUIRES_OK(context, resource->Next([&](const TensorShape& shape, Tensor** record) -> Status { TF_RETURN_IF_ERROR(context->allocate_output(0, shape, record)); - return Status::OK(); + return OkStatus(); })); } @@ -242,7 +242,7 @@ class MongoDBWritableResource : public ResourceBase { // Perform healthcheck before proceeding Healthcheck(); - return Status::OK(); + return OkStatus(); } Status Write(const std::string& record) { @@ -263,7 +263,7 @@ class MongoDBWritableResource : public ResourceBase { bson_destroy(bson_record); - return Status::OK(); + return OkStatus(); } Status DeleteMany(const std::string& record) { @@ -284,7 +284,7 @@ class MongoDBWritableResource : public ResourceBase { bson_destroy(bson_record); - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "MongoDBWritableResource"; } @@ -304,7 +304,7 @@ class MongoDBWritableResource : public ResourceBase { } LOG(INFO) << "MongoDB connection ping successful"; - return Status::OK(); + return OkStatus(); } mutable mutex mu_; @@ -348,7 +348,7 @@ class MongoDBWritableInitOp : public ResourceOpKernel { Status CreateResource(MongoDBWritableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new MongoDBWritableResource(env_); - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/numpy_kernels.cc b/tensorflow_io/core/kernels/numpy_kernels.cc index 2162586a5..d7fcbca6c 100644 --- a/tensorflow_io/core/kernels/numpy_kernels.cc +++ b/tensorflow_io/core/kernels/numpy_kernels.cc @@ -226,7 +226,7 @@ Status ParseNumpyHeader(io::InputStreamInterface* stream, } } - return Status::OK(); + return OkStatus(); } class ZipObjectInputStream : public io::InputStreamInterface { @@ -265,7 +265,7 @@ class ZipObjectInputStream : public io::InputStreamInterface { if (bytes_read < bytes_to_read) { return errors::OutOfRange("EOF reached"); } - return Status::OK(); + return OkStatus(); } virtual int64 Tell() const override { return offset_; } @@ -369,7 +369,7 @@ Status NumpyInfo(const string& filename, const int64 size, } } - return Status::OK(); + return OkStatus(); } class NumpyInfoOp : public OpKernel { @@ -681,7 +681,7 @@ class NumpyReadOp : public OpKernel { return errors::InvalidArgument("unsupported type: ", dtype); } - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/orc/orc_kernels.cc b/tensorflow_io/core/kernels/orc/orc_kernels.cc index c3b86ce9c..06021e94f 100644 --- a/tensorflow_io/core/kernels/orc/orc_kernels.cc +++ b/tensorflow_io/core/kernels/orc/orc_kernels.cc @@ -135,7 +135,7 @@ class ORCReadable : public IOReadableInterface { } } - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, const string& component, @@ -147,7 +147,7 @@ class ORCReadable : public IOReadableInterface { (*record_read) = 0; if (start >= shapes_[column_index].dim_size(0)) { - return Status::OK(); + return OkStatus(); } const string& column = component; int64 element_start = start < shapes_[column_index].dim_size(0) @@ -161,7 +161,7 @@ class ORCReadable : public IOReadableInterface { " selection is out of boundary"); } if (element_start == element_stop) { - return Status::OK(); + return OkStatus(); } #define PROCESS_VALUE(VTYPE) \ @@ -197,7 +197,7 @@ class ORCReadable : public IOReadableInterface { } (*record_read) = element_stop - element_start; - return Status::OK(); + return OkStatus(); } Status Components(std::vector* components) override { @@ -205,7 +205,7 @@ class ORCReadable : public IOReadableInterface { for (size_t i = 0; i < columns_.size(); i++) { components->push_back(columns_[i]); } - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, PartialTensorShape* shape, @@ -216,7 +216,7 @@ class ORCReadable : public IOReadableInterface { int64 column_index = columns_index_[component]; *shape = shapes_[column_index]; *dtype = dtypes_[column_index]; - return Status::OK(); + return OkStatus(); } string DebugString() const override { diff --git a/tensorflow_io/core/kernels/parquet_kernels.cc b/tensorflow_io/core/kernels/parquet_kernels.cc index 7a6cb76bc..22de43493 100644 --- a/tensorflow_io/core/kernels/parquet_kernels.cc +++ b/tensorflow_io/core/kernels/parquet_kernels.cc @@ -95,7 +95,7 @@ class ParquetReadableResource : public ResourceBase { .get() ->ToDotString()] = i; } - return Status::OK(); + return OkStatus(); } Status Components(std::vector* components) { @@ -105,7 +105,7 @@ class ParquetReadableResource : public ResourceBase { for (size_t i = 0; i < columns_.size(); i++) { components->push_back(columns_[i]); } - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, TensorShape* shape, DataType* dtype) { @@ -117,7 +117,7 @@ class ParquetReadableResource : public ResourceBase { int64 column_index = columns_index_[component]; *shape = shapes_[column_index]; *dtype = dtypes_[column_index]; - return Status::OK(); + return OkStatus(); } Status Read(const string& component, @@ -279,7 +279,7 @@ class ParquetReadableResource : public ResourceBase { } row_group_offset += row_group_reader->metadata()->num_rows(); } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "ParquetReadableResource"; } @@ -349,7 +349,7 @@ class ParquetReadableInfoOp } dtype_tensor->flat()(i) = dtypes[i]; } - return Status::OK(); + return OkStatus(); } }; @@ -406,9 +406,9 @@ class ParquetReadableReadOp component, start, shape, [&](const TensorShape& shape, Tensor** value) -> Status { TF_RETURN_IF_ERROR(context->allocate_output(0, shape, value)); - return Status::OK(); + return OkStatus(); })); - return Status::OK(); + return OkStatus(); } }; diff --git a/tensorflow_io/core/kernels/pcap_kernels.cc b/tensorflow_io/core/kernels/pcap_kernels.cc index 9ee28bc9a..b4c3f2904 100644 --- a/tensorflow_io/core/kernels/pcap_kernels.cc +++ b/tensorflow_io/core/kernels/pcap_kernels.cc @@ -63,7 +63,7 @@ class PcapInputStream : public io::BufferedInputStream { record_read = 1; // this method reads one packet at a time from the input buffer - return Status::OK(); + return OkStatus(); } Status ReadHeader() { @@ -83,7 +83,7 @@ class PcapInputStream : public io::BufferedInputStream { // multiple pcap packets EndianSwap(header->snaplen); } - return Status::OK(); + return OkStatus(); } private: @@ -197,20 +197,20 @@ class PcapReadable : public IOReadableInterface { record_index_ = 0; record_final_ = false; - return Status::OK(); + return OkStatus(); } Status Spec(const string& component, PartialTensorShape* shape, DataType* dtype, bool label) override { *shape = PartialTensorShape({-1}); *dtype = label ? DT_DOUBLE : DT_STRING; - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, const string& component, int64* record_read, Tensor* value, Tensor* label) override { (*record_read) = 0; if (record_final_) { - return Status::OK(); + return OkStatus(); } if (start != record_index_) { return errors::InvalidArgument( @@ -240,7 +240,7 @@ class PcapReadable : public IOReadableInterface { } record_index_ += (*record_read); - return Status::OK(); + return OkStatus(); } string DebugString() const override { diff --git a/tensorflow_io/core/kernels/prometheus_kernels.cc b/tensorflow_io/core/kernels/prometheus_kernels.cc index f916eac55..3b588f168 100644 --- a/tensorflow_io/core/kernels/prometheus_kernels.cc +++ b/tensorflow_io/core/kernels/prometheus_kernels.cc @@ -131,13 +131,13 @@ class PrometheusReadableResource : public ResourceBase { metrics->tensor()(index, 1) = instance; metrics->tensor()(index, 2) = name; } - return Status::OK(); + return OkStatus(); } Status Spec(int64* start, int64* stop) { mutex_lock l(mu_); *start = start_; *stop = stop_; - return Status::OK(); + return OkStatus(); } Status Read(const int64 start, const int64 stop, std::vector& jobs, std::vector& instances, std::vector& names, @@ -186,7 +186,7 @@ class PrometheusReadableResource : public ResourceBase { } } - return Status::OK(); + return OkStatus(); } string DebugString() const override { mutex_lock l(mu_); @@ -234,13 +234,13 @@ class PrometheusReadableInitOp input, metadata, [&](const TensorShape& shape, Tensor** metrics) -> Status { TF_RETURN_IF_ERROR(context->allocate_output(1, shape, metrics)); - return Status::OK(); + return OkStatus(); })); } Status CreateResource(PrometheusReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new PrometheusReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -320,7 +320,7 @@ class PrometheusReadableReadOp : public OpKernel { 0, timestamp_shape, timestamp)); TF_RETURN_IF_ERROR( context->allocate_output(1, value_shape, value)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/pubsub_kernels.cc b/tensorflow_io/core/kernels/pubsub_kernels.cc index e199c4b99..26777085d 100644 --- a/tensorflow_io/core/kernels/pubsub_kernels.cc +++ b/tensorflow_io/core/kernels/pubsub_kernels.cc @@ -74,7 +74,7 @@ class PubSubReadableResource : public ResourceBase { } stub_ = Subscriber::NewStub(grpc::CreateChannel(endpoint, creds)); - return Status::OK(); + return OkStatus(); } Status Read(std::function @@ -108,7 +108,7 @@ class PubSubReadableResource : public ResourceBase { TF_RETURN_IF_ERROR(allocate_func(TensorShape({0}), &id_tensor, &data_tensor, &time_tensor)); stub_.reset(nullptr); - return Status::OK(); + return OkStatus(); } if (response.received_messages().size() != 0) { TF_RETURN_IF_ERROR(allocate_func(TensorShape({1}), &id_tensor, @@ -131,10 +131,10 @@ class PubSubReadableResource : public ResourceBase { ClientContext ack_context; status = stub_->Acknowledge(&ack_context, acknowledge, &empty); - return Status::OK(); + return OkStatus(); } } - return Status::OK(); + return OkStatus(); } string DebugString() const override { mutex_lock l(mu_); @@ -180,7 +180,7 @@ class PubSubReadableInitOp : public ResourceOpKernel { Status CreateResource(PubSubReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new PubSubReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -208,7 +208,7 @@ class PubSubReadableReadOp : public OpKernel { TF_RETURN_IF_ERROR(context->allocate_output(0, shape, id_tensor)); TF_RETURN_IF_ERROR(context->allocate_output(1, shape, data_tensor)); TF_RETURN_IF_ERROR(context->allocate_output(2, shape, time_tensor)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/pulsar_kernel.cc b/tensorflow_io/core/kernels/pulsar_kernel.cc index bc35b36c7..d6474ec71 100644 --- a/tensorflow_io/core/kernels/pulsar_kernel.cc +++ b/tensorflow_io/core/kernels/pulsar_kernel.cc @@ -62,7 +62,7 @@ class PulsarReadableResource final : public PulsarResourceBase { LOG(INFO) << "Subscribing to the pulsar topic: " << topic << " with subscription: " << subscription; - return Status::OK(); + return OkStatus(); } Status Next(const int32 timeout, const int32 poll_timeout, @@ -116,7 +116,7 @@ class PulsarReadableResource final : public PulsarResourceBase { key_tensor->flat()(i) = keys[i]; } - return Status::OK(); + return OkStatus(); } std::string DebugString() const override { return "PulsarReadableResource"; } @@ -160,7 +160,7 @@ class PulsarReadableInitOp : public ResourceOpKernel { Status CreateResource(PulsarReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new PulsarReadableResource(); - return Status::OK(); + return OkStatus(); } private: @@ -197,7 +197,7 @@ class PulsarReadableNextOp : public OpKernel { TF_RETURN_IF_ERROR(context->allocate_output(1, shape, key)); TF_RETURN_IF_ERROR( context->allocate_output(2, TensorShape({}), continue_fetch)); - return Status::OK(); + return OkStatus(); })); } @@ -223,7 +223,7 @@ class PulsarWritableResource final : public PulsarResourceBase { } LOG(INFO) << "Created producer on pulsar topic: " << topic; - return Status::OK(); + return OkStatus(); } Status WriteAsync(const std::string& value, const std::string& key) { @@ -248,7 +248,7 @@ class PulsarWritableResource final : public PulsarResourceBase { " error: ", pulsar::strResult(send_async_result)); } index_++; - return Status::OK(); + return OkStatus(); } Status Flush() { @@ -257,7 +257,7 @@ class PulsarWritableResource final : public PulsarResourceBase { if (result != pulsar::ResultOk) { return errors::Internal("failed to flush: ", pulsar::strResult(result)); } - return Status::OK(); + return OkStatus(); } std::string DebugString() const override { return "PulsarWritableResource"; } @@ -290,7 +290,7 @@ class PulsarWritableInitOp : public ResourceOpKernel { Status CreateResource(PulsarWritableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new PulsarWritableResource(); - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/serialization_kernels.cc b/tensorflow_io/core/kernels/serialization_kernels.cc index 86a3c248f..020df3961 100644 --- a/tensorflow_io/core/kernels/serialization_kernels.cc +++ b/tensorflow_io/core/kernels/serialization_kernels.cc @@ -235,7 +235,7 @@ class DecodeAvroOp : public OpKernel { datum.type()); } - return Status::OK(); + return OkStatus(); } Status ProcessRecord(const int64 index, @@ -247,7 +247,7 @@ class DecodeAvroOp : public OpKernel { const avro::GenericDatum& field = record.fieldAt(i); TF_RETURN_IF_ERROR(ProcessEntry(index, values, entry, field)); } - return Status::OK(); + return OkStatus(); } Status ProcessPrimitive(const int64 index, std::unordered_map& values, @@ -310,7 +310,7 @@ class DecodeAvroOp : public OpKernel { return errors::InvalidArgument("data type not supported: ", datum.type()); } - return Status::OK(); + return OkStatus(); } Status ProcessNull(const int64 index, std::unordered_map& values, @@ -344,7 +344,7 @@ class DecodeAvroOp : public OpKernel { return errors::InvalidArgument("data type not supported: ", value_tensor->dtype()); } - return Status::OK(); + return OkStatus(); } private: @@ -436,7 +436,7 @@ class EncodeAvroOp : public OpKernel { datum.type()); } - return Status::OK(); + return OkStatus(); } Status ProcessRecord(const int64 index, @@ -448,7 +448,7 @@ class EncodeAvroOp : public OpKernel { avro::GenericDatum& field = record.fieldAt(i); TF_RETURN_IF_ERROR(ProcessEntry(index, values, entry, field)); } - return Status::OK(); + return OkStatus(); } Status ProcessPrimitive( const int64 index, @@ -511,7 +511,7 @@ class EncodeAvroOp : public OpKernel { return errors::InvalidArgument("data type not supported: ", datum.type()); } - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/sql_kernels.cc b/tensorflow_io/core/kernels/sql_kernels.cc index fd554d234..92a224ad0 100644 --- a/tensorflow_io/core/kernels/sql_kernels.cc +++ b/tensorflow_io/core/kernels/sql_kernels.cc @@ -58,7 +58,7 @@ Status SqlDataCopy(int oid, char* data, Tensor* value) { if (!result) { return errors::InvalidArgument("unable to convert data"); } - return Status::OK(); + return OkStatus(); }; class SqlIterableResource : public ResourceBase { @@ -123,7 +123,7 @@ class SqlIterableResource : public ResourceBase { dtypes->push_back(dtypes_[i]); } - return Status::OK(); + return OkStatus(); } Status Read(const int64 index, const Tensor* field_tensor, std::function @@ -138,7 +138,7 @@ class SqlIterableResource : public ResourceBase { char* data = PQgetvalue(result_.get(), index, field_number); TF_RETURN_IF_ERROR(SqlDataCopy(field_type, data, value)); } - return Status::OK(); + return OkStatus(); } string DebugString() const override { return "SqlIterableResource"; } @@ -202,7 +202,7 @@ class SqlIterableInitOp : public ResourceOpKernel { Status CreateResource(SqlIterableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new SqlIterableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -236,7 +236,7 @@ class SqlIterableReadOp : public OpKernel { [&](const int64 field_index, Tensor** value) -> Status { TF_RETURN_IF_ERROR(context->allocate_output( field_index, TensorShape({1}), value)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/kernels/tests/bigquery_test_client_op.cc b/tensorflow_io/core/kernels/tests/bigquery_test_client_op.cc index 076ae8187..51dda0399 100644 --- a/tensorflow_io/core/kernels/tests/bigquery_test_client_op.cc +++ b/tensorflow_io/core/kernels/tests/bigquery_test_client_op.cc @@ -67,7 +67,7 @@ class BigQueryTestClientOp : public OpKernel { return absl::make_unique( channel); }); - return Status::OK(); + return OkStatus(); })); core::ScopedUnref unref_resource(resource); diff --git a/tensorflow_io/core/kernels/text_kernels.cc b/tensorflow_io/core/kernels/text_kernels.cc index dea82134f..f96f952d8 100644 --- a/tensorflow_io/core/kernels/text_kernels.cc +++ b/tensorflow_io/core/kernels/text_kernels.cc @@ -63,7 +63,7 @@ class FilenoInputStream : public io::InputStreamInterface { if (bytes_read < bytes_to_read) { return errors::OutOfRange("EOF reached"); } - return Status::OK(); + return OkStatus(); } virtual int64 Tell() const override { return offset_; } @@ -110,7 +110,7 @@ class ReadTextOp : public OpKernel { std::unique_ptr stream( new tensorflow::io::BufferedInputStream(input_stream.get(), 4096)); - Status status = Status::OK(); + Status status = OkStatus(); while (status.ok()) { string line; status = stream->ReadLine(&line); diff --git a/tensorflow_io/core/kernels/text_sequence.cc b/tensorflow_io/core/kernels/text_sequence.cc index ef80accfa..2eeb72b94 100644 --- a/tensorflow_io/core/kernels/text_sequence.cc +++ b/tensorflow_io/core/kernels/text_sequence.cc @@ -39,7 +39,7 @@ class TextOutputSequence : public OutputSequence { } TF_RETURN_IF_ERROR(file->Close()); } - return Status::OK(); + return OkStatus(); } #if TF_MAJOR_VERSION == 1 && TF_MINOR_VERSION == 13 virtual string DebugString() { @@ -54,7 +54,7 @@ class TextOutputSequence : public OutputSequence { return errors::Unimplemented("only one file is supported: ", destination_.size()); } - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/video_kernels.cc b/tensorflow_io/core/kernels/video_kernels.cc index e2f668951..88d4ac8d2 100644 --- a/tensorflow_io/core/kernels/video_kernels.cc +++ b/tensorflow_io/core/kernels/video_kernels.cc @@ -88,7 +88,7 @@ class VideoCaptureReadableResource : public ResourceBase { bytes_ = static_cast(bytes); width_ = static_cast(width); height_ = static_cast(height); - return Status::OK(); + return OkStatus(); } Status Read( std::function @@ -104,7 +104,7 @@ class VideoCaptureReadableResource : public ResourceBase { static_cast(bytes_)); value_tensor->flat()(0) = buffer; - return Status::OK(); + return OkStatus(); } string DebugString() const override { mutex_lock l(mu_); @@ -142,7 +142,7 @@ class VideoCaptureReadableInitOp Status CreateResource(VideoCaptureReadableResource** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new VideoCaptureReadableResource(env_); - return Status::OK(); + return OkStatus(); } private: @@ -167,7 +167,7 @@ class VideoCaptureReadableReadOp : public OpKernel { context, resource->Read([&](const TensorShape& shape, Tensor** value_tensor) -> Status { TF_RETURN_IF_ERROR(context->allocate_output(0, shape, value_tensor)); - return Status::OK(); + return OkStatus(); })); } diff --git a/tensorflow_io/core/ops/arrow_ops.cc b/tensorflow_io/core/ops/arrow_ops.cc index ceff0a7c7..34d027032 100644 --- a/tensorflow_io/core/ops/arrow_ops.cc +++ b/tensorflow_io/core/ops/arrow_ops.cc @@ -96,7 +96,7 @@ REGISTER_OP("IO>ListFeatherColumns") c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim()})); c->set_output(2, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FeatherReadableInit") @@ -108,7 +108,7 @@ REGISTER_OP("IO>FeatherReadableInit") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FeatherReadableSpec") @@ -119,7 +119,7 @@ REGISTER_OP("IO>FeatherReadableSpec") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FeatherReadableRead") @@ -136,7 +136,7 @@ REGISTER_OP("IO>FeatherReadableRead") shape_inference::ShapeHandle entry; TF_RETURN_IF_ERROR(c->MakeShapeFromPartialTensorShape(shape, &entry)); c->set_output(0, entry); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>ArrowReadableFromMemoryInit") @@ -150,7 +150,7 @@ REGISTER_OP("IO>ArrowReadableFromMemoryInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>ArrowReadableSpec") @@ -162,7 +162,7 @@ REGISTER_OP("IO>ArrowReadableSpec") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>ArrowReadableRead") @@ -179,13 +179,13 @@ REGISTER_OP("IO>ArrowReadableRead") TF_RETURN_IF_ERROR(c->MakeShapeFromShapeTensor(3, &full)); if (!(c->RankKnown(full) && c->Rank(full) > 0)) { c->set_output(0, full); - return Status::OK(); + return OkStatus(); } // TODO: replace dims up until rank(start|stop) shape_inference::ShapeHandle shape; TF_RETURN_IF_ERROR(c->ReplaceDim(full, 0, c->UnknownDim(), &shape)); c->set_output(0, shape); - return Status::OK(); + return OkStatus(); }); } // namespace tensorflow diff --git a/tensorflow_io/core/ops/audio_ops.cc b/tensorflow_io/core/ops/audio_ops.cc index c34ce0e79..2ed30e7f0 100644 --- a/tensorflow_io/core/ops/audio_ops.cc +++ b/tensorflow_io/core/ops/audio_ops.cc @@ -28,7 +28,7 @@ REGISTER_OP("IO>AudioReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioReadableSpec") @@ -40,7 +40,7 @@ REGISTER_OP("IO>AudioReadableSpec") c->set_output(0, c->MakeShape({2})); c->set_output(1, c->MakeShape({})); c->set_output(2, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioReadableRead") @@ -51,7 +51,7 @@ REGISTER_OP("IO>AudioReadableRead") .Attr("dtype: type") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioResample") @@ -62,7 +62,7 @@ REGISTER_OP("IO>AudioResample") .Attr("T: type") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioDecodeWAV") @@ -75,14 +75,14 @@ REGISTER_OP("IO>AudioDecodeWAV") TF_RETURN_IF_ERROR(c->MakeShapeFromShapeTensor(1, &shape)); if (!c->RankKnown(shape)) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); } if (c->Rank(shape) != 2) { return errors::InvalidArgument("rank must be two, received ", c->DebugString(shape)); } c->set_output(0, shape); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioEncodeWAV") @@ -92,7 +92,7 @@ REGISTER_OP("IO>AudioEncodeWAV") .Attr("dtype: {uint8, int16, int32, float32}") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioDecodeFlac") @@ -105,14 +105,14 @@ REGISTER_OP("IO>AudioDecodeFlac") TF_RETURN_IF_ERROR(c->MakeShapeFromShapeTensor(1, &shape)); if (!c->RankKnown(shape)) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); } if (c->Rank(shape) != 2) { return errors::InvalidArgument("rank must be two, received ", c->DebugString(shape)); } c->set_output(0, shape); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioEncodeFlac") @@ -122,7 +122,7 @@ REGISTER_OP("IO>AudioEncodeFlac") .Attr("dtype: {uint8, int16, int32}") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioDecodeVorbis") @@ -134,14 +134,14 @@ REGISTER_OP("IO>AudioDecodeVorbis") TF_RETURN_IF_ERROR(c->MakeShapeFromShapeTensor(1, &shape)); if (!c->RankKnown(shape)) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); } if (c->Rank(shape) != 2) { return errors::InvalidArgument("rank must be two, received ", c->DebugString(shape)); } c->set_output(0, shape); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioEncodeVorbis") @@ -150,7 +150,7 @@ REGISTER_OP("IO>AudioEncodeVorbis") .Output("value: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioDecodeMP3") @@ -162,14 +162,14 @@ REGISTER_OP("IO>AudioDecodeMP3") TF_RETURN_IF_ERROR(c->MakeShapeFromShapeTensor(1, &shape)); if (!c->RankKnown(shape)) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); } if (c->Rank(shape) != 2) { return errors::InvalidArgument("rank must be two, received ", c->DebugString(shape)); } c->set_output(0, shape); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioEncodeMP3") @@ -178,7 +178,7 @@ REGISTER_OP("IO>AudioEncodeMP3") .Output("value: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioDecodeAAC") @@ -190,14 +190,14 @@ REGISTER_OP("IO>AudioDecodeAAC") TF_RETURN_IF_ERROR(c->MakeShapeFromShapeTensor(1, &shape)); if (!c->RankKnown(shape)) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); } if (c->Rank(shape) != 2) { return errors::InvalidArgument("rank must be two, received ", c->DebugString(shape)); } c->set_output(0, shape); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AudioEncodeAAC") @@ -206,7 +206,7 @@ REGISTER_OP("IO>AudioEncodeAAC") .Output("value: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/avro_ops.cc b/tensorflow_io/core/ops/avro_ops.cc index 34a8b19fb..658bdeb3d 100644 --- a/tensorflow_io/core/ops/avro_ops.cc +++ b/tensorflow_io/core/ops/avro_ops.cc @@ -37,7 +37,7 @@ Status AddDenseOutputShapes(const std::vector& dense_shapes, TF_RETURN_IF_ERROR(c->Concatenate(prefix, s, &s)); c->set_output((*output_idx)++, s); } - return Status::OK(); + return OkStatus(); } shape_inference::DimensionOrConstant ComputeSparseRank( @@ -164,7 +164,7 @@ REGISTER_OP("IO>ParseAvro") AddSparseOutputShapes(num_sparse, input, sparse_ranks, c, &output_idx); TF_RETURN_IF_ERROR( AddDenseOutputShapes(dense_shapes, input, c, &output_idx)); - return Status::OK(); + return OkStatus(); }); // Adjusted from TFRecordDataset here @@ -255,7 +255,7 @@ REGISTER_OP("IO>ListAvroColumns") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>ReadAvro") @@ -269,7 +269,7 @@ REGISTER_OP("IO>ReadAvro") .Output("output: dtype") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AvroReadableInit") @@ -282,7 +282,7 @@ REGISTER_OP("IO>AvroReadableInit") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AvroReadableSpec") @@ -293,7 +293,7 @@ REGISTER_OP("IO>AvroReadableSpec") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AvroReadableRead") @@ -311,7 +311,7 @@ REGISTER_OP("IO>AvroReadableRead") shape_inference::ShapeHandle entry; TF_RETURN_IF_ERROR(c->MakeShapeFromPartialTensorShape(shape, &entry)); c->set_output(0, entry); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>AvroReadablePartitions") @@ -319,7 +319,7 @@ REGISTER_OP("IO>AvroReadablePartitions") .Output("partitions: int64") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); } // namespace tensorflow diff --git a/tensorflow_io/core/ops/bigquery_ops.cc b/tensorflow_io/core/ops/bigquery_ops.cc index 9d2c1d20d..8a148cb67 100644 --- a/tensorflow_io/core/ops/bigquery_ops.cc +++ b/tensorflow_io/core/ops/bigquery_ops.cc @@ -43,7 +43,7 @@ REGISTER_OP("IO>BigQueryReadSession") .SetShapeFn([](tensorflow::shape_inference::InferenceContext* c) { c->set_output(0, c->Vector(c->UnknownDim())); c->set_output(1, c->Scalar()); - return tensorflow::Status::OK(); + return tensorflow::OkStatus(); }); REGISTER_OP("IO>BigQueryDataset") diff --git a/tensorflow_io/core/ops/bigtable_ops.cc b/tensorflow_io/core/ops/bigtable_ops.cc index 8d57aa134..b67dc1289 100644 --- a/tensorflow_io/core/ops/bigtable_ops.cc +++ b/tensorflow_io/core/ops/bigtable_ops.cc @@ -162,7 +162,7 @@ REGISTER_OP("BigtableSplitRowSetEvenly") .SetIsStateful() .SetShapeFn([](tensorflow::shape_inference::InferenceContext* c) { c->set_output(0, c->Vector(c->UnknownDim())); - return tensorflow::Status::OK(); + return tensorflow::OkStatus(); }) .Doc(R"doc( Retrieves SampleRowKeys from bigtable, checks which tablets contain row keys diff --git a/tensorflow_io/core/ops/core_ops.cc b/tensorflow_io/core/ops/core_ops.cc index dc9044e60..7b4a13eae 100644 --- a/tensorflow_io/core/ops/core_ops.cc +++ b/tensorflow_io/core/ops/core_ops.cc @@ -28,7 +28,7 @@ REGISTER_OP("IO>ListArchiveEntries") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>ReadArchive") @@ -39,7 +39,7 @@ REGISTER_OP("IO>ReadArchive") .Output("output: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); } // namespace tensorflow diff --git a/tensorflow_io/core/ops/elasticsearch_ops.cc b/tensorflow_io/core/ops/elasticsearch_ops.cc index 98e520553..f7d7d7a5b 100644 --- a/tensorflow_io/core/ops/elasticsearch_ops.cc +++ b/tensorflow_io/core/ops/elasticsearch_ops.cc @@ -34,7 +34,7 @@ REGISTER_OP("IO>ElasticsearchReadableInit") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>ElasticsearchReadableNext") @@ -44,7 +44,7 @@ REGISTER_OP("IO>ElasticsearchReadableNext") .Output("items: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/ffmpeg_ops.cc b/tensorflow_io/core/ops/ffmpeg_ops.cc index dcb976d36..4fddd94d1 100644 --- a/tensorflow_io/core/ops/ffmpeg_ops.cc +++ b/tensorflow_io/core/ops/ffmpeg_ops.cc @@ -28,7 +28,7 @@ REGISTER_OP("IO>FfmpegReadableInit") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FfmpegReadableSpec") @@ -41,7 +41,7 @@ REGISTER_OP("IO>FfmpegReadableSpec") c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({})); c->set_output(2, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FfmpegReadableRead") @@ -58,7 +58,7 @@ REGISTER_OP("IO>FfmpegReadableRead") shape_inference::ShapeHandle entry; TF_RETURN_IF_ERROR(c->MakeShapeFromPartialTensorShape(shape, &entry)); c->set_output(0, entry); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FfmpegDecodeVideo") @@ -68,7 +68,7 @@ REGISTER_OP("IO>FfmpegDecodeVideo") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim(), c->UnknownDim(), 3})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FfmpegAudioReadableInit") @@ -79,7 +79,7 @@ REGISTER_OP("IO>FfmpegAudioReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FfmpegAudioReadableNext") @@ -89,7 +89,7 @@ REGISTER_OP("IO>FfmpegAudioReadableNext") .Attr("dtype: type") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FfmpegVideoReadableInit") @@ -100,7 +100,7 @@ REGISTER_OP("IO>FfmpegVideoReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FfmpegVideoReadableNext") @@ -111,7 +111,7 @@ REGISTER_OP("IO>FfmpegVideoReadableNext") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim(), c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); } // namespace tensorflow diff --git a/tensorflow_io/core/ops/file_ops.cc b/tensorflow_io/core/ops/file_ops.cc index 37a3f8eb9..5f9445b0f 100644 --- a/tensorflow_io/core/ops/file_ops.cc +++ b/tensorflow_io/core/ops/file_ops.cc @@ -28,7 +28,7 @@ REGISTER_OP("IO>FileInfo") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FileRead") @@ -39,7 +39,7 @@ REGISTER_OP("IO>FileRead") .Output("value: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>FileInit") diff --git a/tensorflow_io/core/ops/genome_ops.cc b/tensorflow_io/core/ops/genome_ops.cc index bbb7fb771..7c985f525 100644 --- a/tensorflow_io/core/ops/genome_ops.cc +++ b/tensorflow_io/core/ops/genome_ops.cc @@ -26,7 +26,7 @@ REGISTER_OP("IO>ReadFastq") .SetShapeFn([](::tensorflow::shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); } // namespace tensorflow diff --git a/tensorflow_io/core/ops/grpc_ops.cc b/tensorflow_io/core/ops/grpc_ops.cc index 5652930cf..b5a88ba6c 100644 --- a/tensorflow_io/core/ops/grpc_ops.cc +++ b/tensorflow_io/core/ops/grpc_ops.cc @@ -29,7 +29,7 @@ REGISTER_OP("IO>GRPCReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>GRPCReadableRead") @@ -43,7 +43,7 @@ REGISTER_OP("IO>GRPCReadableRead") shape_inference::ShapeHandle shape; TF_RETURN_IF_ERROR(c->MakeShapeFromShapeTensor(2, &shape)); c->set_output(0, shape); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/hdf5_ops.cc b/tensorflow_io/core/ops/hdf5_ops.cc index 6e8219c7c..3ba3950c8 100644 --- a/tensorflow_io/core/ops/hdf5_ops.cc +++ b/tensorflow_io/core/ops/hdf5_ops.cc @@ -32,7 +32,7 @@ REGISTER_OP("IO>HDF5ReadableInfo") c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); c->set_output(2, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>HDF5ReadableRead") @@ -50,13 +50,13 @@ REGISTER_OP("IO>HDF5ReadableRead") TF_RETURN_IF_ERROR(c->MakeShapeFromShapeTensor(3, &full)); if (!(c->RankKnown(full) && c->Rank(full) > 0)) { c->set_output(0, full); - return Status::OK(); + return OkStatus(); } // TODO: replace dims up until rank(start|stop) shape_inference::ShapeHandle shape; TF_RETURN_IF_ERROR(c->ReplaceDim(full, 0, c->UnknownDim(), &shape)); c->set_output(0, shape); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/image_ops.cc b/tensorflow_io/core/ops/image_ops.cc index 43f899e62..09a1fe2b4 100644 --- a/tensorflow_io/core/ops/image_ops.cc +++ b/tensorflow_io/core/ops/image_ops.cc @@ -30,7 +30,7 @@ REGISTER_OP("IO>DecodeTiffInfo") TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused)); c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodeTiff") @@ -43,7 +43,7 @@ REGISTER_OP("IO>DecodeTiff") TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 0, &unused)); c->set_output( 0, c->MakeShape({c->UnknownDim(), c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>EncodeBmp") @@ -53,7 +53,7 @@ REGISTER_OP("IO>EncodeBmp") shape_inference::ShapeHandle unused; TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 3, &unused)); c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodeWebP") @@ -65,7 +65,7 @@ REGISTER_OP("IO>DecodeWebP") c->set_output( 0, c->MakeShape({shape_inference::InferenceContext::kUnknownDim, shape_inference::InferenceContext::kUnknownDim, 4})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodePnm") @@ -77,7 +77,7 @@ REGISTER_OP("IO>DecodePnm") TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused)); c->set_output( 0, c->MakeShape({c->UnknownDim(), c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DrawBoundingBoxesV3") @@ -99,7 +99,7 @@ REGISTER_OP("IO>DecodeJpegExif") shape_inference::ShapeHandle unused; TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused)); c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodeExrInfo") @@ -113,7 +113,7 @@ REGISTER_OP("IO>DecodeExrInfo") c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); c->set_output(2, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodeExr") @@ -129,7 +129,7 @@ REGISTER_OP("IO>DecodeExr") TF_RETURN_IF_ERROR(c->WithRank(c->input(2), 0, &unused)); c->set_output( 0, c->MakeShape({c->UnknownDim(), c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodeHdr") @@ -139,7 +139,7 @@ REGISTER_OP("IO>DecodeHdr") shape_inference::ShapeHandle unused; TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused)); c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim(), 3})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodeDICOMImage") @@ -154,7 +154,7 @@ REGISTER_OP("IO>DecodeDICOMImage") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim(), c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }) .Doc(R"doc( loads a dicom image file and returns its pixel information in the specified output format @@ -167,7 +167,7 @@ REGISTER_OP("IO>DecodeDICOMData") .Output("tag_values: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->input(1)); - return Status::OK(); + return OkStatus(); }) .Doc(R"doc( loads a dicom file and returns the specified tags values as string. @@ -181,7 +181,7 @@ REGISTER_OP("IO>DecodeNV12") shape_inference::ShapeHandle unused; TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused)); c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim(), 3})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodeYUY2") @@ -192,7 +192,7 @@ REGISTER_OP("IO>DecodeYUY2") shape_inference::ShapeHandle unused; TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused)); c->set_output(0, c->MakeShape({c->UnknownDim(), c->UnknownDim(), 3})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodeAVIF") @@ -204,7 +204,7 @@ REGISTER_OP("IO>DecodeAVIF") c->set_output( 0, c->MakeShape({shape_inference::InferenceContext::kUnknownDim, shape_inference::InferenceContext::kUnknownDim, 4})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodeJPEG2K") @@ -218,7 +218,7 @@ REGISTER_OP("IO>DecodeJPEG2K") 0, c->MakeShape({shape_inference::InferenceContext::kUnknownDim, shape_inference::InferenceContext::kUnknownDim, shape_inference::InferenceContext::kUnknownDim})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>EncodeGif") @@ -228,7 +228,7 @@ REGISTER_OP("IO>EncodeGif") shape_inference::ShapeHandle unused; TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 4, &unused)); c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/json_ops.cc b/tensorflow_io/core/ops/json_ops.cc index c80cc1647..5255bb32d 100644 --- a/tensorflow_io/core/ops/json_ops.cc +++ b/tensorflow_io/core/ops/json_ops.cc @@ -31,7 +31,7 @@ REGISTER_OP("IO>JSONReadableInit") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>JSONReadableSpec") @@ -42,7 +42,7 @@ REGISTER_OP("IO>JSONReadableSpec") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>JSONReadableRead") @@ -59,7 +59,7 @@ REGISTER_OP("IO>JSONReadableRead") shape_inference::ShapeHandle entry; TF_RETURN_IF_ERROR(c->MakeShapeFromPartialTensorShape(shape, &entry)); c->set_output(0, entry); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/kafka_ops.cc b/tensorflow_io/core/ops/kafka_ops.cc index dae76ae91..403ee0240 100644 --- a/tensorflow_io/core/ops/kafka_ops.cc +++ b/tensorflow_io/core/ops/kafka_ops.cc @@ -31,7 +31,7 @@ REGISTER_OP("IO>KafkaReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KafkaReadableNext") @@ -42,7 +42,7 @@ REGISTER_OP("IO>KafkaReadableNext") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KafkaReadableRead") @@ -54,7 +54,7 @@ REGISTER_OP("IO>KafkaReadableRead") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KafkaReadableSpec") @@ -66,7 +66,7 @@ REGISTER_OP("IO>KafkaReadableSpec") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KafkaIterableInit") @@ -79,7 +79,7 @@ REGISTER_OP("IO>KafkaIterableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>LayerKafkaCall") @@ -111,7 +111,7 @@ REGISTER_OP("IO>KafkaGroupReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KafkaGroupReadableNext") @@ -126,7 +126,7 @@ REGISTER_OP("IO>KafkaGroupReadableNext") c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim()})); c->set_output(2, c->Scalar()); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/kafka_ops_deprecated.cc b/tensorflow_io/core/ops/kafka_ops_deprecated.cc index b0da1e9d0..c24681e8f 100644 --- a/tensorflow_io/core/ops/kafka_ops_deprecated.cc +++ b/tensorflow_io/core/ops/kafka_ops_deprecated.cc @@ -63,7 +63,7 @@ REGISTER_OP("IO>WriteKafka") TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 0, &unused)); TF_RETURN_IF_ERROR(c->WithRank(c->input(2), 0, &unused)); c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KafkaEncodeAvro") @@ -73,7 +73,7 @@ REGISTER_OP("IO>KafkaEncodeAvro") .Attr("dtype: list({float,double,int32,int64,string})") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->input(0)); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KafkaDecodeAvroInit") @@ -83,7 +83,7 @@ REGISTER_OP("IO>KafkaDecodeAvroInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KafkaDecodeAvro") @@ -96,7 +96,7 @@ REGISTER_OP("IO>KafkaDecodeAvro") for (int64 i = 0; i < c->num_outputs(); i++) { c->set_output(i, c->input(0)); } - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KafkaOutputSequence") @@ -111,7 +111,7 @@ REGISTER_OP("IO>KafkaOutputSequence") TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused)); TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 0, &unused)); c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KafkaOutputSequenceSetItem") diff --git a/tensorflow_io/core/ops/kinesis_ops.cc b/tensorflow_io/core/ops/kinesis_ops.cc index e6d5e3706..c7d14b0bd 100644 --- a/tensorflow_io/core/ops/kinesis_ops.cc +++ b/tensorflow_io/core/ops/kinesis_ops.cc @@ -30,7 +30,7 @@ REGISTER_OP("IO>KinesisReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>KinesisReadableRead") @@ -45,7 +45,7 @@ REGISTER_OP("IO>KinesisReadableRead") c->set_output(1, c->MakeShape({c->UnknownDim()})); c->set_output(2, c->MakeShape({c->UnknownDim()})); c->set_output(3, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/libsvm_ops.cc b/tensorflow_io/core/ops/libsvm_ops.cc index 4b4ba378f..f0c8629a1 100644 --- a/tensorflow_io/core/ops/libsvm_ops.cc +++ b/tensorflow_io/core/ops/libsvm_ops.cc @@ -38,7 +38,7 @@ REGISTER_OP("IO>DecodeLibsvm") c->set_output(2, c->Vector(InferenceContext::kUnknownDim)); c->set_output(3, c->Vector(InferenceContext::kUnknownDim)); - return Status::OK(); + return OkStatus(); }) .Doc(R"doc( diff --git a/tensorflow_io/core/ops/lmdb_ops.cc b/tensorflow_io/core/ops/lmdb_ops.cc index d66bf8fec..adfa7ed11 100644 --- a/tensorflow_io/core/ops/lmdb_ops.cc +++ b/tensorflow_io/core/ops/lmdb_ops.cc @@ -27,7 +27,7 @@ REGISTER_OP("IO>LMDBInput") .Attr("schema: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>LMDBDatasetV2") @@ -40,7 +40,7 @@ REGISTER_OP("IO>LMDBDatasetV2") .SetIsStateful() .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>LMDBReadableInit") @@ -50,7 +50,7 @@ REGISTER_OP("IO>LMDBReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>LMDBReadableRead") @@ -66,7 +66,7 @@ REGISTER_OP("IO>LMDBReadableRead") shape_inference::ShapeHandle entry; TF_RETURN_IF_ERROR(c->MakeShapeFromPartialTensorShape(shape, &entry)); c->set_output(0, entry); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>LMDBMappingInit") @@ -77,7 +77,7 @@ REGISTER_OP("IO>LMDBMappingInit") .SetIsStateful() .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>LMDBMappingRead") @@ -86,7 +86,7 @@ REGISTER_OP("IO>LMDBMappingRead") .Output("value: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->input(1)); - return Status::OK(); + return OkStatus(); }); } // namespace tensorflow diff --git a/tensorflow_io/core/ops/mongodb_ops.cc b/tensorflow_io/core/ops/mongodb_ops.cc index 57b607e3a..a08883c06 100644 --- a/tensorflow_io/core/ops/mongodb_ops.cc +++ b/tensorflow_io/core/ops/mongodb_ops.cc @@ -34,7 +34,7 @@ REGISTER_OP("IO>MongoDBReadableNext") .Output("record: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>MongoDBWritableInit") diff --git a/tensorflow_io/core/ops/numpy_ops.cc b/tensorflow_io/core/ops/numpy_ops.cc index 773ac56d8..5d9b6af2e 100644 --- a/tensorflow_io/core/ops/numpy_ops.cc +++ b/tensorflow_io/core/ops/numpy_ops.cc @@ -30,7 +30,7 @@ REGISTER_OP("IO>NumpyInfo") c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); c->set_output(2, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>NumpySpec") @@ -41,7 +41,7 @@ REGISTER_OP("IO>NumpySpec") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>NumpyRead") @@ -59,7 +59,7 @@ REGISTER_OP("IO>NumpyRead") shape_inference::ShapeHandle shape; TF_RETURN_IF_ERROR(c->ReplaceDim(full, 0, c->UnknownDim(), &shape)); c->set_output(0, shape); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/obj_ops.cc b/tensorflow_io/core/ops/obj_ops.cc index e3c45653a..c5a6973bd 100644 --- a/tensorflow_io/core/ops/obj_ops.cc +++ b/tensorflow_io/core/ops/obj_ops.cc @@ -28,7 +28,7 @@ REGISTER_OP("IO>DecodeObj") shape_inference::ShapeHandle unused; TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused)); c->set_output(0, c->MakeShape({c->UnknownDim(), 3})); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/operation_ops.cc b/tensorflow_io/core/ops/operation_ops.cc index 643591314..b1cdc6b69 100644 --- a/tensorflow_io/core/ops/operation_ops.cc +++ b/tensorflow_io/core/ops/operation_ops.cc @@ -36,11 +36,11 @@ REGISTER_OP("IO>OrderIndices") dims.emplace_back(c->UnknownDim()); } c->set_output(0, c->MakeShape(dims)); - return Status::OK(); + return OkStatus(); } c->set_output(0, c->UnknownShape()); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/orc_ops.cc b/tensorflow_io/core/ops/orc_ops.cc index 9bc292459..f75915f06 100644 --- a/tensorflow_io/core/ops/orc_ops.cc +++ b/tensorflow_io/core/ops/orc_ops.cc @@ -27,7 +27,7 @@ REGISTER_OP("IO>ORCReadableInit") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>ORCReadableSpec") @@ -38,7 +38,7 @@ REGISTER_OP("IO>ORCReadableSpec") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>ORCReadableRead") @@ -55,6 +55,6 @@ REGISTER_OP("IO>ORCReadableRead") shape_inference::ShapeHandle entry; TF_RETURN_IF_ERROR(c->MakeShapeFromPartialTensorShape(shape, &entry)); c->set_output(0, entry); - return Status::OK(); + return OkStatus(); }); } // namespace tensorflow \ No newline at end of file diff --git a/tensorflow_io/core/ops/parquet_ops.cc b/tensorflow_io/core/ops/parquet_ops.cc index 320983037..60d14b20e 100644 --- a/tensorflow_io/core/ops/parquet_ops.cc +++ b/tensorflow_io/core/ops/parquet_ops.cc @@ -32,7 +32,7 @@ REGISTER_OP("IO>ParquetReadableInfo") c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); c->set_output(2, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>ParquetReadableRead") @@ -50,13 +50,13 @@ REGISTER_OP("IO>ParquetReadableRead") TF_RETURN_IF_ERROR(c->MakeShapeFromShapeTensor(3, &full)); if (!(c->RankKnown(full) && c->Rank(full) > 0)) { c->set_output(0, full); - return Status::OK(); + return OkStatus(); } // TODO: replace dims up until rank(start|stop) shape_inference::ShapeHandle shape; TF_RETURN_IF_ERROR(c->ReplaceDim(full, 0, c->UnknownDim(), &shape)); c->set_output(0, shape); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/pcap_ops.cc b/tensorflow_io/core/ops/pcap_ops.cc index 53538fccc..b794a0f28 100644 --- a/tensorflow_io/core/ops/pcap_ops.cc +++ b/tensorflow_io/core/ops/pcap_ops.cc @@ -26,7 +26,7 @@ REGISTER_OP("IO>PcapReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>PcapReadableRead") @@ -39,7 +39,7 @@ REGISTER_OP("IO>PcapReadableRead") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); } // namespace tensorflow diff --git a/tensorflow_io/core/ops/prometheus_ops.cc b/tensorflow_io/core/ops/prometheus_ops.cc index 83639174a..5fa924172 100644 --- a/tensorflow_io/core/ops/prometheus_ops.cc +++ b/tensorflow_io/core/ops/prometheus_ops.cc @@ -31,7 +31,7 @@ REGISTER_OP("IO>PrometheusReadableInit") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->MakeShape({c->UnknownDim(), 3})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>PrometheusReadableSpec") @@ -41,7 +41,7 @@ REGISTER_OP("IO>PrometheusReadableSpec") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>PrometheusReadableRead") @@ -54,7 +54,7 @@ REGISTER_OP("IO>PrometheusReadableRead") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim(), c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>PrometheusScrape") @@ -66,7 +66,7 @@ REGISTER_OP("IO>PrometheusScrape") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->Scalar()); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/pubsub_ops.cc b/tensorflow_io/core/ops/pubsub_ops.cc index dc4a5f7a9..2c34484ef 100644 --- a/tensorflow_io/core/ops/pubsub_ops.cc +++ b/tensorflow_io/core/ops/pubsub_ops.cc @@ -29,7 +29,7 @@ REGISTER_OP("IO>PubSubReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>PubSubReadableRead") @@ -42,7 +42,7 @@ REGISTER_OP("IO>PubSubReadableRead") c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim()})); c->set_output(2, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/pulsar_ops.cc b/tensorflow_io/core/ops/pulsar_ops.cc index 5ee047d0d..72d540f6b 100644 --- a/tensorflow_io/core/ops/pulsar_ops.cc +++ b/tensorflow_io/core/ops/pulsar_ops.cc @@ -31,7 +31,7 @@ REGISTER_OP("IO>PulsarReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>PulsarReadableNext") @@ -45,7 +45,7 @@ REGISTER_OP("IO>PulsarReadableNext") c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({c->UnknownDim()})); c->set_output(2, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>PulsarWritableInit") @@ -56,7 +56,7 @@ REGISTER_OP("IO>PulsarWritableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>PulsarWritableWrite") diff --git a/tensorflow_io/core/ops/serialization_ops.cc b/tensorflow_io/core/ops/serialization_ops.cc index db45534fe..5a4793e20 100644 --- a/tensorflow_io/core/ops/serialization_ops.cc +++ b/tensorflow_io/core/ops/serialization_ops.cc @@ -33,7 +33,7 @@ REGISTER_OP("IO>DecodeJSON") for (size_t i = 0; i < c->num_outputs(); ++i) { c->set_output(static_cast(i), c->MakeShape({c->UnknownDim()})); } - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>DecodeAvro") @@ -60,7 +60,7 @@ REGISTER_OP("IO>DecodeAvro") c->MakeShapeFromPartialTensorShape(shapes[i], &shape)); c->set_output(static_cast(i), shape); } - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>EncodeAvro") @@ -71,7 +71,7 @@ REGISTER_OP("IO>EncodeAvro") .Attr("dtype: list({bool,int32,int64,float,double,string})") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->input(0)); - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/sql_ops.cc b/tensorflow_io/core/ops/sql_ops.cc index 11093cb5e..5ebe824e4 100644 --- a/tensorflow_io/core/ops/sql_ops.cc +++ b/tensorflow_io/core/ops/sql_ops.cc @@ -36,7 +36,7 @@ REGISTER_OP("IO>SqlIterableInit") c->set_output(1, c->Scalar()); c->set_output(2, c->MakeShape({})); c->set_output(3, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>SqlIterableRead") @@ -50,7 +50,7 @@ REGISTER_OP("IO>SqlIterableRead") for (int64 i = 0; i < c->num_outputs(); ++i) { c->set_output(i, c->MakeShape({c->UnknownDim()})); } - return Status::OK(); + return OkStatus(); }); } // namespace diff --git a/tensorflow_io/core/ops/text_ops.cc b/tensorflow_io/core/ops/text_ops.cc index 1a8447000..dc8fa99bb 100644 --- a/tensorflow_io/core/ops/text_ops.cc +++ b/tensorflow_io/core/ops/text_ops.cc @@ -29,7 +29,7 @@ REGISTER_OP("IO>RE2FullMatch") if (!c->RankKnown(c->input(0))) { c->set_output(0, c->UnknownShape()); c->set_output(1, c->UnknownShape()); - return Status::OK(); + return OkStatus(); } string pattern; TF_RETURN_IF_ERROR(c->GetAttr("pattern", &pattern)); @@ -45,7 +45,7 @@ REGISTER_OP("IO>RE2FullMatch") shape, c->Vector(re.NumberOfCapturingGroups()), &output_shape)); c->set_output(0, c->input(0)); c->set_output(1, output_shape); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>ReadText") @@ -56,7 +56,7 @@ REGISTER_OP("IO>ReadText") .Output("output: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>TextOutputSequence") @@ -83,7 +83,7 @@ REGISTER_OP("IO>CSVReadableInit") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>CSVReadableSpec") @@ -94,7 +94,7 @@ REGISTER_OP("IO>CSVReadableSpec") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); c->set_output(1, c->MakeShape({})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>CSVReadableRead") @@ -112,7 +112,7 @@ REGISTER_OP("IO>CSVReadableRead") shape_inference::ShapeHandle entry; TF_RETURN_IF_ERROR(c->MakeShapeFromPartialTensorShape(shape, &entry)); c->set_output(0, entry); - return Status::OK(); + return OkStatus(); }); } // namespace tensorflow diff --git a/tensorflow_io/core/ops/video_ops.cc b/tensorflow_io/core/ops/video_ops.cc index 01c90a3f7..432b3284c 100644 --- a/tensorflow_io/core/ops/video_ops.cc +++ b/tensorflow_io/core/ops/video_ops.cc @@ -28,7 +28,7 @@ REGISTER_OP("IO>VideoCaptureReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>VideoCaptureReadableRead") @@ -37,7 +37,7 @@ REGISTER_OP("IO>VideoCaptureReadableRead") .Output("value: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>VideoReadableInit") @@ -47,7 +47,7 @@ REGISTER_OP("IO>VideoReadableInit") .Attr("shared_name: string = ''") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->Scalar()); - return Status::OK(); + return OkStatus(); }); REGISTER_OP("IO>VideoReadableRead") @@ -56,7 +56,7 @@ REGISTER_OP("IO>VideoReadableRead") .Output("value: string") .SetShapeFn([](shape_inference::InferenceContext* c) { c->set_output(0, c->MakeShape({c->UnknownDim()})); - return Status::OK(); + return OkStatus(); }); } // namespace From 95aff411b066057c5a5c354eb092430a83b709fe Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sat, 25 Feb 2023 13:25:11 -0800 Subject: [PATCH 04/24] Use tsl for block cache Signed-off-by: Yong Tang --- .../core/kernels/gsmemcachedfs/memcached_file_block_cache.h | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h index b84b14d42..d7a363e42 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h +++ b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h @@ -25,7 +25,7 @@ limitations under the License. #include #include "absl/container/flat_hash_map.h" -#include "tensorflow/core/platform/cloud/file_block_cache.h" +#include "tensorflow/tsl/platform/cloud/ram_file_block_cache.h" #include "tensorflow_io/core/kernels/gsmemcachedfs/memcached_dao_interface.h" namespace tensorflow { From d98ed482ee839911d89938f81b3196544ca98fc7 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sat, 25 Feb 2023 13:42:20 -0800 Subject: [PATCH 05/24] Fix Status::OK Signed-off-by: Yong Tang --- .../core/kernels/arrow/arrow_kernels.h | 2 +- .../kernels/avro/utils/avro_record_reader.h | 2 +- .../core/kernels/avro/utils/value_buffer.h | 22 +++++++++---------- .../core/kernels/bigquery/bigquery_lib.h | 18 +++++++-------- tensorflow_io/core/kernels/io_interface.h | 2 +- tensorflow_io/core/kernels/io_stream.h | 4 ++-- tensorflow_io/core/kernels/sequence_ops.h | 4 ++-- tensorflow_io/core/kernels/video_kernels.h | 4 ++-- 8 files changed, 29 insertions(+), 29 deletions(-) diff --git a/tensorflow_io/core/kernels/arrow/arrow_kernels.h b/tensorflow_io/core/kernels/arrow/arrow_kernels.h index d910c193e..4b0c20c91 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_kernels.h +++ b/tensorflow_io/core/kernels/arrow/arrow_kernels.h @@ -35,7 +35,7 @@ class ArrowRandomAccessFile : public ::arrow::io::RandomAccessFile { : file_(file), size_(size), position_(0) {} ~ArrowRandomAccessFile() {} - arrow::Status Close() override { return arrow::Status::OK(); } + arrow::Status Close() override { return arrow::OkStatus(); } bool closed() const override { return false; } arrow::Result Tell() const override { return position_; } arrow::Status Seek(int64_t position) override { diff --git a/tensorflow_io/core/kernels/avro/utils/avro_record_reader.h b/tensorflow_io/core/kernels/avro/utils/avro_record_reader.h index 3e6f7c00f..df133a714 100644 --- a/tensorflow_io/core/kernels/avro/utils/avro_record_reader.h +++ b/tensorflow_io/core/kernels/avro/utils/avro_record_reader.h @@ -92,7 +92,7 @@ class SequentialAvroRecordReader { "Trying to seek offset: ", offset, " which is less than the current offset: ", offset_); offset_ = offset; - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/avro/utils/value_buffer.h b/tensorflow_io/core/kernels/avro/utils/value_buffer.h index 7727e8946..23791e1d7 100644 --- a/tensorflow_io/core/kernels/avro/utils/value_buffer.h +++ b/tensorflow_io/core/kernels/avro/utils/value_buffer.h @@ -356,7 +356,7 @@ Status ValueBuffer::ResolveDenseShapeWithBatch( } *shape = data_shape; } - return Status::OK(); + return OkStatus(); } // TODO(fraudies): Move validation of user defined shape and defaults into the @@ -381,7 +381,7 @@ Status ValueBuffer::ResolveDenseShape( PartialTensorShape tmp_shape; // Honor any partially defined shape from user and supplement with that from // default - if (user_shape.MergeWith(default_shape, &tmp_shape) == Status::OK()) { + if (user_shape.MergeWith(default_shape, &tmp_shape) == OkStatus()) { // Merged convert partial shape into shape if (!tmp_shape.AsTensorShape(shape)) { return errors::InvalidArgument("Expected ", tmp_shape, @@ -404,7 +404,7 @@ Status ValueBuffer::ResolveDenseShape( PartialTensorShape tmp_shape; // Honor any partially defined shape from user and supplement with that from // data - if (user_shape.MergeWith(dense_shape, &tmp_shape) == Status::OK()) { + if (user_shape.MergeWith(dense_shape, &tmp_shape) == OkStatus()) { if (!tmp_shape.AsTensorShape(shape)) { return errors::InvalidArgument("Expected ", tmp_shape, " to be fully defined" @@ -416,13 +416,13 @@ Status ValueBuffer::ResolveDenseShape( } } - return Status::OK(); + return OkStatus(); } template Status ValueBuffer::GetSparseValueShape(TensorShape* shape) const { (*shape).AddDim(GetNumberOfElements()); - return Status::OK(); + return OkStatus(); } template @@ -433,7 +433,7 @@ Status ValueBuffer::GetSparseIndexShape(TensorShape* shape) const { if (n_dim > 1) { (*shape).AddDim(shape_builder_.GetNumberOfDimensions()); } - return Status::OK(); + return OkStatus(); } // Assumes dense_shape has been allocated appropriate space -- not checked @@ -448,7 +448,7 @@ Status ValueBuffer::GetDenseShapeForSparse(Tensor* dense_shape) const { for (size_t i_dim = 0; i_dim < n_dim; ++i_dim) { tensor_flat(i_dim) = shape.dim_size(i_dim); } - return Status::OK(); + return OkStatus(); } template @@ -467,7 +467,7 @@ Status ValueBuffer::MakeDense(Tensor* tensor, // Fill in the values into the tensor from the buffer TF_RETURN_IF_ERROR(FillInFromBuffer(tensor)); - return Status::OK(); + return OkStatus(); } // Builds the tensor for the values and the indices from the value buffer @@ -484,7 +484,7 @@ Status ValueBuffer::MakeSparse(Tensor* values, Tensor* indices) const { // Create indices TF_RETURN_IF_ERROR(shape_builder_.GetIndices(indices)); - return Status::OK(); + return OkStatus(); } template @@ -511,7 +511,7 @@ Status ValueBuffer::FillInFromBuffer(Tensor* tensor) const { source_offset += length; } - return Status::OK(); + return OkStatus(); } template @@ -544,7 +544,7 @@ Status ValueBuffer::FillInFromDefault(Tensor* tensor, } } - return Status::OK(); + return OkStatus(); } template diff --git a/tensorflow_io/core/kernels/bigquery/bigquery_lib.h b/tensorflow_io/core/kernels/bigquery/bigquery_lib.h index 68a80a3fd..d4b9305d0 100644 --- a/tensorflow_io/core/kernels/bigquery/bigquery_lib.h +++ b/tensorflow_io/core/kernels/bigquery/bigquery_lib.h @@ -123,7 +123,7 @@ class BigQueryReaderDatasetIteratorBase : public DatasetIterator { TF_RETURN_IF_ERROR(EnsureHasRow(end_of_sequence)); if (*end_of_sequence) { VLOG(3) << "end of sequence"; - return Status::OK(); + return OkStatus(); } auto status = @@ -153,7 +153,7 @@ class BigQueryReaderDatasetIteratorBase : public DatasetIterator { } virtual Status EnsureReaderInitialized() TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) { if (reader_) { - return Status::OK(); + return OkStatus(); } apiv1beta1::ReadRowsRequest readRowsRequest; @@ -179,7 +179,7 @@ class BigQueryReaderDatasetIteratorBase : public DatasetIterator { ->GetStub(readRowsRequest.read_position().stream().name()) ->ReadRows(read_rows_context_.get(), readRowsRequest); - return Status::OK(); + return OkStatus(); } virtual Status EnsureHasRow(bool *end_of_sequence) = 0; @@ -214,7 +214,7 @@ class BigQueryReaderArrowDatasetIterator if (this->response_ && this->response_->has_arrow_record_batch() && this->current_row_index_ < this->response_->arrow_record_batch().row_count()) { - return Status::OK(); + return OkStatus(); } this->response_ = absl::make_unique(); @@ -243,7 +243,7 @@ class BigQueryReaderArrowDatasetIterator VLOG(3) << "got record batch, rows:" << record_batch_->num_rows(); - return Status::OK(); + return OkStatus(); } Status ReadRecord(IteratorContext *ctx, std::vector *out_tensors, @@ -292,7 +292,7 @@ class BigQueryReaderArrowDatasetIterator out_tensors->emplace_back(std::move(tensor)); } - return Status::OK(); + return OkStatus(); } private: @@ -316,7 +316,7 @@ class BigQueryReaderAvroDatasetIterator TF_EXCLUSIVE_LOCKS_REQUIRED(this->mu_) override { if (this->response_ && this->current_row_index_ < this->response_->avro_rows().row_count()) { - return Status::OK(); + return OkStatus(); } this->response_ = absl::make_unique(); @@ -335,7 +335,7 @@ class BigQueryReaderAvroDatasetIterator this->decoder_->init(*memory_input_stream_); this->datum_ = absl::make_unique(*this->dataset()->avro_schema()); - return Status::OK(); + return OkStatus(); } Status ReadRecord(IteratorContext *ctx, std::vector *out_tensors, @@ -559,7 +559,7 @@ class BigQueryReaderAvroDatasetIterator } } - return Status::OK(); + return OkStatus(); } private: diff --git a/tensorflow_io/core/kernels/io_interface.h b/tensorflow_io/core/kernels/io_interface.h index 9239bc9b9..c9c81071c 100644 --- a/tensorflow_io/core/kernels/io_interface.h +++ b/tensorflow_io/core/kernels/io_interface.h @@ -121,7 +121,7 @@ class IOInterfaceInitOp : public ResourceOpKernel { Status CreateResource(Type** resource) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *resource = new Type(env_); - return Status::OK(); + return OkStatus(); } mutex mu_; Env* env_; diff --git a/tensorflow_io/core/kernels/io_stream.h b/tensorflow_io/core/kernels/io_stream.h index 54c963159..59b101601 100644 --- a/tensorflow_io/core/kernels/io_stream.h +++ b/tensorflow_io/core/kernels/io_stream.h @@ -32,7 +32,7 @@ class SizedRandomAccessFile : public tensorflow::RandomAccessFile { : file_(nullptr), size_(optional_memory_size), buff_((const char*)(optional_memory_buff)), - size_status_(Status::OK()) { + size_status_(OkStatus()) { if (size_ == 0) { size_status_ = env->GetFileSize(filename, &size_); if (size_status_.ok()) { @@ -58,7 +58,7 @@ class SizedRandomAccessFile : public tensorflow::RandomAccessFile { if (bytes_to_read < n) { return errors::OutOfRange("EOF reached"); } - return Status::OK(); + return OkStatus(); } Status GetFileSize(uint64* size) { if (size_status_.ok()) { diff --git a/tensorflow_io/core/kernels/sequence_ops.h b/tensorflow_io/core/kernels/sequence_ops.h index 79466182c..4ea97d8e9 100644 --- a/tensorflow_io/core/kernels/sequence_ops.h +++ b/tensorflow_io/core/kernels/sequence_ops.h @@ -50,7 +50,7 @@ class OutputSequence : public ResourceBase { if (fifo_.front().get() != nullptr) { TF_RETURN_IF_ERROR(Output()); } - return Status::OK(); + return OkStatus(); } virtual string DebugString() const { return strings::StrCat("OutputSequence[]"); @@ -76,7 +76,7 @@ class OutputSequenceOp : public ResourceOpKernel { Status CreateResource(T** sequence) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) override { *sequence = new T(env_); - return Status::OK(); + return OkStatus(); } Env* env_; mutex mu_; diff --git a/tensorflow_io/core/kernels/video_kernels.h b/tensorflow_io/core/kernels/video_kernels.h index 91838dbe5..e4ebb3c49 100644 --- a/tensorflow_io/core/kernels/video_kernels.h +++ b/tensorflow_io/core/kernels/video_kernels.h @@ -124,7 +124,7 @@ class VideoCaptureContext { *width = fmt.fmt.pix.width; *height = fmt.fmt.pix.height; - return Status::OK(); + return OkStatus(); } Status Read(void* data, size_t size) { do { @@ -166,7 +166,7 @@ class VideoCaptureContext { // Data Obtained, break break; } while (true); - return Status::OK(); + return OkStatus(); } protected: From fc4cd0b954b6c5cc47a6bc5eba145e4ef6c400ff Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sat, 25 Feb 2023 14:12:32 -0800 Subject: [PATCH 06/24] Fix Signed-off-by: Yong Tang --- .../core/kernels/gsmemcachedfs/memcached_file_system.h | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.h b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.h index 7b37b871b..0ead14242 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.h +++ b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.h @@ -21,7 +21,7 @@ limitations under the License. #include #include "absl/memory/memory.h" -#include "tensorflow/core/platform/cloud/gcs_file_system.h" +#include "tensorflow/tsl/platform/cloud/gcs_file_system.h" #include "tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.h" #include "tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h" From 1c2ee76541769a0f410799d465ce7ea8ee0c906a Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sat, 25 Feb 2023 14:40:58 -0800 Subject: [PATCH 07/24] Fix Signed-off-by: Yong Tang --- tensorflow_io/core/kernels/arrow/arrow_kernels.h | 2 +- .../kernels/arrow/arrow_stream_client_unix.cc | 4 ++-- .../kernels/arrow/arrow_stream_client_windows.cc | 4 ++-- tensorflow_io/core/kernels/arrow/arrow_util.cc | 16 ++++++++-------- 4 files changed, 13 insertions(+), 13 deletions(-) diff --git a/tensorflow_io/core/kernels/arrow/arrow_kernels.h b/tensorflow_io/core/kernels/arrow/arrow_kernels.h index 4b0c20c91..d910c193e 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_kernels.h +++ b/tensorflow_io/core/kernels/arrow/arrow_kernels.h @@ -35,7 +35,7 @@ class ArrowRandomAccessFile : public ::arrow::io::RandomAccessFile { : file_(file), size_(size), position_(0) {} ~ArrowRandomAccessFile() {} - arrow::Status Close() override { return arrow::OkStatus(); } + arrow::Status Close() override { return arrow::Status::OK(); } bool closed() const override { return false; } arrow::Result Tell() const override { return position_; } arrow::Status Seek(int64_t position) override { diff --git a/tensorflow_io/core/kernels/arrow/arrow_stream_client_unix.cc b/tensorflow_io/core/kernels/arrow/arrow_stream_client_unix.cc index 78cd85465..bc171a6cb 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_stream_client_unix.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_stream_client_unix.cc @@ -95,7 +95,7 @@ arrow::Status ArrowStreamClient::Connect() { socket_family); } - return arrow::OkStatus(); + return arrow::Status::OK(); } arrow::Status ArrowStreamClient::Close() { @@ -106,7 +106,7 @@ arrow::Status ArrowStreamClient::Close() { return arrow::Status::IOError("Failed to correctly close connection"); } - return arrow::OkStatus(); + return arrow::Status::OK(); } bool ArrowStreamClient::closed() const { return sock_ == -1; } diff --git a/tensorflow_io/core/kernels/arrow/arrow_stream_client_windows.cc b/tensorflow_io/core/kernels/arrow/arrow_stream_client_windows.cc index 6af6c91b6..64c465205 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_stream_client_windows.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_stream_client_windows.cc @@ -115,7 +115,7 @@ arrow::Status ArrowStreamClient::Connect() { return arrow::Status::IOError("Unable to connect to server"); } - return arrow::OkStatus(); + return arrow::Status::OK(); } arrow::Status ArrowStreamClient::Close() { @@ -128,7 +128,7 @@ arrow::Status ArrowStreamClient::Close() { std::to_string(WSAGetLastError())); } - return arrow::OkStatus(); + return arrow::Status::OK(); } bool ArrowStreamClient::closed() const { return sock_ == -1; } diff --git a/tensorflow_io/core/kernels/arrow/arrow_util.cc b/tensorflow_io/core/kernels/arrow/arrow_util.cc index ba24f2a73..931d8f796 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_util.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_util.cc @@ -101,7 +101,7 @@ class ArrowAssignSpecImpl : public arrow::ArrayVisitor { return ::arrow::adapters::tensorflow::GetTensorFlowType(array.type(), out_dtype_); } - return arrow::OkStatus(); + return arrow::Status::OK(); } #define VISIT_PRIMITIVE(TYPE) \ @@ -202,7 +202,7 @@ class ArrowAssignTensorImpl : public arrow::ArrayVisitor { memcpy(dst, &value, sizeof(value)); } - return arrow::OkStatus(); + return arrow::Status::OK(); } template @@ -227,7 +227,7 @@ class ArrowAssignTensorImpl : public arrow::ArrayVisitor { void* dst = const_cast(out_tensor_->tensor_data().data()); std::memcpy(dst, src, out_tensor_->NumElements() * type_width); - return arrow::OkStatus(); + return arrow::Status::OK(); } #define VISIT_FIXED_WIDTH(TYPE) \ @@ -288,7 +288,7 @@ class ArrowAssignTensorImpl : public arrow::ArrayVisitor { output_flat(j) = array.GetString(i_ + j); } - return arrow::OkStatus(); + return arrow::Status::OK(); } virtual arrow::Status Visit(const arrow::BinaryArray& array) override { @@ -299,7 +299,7 @@ class ArrowAssignTensorImpl : public arrow::ArrayVisitor { output_flat(j) = array.GetString(i_ + j); } - return arrow::OkStatus(); + return arrow::Status::OK(); } private: @@ -351,7 +351,7 @@ class ArrowArrayTypeCheckerImpl : public arrow::TypeVisitor { std::to_string(expected_type_) + ", but got dtype=" + std::to_string(converted_type)); } - return arrow::OkStatus(); + return arrow::Status::OK(); } private: @@ -384,7 +384,7 @@ class ArrowMakeArrayDataImpl : public arrow::TypeVisitor { // TODO null count == 0 *out_data_ = arrow::ArrayData::Make(type_, lengths_[0], std::move(buffers_), 0); - return arrow::OkStatus(); + return arrow::Status::OK(); } #define VISIT_PRIMITIVE(TYPE) \ @@ -428,7 +428,7 @@ class ArrowMakeArrayDataImpl : public arrow::TypeVisitor { *out_data_ = arrow::ArrayData::Make(list_type, list_length, std::move(list_bufs), {child_data}, 0); - return arrow::OkStatus(); + return arrow::Status::OK(); } private: From fddc8c7abc6e4d862fde6217e970730f30cf40ad Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sat, 25 Feb 2023 15:07:09 -0800 Subject: [PATCH 08/24] ValueOrDie Signed-off-by: Yong Tang --- tensorflow_io/core/kernels/arrow/arrow_kernels.cc | 2 +- .../core/kernels/bigtable/bigtable_resource_kernel.h | 2 +- tensorflow_io/core/kernels/bigtable/serialization.cc | 10 +++++----- tensorflow_io/core/kernels/genome_fastq_kernels.cc | 8 ++++---- tensorflow_io/core/kernels/json_kernels.cc | 4 ++-- 5 files changed, 13 insertions(+), 13 deletions(-) diff --git a/tensorflow_io/core/kernels/arrow/arrow_kernels.cc b/tensorflow_io/core/kernels/arrow/arrow_kernels.cc index 1ed8f3bc5..55c4948db 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_kernels.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_kernels.cc @@ -551,7 +551,7 @@ class FeatherReadable : public IOReadableInterface { return errors::Internal(maybe_reader.status().ToString()); } std::shared_ptr reader = - maybe_reader.ValueOrDie(); + maybe_reader.value(); std::shared_ptr schema = reader->schema(); std::shared_ptr table; diff --git a/tensorflow_io/core/kernels/bigtable/bigtable_resource_kernel.h b/tensorflow_io/core/kernels/bigtable/bigtable_resource_kernel.h index 030596808..991f59df1 100644 --- a/tensorflow_io/core/kernels/bigtable/bigtable_resource_kernel.h +++ b/tensorflow_io/core/kernels/bigtable/bigtable_resource_kernel.h @@ -44,7 +44,7 @@ class AbstractBigtableResourceOp : public OpKernel { OP_REQUIRES_OK(context, maybe_resource.status()); OP_REQUIRES_OK(context, mgr->Create(cinfo.container(), cinfo.name(), - maybe_resource.ValueOrDie())); + maybe_resource.value())); OP_REQUIRES_OK(context, MakeResourceHandleToOutput( context, 0, cinfo.container(), cinfo.name(), diff --git a/tensorflow_io/core/kernels/bigtable/serialization.cc b/tensorflow_io/core/kernels/bigtable/serialization.cc index 3beb9e81f..18f24f429 100644 --- a/tensorflow_io/core/kernels/bigtable/serialization.cc +++ b/tensorflow_io/core/kernels/bigtable/serialization.cc @@ -150,7 +150,7 @@ Status PutCellValueInTensor(Tensor& tensor, size_t index, DataType cell_type, if (!maybe_parsed_data.ok()) { return maybe_parsed_data.status(); } - tensor_data(index) = maybe_parsed_data.ValueOrDie(); + tensor_data(index) = maybe_parsed_data.value(); } break; case DT_INT32: { auto tensor_data = tensor.tensor(); @@ -158,7 +158,7 @@ Status PutCellValueInTensor(Tensor& tensor, size_t index, DataType cell_type, if (!maybe_parsed_data.ok()) { return maybe_parsed_data.status(); } - tensor_data(index) = maybe_parsed_data.ValueOrDie(); + tensor_data(index) = maybe_parsed_data.value(); } break; case DT_INT64: { auto tensor_data = tensor.tensor(); @@ -166,7 +166,7 @@ Status PutCellValueInTensor(Tensor& tensor, size_t index, DataType cell_type, if (!maybe_parsed_data.ok()) { return maybe_parsed_data.status(); } - tensor_data(index) = maybe_parsed_data.ValueOrDie(); + tensor_data(index) = maybe_parsed_data.value(); } break; case DT_FLOAT: { auto tensor_data = tensor.tensor(); @@ -174,7 +174,7 @@ Status PutCellValueInTensor(Tensor& tensor, size_t index, DataType cell_type, if (!maybe_parsed_data.ok()) { return maybe_parsed_data.status(); } - tensor_data(index) = maybe_parsed_data.ValueOrDie(); + tensor_data(index) = maybe_parsed_data.value(); } break; case DT_DOUBLE: { auto tensor_data = tensor.tensor(); @@ -182,7 +182,7 @@ Status PutCellValueInTensor(Tensor& tensor, size_t index, DataType cell_type, if (!maybe_parsed_data.ok()) { return maybe_parsed_data.status(); } - tensor_data(index) = maybe_parsed_data.ValueOrDie(); + tensor_data(index) = maybe_parsed_data.value(); } break; default: return errors::Unimplemented("Data type not supported."); diff --git a/tensorflow_io/core/kernels/genome_fastq_kernels.cc b/tensorflow_io/core/kernels/genome_fastq_kernels.cc index 530798287..7124d5265 100644 --- a/tensorflow_io/core/kernels/genome_fastq_kernels.cc +++ b/tensorflow_io/core/kernels/genome_fastq_kernels.cc @@ -44,20 +44,20 @@ class FastqOp : public OpKernel { std::unique_ptr reader = std::move(FastqReader::FromFile( filename, nucleus::genomics::v1::FastqReaderOptions()) - .ValueOrDie()); + .value()); std::vector sequences; std::vector quality; std::shared_ptr fastq_iterable = - reader->Iterate().ValueOrDie(); + reader->Iterate().value(); for (const nucleus::StatusOr maybe_sequence : fastq_iterable) { OP_REQUIRES( context, maybe_sequence.ok(), errors::Internal("internal error: ", maybe_sequence.error_message())); - sequences.push_back(maybe_sequence.ValueOrDie()->sequence()); - quality.push_back(maybe_sequence.ValueOrDie()->quality()); + sequences.push_back(maybe_sequence.value()->sequence()); + quality.push_back(maybe_sequence.value()->quality()); } TensorShape output_shape({static_cast(sequences.size())}); diff --git a/tensorflow_io/core/kernels/json_kernels.cc b/tensorflow_io/core/kernels/json_kernels.cc index 7af956970..b053478d0 100644 --- a/tensorflow_io/core/kernels/json_kernels.cc +++ b/tensorflow_io/core/kernels/json_kernels.cc @@ -114,13 +114,13 @@ class JSONReadable : public IOReadableInterface { if (!reader_result.ok()) { return errors::InvalidArgument("unable to make a TableReader: ", status); } else { - reader_ = reader_result.ValueOrDie(); + reader_ = reader_result.value(); } auto table_result = reader_->Read(); if (!table_result.ok()) { return errors::InvalidArgument("unable to read table: ", status); } else { - table_ = table_result.ValueOrDie(); + table_ = table_result.value(); } shapes_.clear(); From 7e39fddf39cd4f2134c1d14800effb5ffce999b7 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sat, 25 Feb 2023 15:17:38 -0800 Subject: [PATCH 09/24] Fix again Signed-off-by: Yong Tang --- tensorflow_io/core/kernels/elasticsearch_kernels.cc | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tensorflow_io/core/kernels/elasticsearch_kernels.cc b/tensorflow_io/core/kernels/elasticsearch_kernels.cc index 6862a9114..c092511ee 100644 --- a/tensorflow_io/core/kernels/elasticsearch_kernels.cc +++ b/tensorflow_io/core/kernels/elasticsearch_kernels.cc @@ -19,7 +19,7 @@ limitations under the License. #include "rapidjson/writer.h" #include "tensorflow/core/framework/resource_mgr.h" #include "tensorflow/core/framework/resource_op_kernel.h" -#include "tensorflow/core/platform/cloud/curl_http_request.h" +#include "tensorflow/tsl/platform/cloud/curl_http_request.h" namespace tensorflow { namespace io { From 4da4a04c3f83b0b3fd1d20a6818114bc58ffc7b1 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sat, 25 Feb 2023 19:01:16 -0800 Subject: [PATCH 10/24] Fix Signed-off-by: Yong Tang --- tensorflow_io/core/kernels/arrow/arrow_kernels.cc | 2 +- tensorflow_io/core/kernels/genome_fastq_kernels.cc | 8 ++++---- tensorflow_io/core/kernels/json_kernels.cc | 4 ++-- 3 files changed, 7 insertions(+), 7 deletions(-) diff --git a/tensorflow_io/core/kernels/arrow/arrow_kernels.cc b/tensorflow_io/core/kernels/arrow/arrow_kernels.cc index 55c4948db..1ed8f3bc5 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_kernels.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_kernels.cc @@ -551,7 +551,7 @@ class FeatherReadable : public IOReadableInterface { return errors::Internal(maybe_reader.status().ToString()); } std::shared_ptr reader = - maybe_reader.value(); + maybe_reader.ValueOrDie(); std::shared_ptr schema = reader->schema(); std::shared_ptr table; diff --git a/tensorflow_io/core/kernels/genome_fastq_kernels.cc b/tensorflow_io/core/kernels/genome_fastq_kernels.cc index 7124d5265..530798287 100644 --- a/tensorflow_io/core/kernels/genome_fastq_kernels.cc +++ b/tensorflow_io/core/kernels/genome_fastq_kernels.cc @@ -44,20 +44,20 @@ class FastqOp : public OpKernel { std::unique_ptr reader = std::move(FastqReader::FromFile( filename, nucleus::genomics::v1::FastqReaderOptions()) - .value()); + .ValueOrDie()); std::vector sequences; std::vector quality; std::shared_ptr fastq_iterable = - reader->Iterate().value(); + reader->Iterate().ValueOrDie(); for (const nucleus::StatusOr maybe_sequence : fastq_iterable) { OP_REQUIRES( context, maybe_sequence.ok(), errors::Internal("internal error: ", maybe_sequence.error_message())); - sequences.push_back(maybe_sequence.value()->sequence()); - quality.push_back(maybe_sequence.value()->quality()); + sequences.push_back(maybe_sequence.ValueOrDie()->sequence()); + quality.push_back(maybe_sequence.ValueOrDie()->quality()); } TensorShape output_shape({static_cast(sequences.size())}); diff --git a/tensorflow_io/core/kernels/json_kernels.cc b/tensorflow_io/core/kernels/json_kernels.cc index b053478d0..7af956970 100644 --- a/tensorflow_io/core/kernels/json_kernels.cc +++ b/tensorflow_io/core/kernels/json_kernels.cc @@ -114,13 +114,13 @@ class JSONReadable : public IOReadableInterface { if (!reader_result.ok()) { return errors::InvalidArgument("unable to make a TableReader: ", status); } else { - reader_ = reader_result.value(); + reader_ = reader_result.ValueOrDie(); } auto table_result = reader_->Read(); if (!table_result.ok()) { return errors::InvalidArgument("unable to read table: ", status); } else { - table_ = table_result.value(); + table_ = table_result.ValueOrDie(); } shapes_.clear(); From 69eff7b1aecff73fa634b3d562afb27d2b3eb32a Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sun, 26 Feb 2023 04:19:32 +0000 Subject: [PATCH 11/24] Additional nucleus patch Signed-off-by: Yong Tang --- third_party/nucleus.patch | 91 +++++++++++++++++++++++++++++++++++++-- 1 file changed, 87 insertions(+), 4 deletions(-) diff --git a/third_party/nucleus.patch b/third_party/nucleus.patch index 6b8034f92..148c0e49a 100644 --- a/third_party/nucleus.patch +++ b/third_party/nucleus.patch @@ -1,6 +1,89 @@ +diff -Naur a/nucleus/io/fastq_reader.cc b/nucleus/io/fastq_reader.cc +--- a/nucleus/io/fastq_reader.cc 2023-02-26 04:04:19.097785713 +0000 ++++ b/nucleus/io/fastq_reader.cc 2023-02-26 04:15:52.293582517 +0000 +@@ -76,7 +76,7 @@ + } + record->set_sequence(string(sequence)); + record->set_quality(string(quality)); +- return tf::Status::OK(); ++ return tsl::OkStatus(); + } + } // namespace + +@@ -145,7 +145,7 @@ + *sequence = sequence_or.ValueOrDie(); + *pad = pad_or.ValueOrDie(); + *quality = quality_or.ValueOrDie(); +- return tf::Status::OK(); ++ return tsl::OkStatus(); + + data_loss: + return tf::errors::DataLoss("Failed to parse FASTQ record"); +diff -Naur a/nucleus/io/fastq_reader.h b/nucleus/io/fastq_reader.h +--- a/nucleus/io/fastq_reader.h 2023-02-26 04:04:19.097785713 +0000 ++++ b/nucleus/io/fastq_reader.h 2023-02-26 04:16:46.494817526 +0000 +@@ -76,7 +76,7 @@ + + // This no-op function is needed only for Python context manager support. Do + // not use it! Returns a Status indicating whether the enter was successful. +- tensorflow::Status PythonEnter() const { return tensorflow::Status::OK(); } ++ tensorflow::Status PythonEnter() const { return tsl::OkStatus(); } + + // Get the options controlling the behavior of this FastqReader. + const nucleus::genomics::v1::FastqReaderOptions& Options() const { +diff -Naur a/nucleus/io/reader_base.cc b/nucleus/io/reader_base.cc +--- a/nucleus/io/reader_base.cc 2023-02-26 04:04:19.097785713 +0000 ++++ b/nucleus/io/reader_base.cc 2023-02-26 04:17:30.943830589 +0000 +@@ -56,7 +56,7 @@ + reader_->live_iterable_ = nullptr; + reader_ = nullptr; + } +- return tensorflow::Status::OK(); ++ return tsl::OkStatus(); + } + + bool IterableBase::IsAlive() const { +@@ -66,11 +66,11 @@ + tensorflow::Status IterableBase::CheckIsAlive() const { + if (!IsAlive()) + return tensorflow::errors::FailedPrecondition("Reader is not alive"); +- return tensorflow::Status::OK(); ++ return tsl::OkStatus(); + } + + tensorflow::Status IterableBase::PythonEnter() { +- return tensorflow::Status::OK(); ++ return tsl::OkStatus(); + } + + tensorflow::Status IterableBase::PythonExit() { +diff -Naur a/nucleus/io/reader_base.h b/nucleus/io/reader_base.h +--- a/nucleus/io/reader_base.h 2023-02-26 04:04:19.097785713 +0000 ++++ b/nucleus/io/reader_base.h 2023-02-26 04:14:44.948047970 +0000 +@@ -129,7 +129,7 @@ + + private: + Record current_record_; +- tensorflow::Status current_status_ = tensorflow::Status::OK(); ++ tensorflow::Status current_status_ = tsl::OkStatus(); + bool IsOK() { return current_status_.ok(); } + + public: +diff -Naur a/nucleus/io/text_reader.cc b/nucleus/io/text_reader.cc +--- a/nucleus/io/text_reader.cc 2023-02-26 04:04:19.101785802 +0000 ++++ b/nucleus/io/text_reader.cc 2023-02-26 04:18:23.589030554 +0000 +@@ -78,7 +78,7 @@ + if (hts_ok < 0) { + return tf::errors::Internal("hts_close() failed with return code ", hts_ok); + } +- return tf::Status::OK(); ++ return tsl::OkStatus(); + } + + TextReader::TextReader(htsFile* hts_file) diff -Naur a/nucleus/protos/BUILD b/nucleus/protos/BUILD ---- a/nucleus/protos/BUILD 2019-09-07 15:31:09.625067415 +0000 -+++ b/nucleus/protos/BUILD 2019-09-07 15:32:35.368864263 +0000 +--- a/nucleus/protos/BUILD 2023-02-26 04:04:19.101785802 +0000 ++++ b/nucleus/protos/BUILD 2023-02-26 04:08:24.867390853 +0000 @@ -1,62 +1,6 @@ -load("//:third_party/clif.bzl", "py_clif_cc", "pyclif_proto_library") -load("@protobuf_archive//:protobuf.bzl", "py_proto_library") @@ -467,8 +550,8 @@ diff -Naur a/nucleus/protos/BUILD b/nucleus/protos/BUILD + deps = [":fastq_proto"], ) diff -Naur a/nucleus/vendor/BUILD b/nucleus/vendor/BUILD ---- a/nucleus/vendor/BUILD 2019-09-07 15:31:09.629067403 +0000 -+++ b/nucleus/vendor/BUILD 2019-09-07 15:38:30.184079565 +0000 +--- a/nucleus/vendor/BUILD 2023-02-26 04:04:19.105785893 +0000 ++++ b/nucleus/vendor/BUILD 2023-02-26 04:08:24.867390853 +0000 @@ -1,4 +1,4 @@ -load("//:third_party/clif.bzl", "py_clif_cc") +#load("//:third_party/clif.bzl", "py_clif_cc") From f5c40ec219ea4c91e70571f1f022fdc1cc3b80df Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sun, 26 Feb 2023 04:40:01 +0000 Subject: [PATCH 12/24] Fix Signed-off-by: Yong Tang --- tensorflow_io/core/kernels/elasticsearch_kernels.cc | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tensorflow_io/core/kernels/elasticsearch_kernels.cc b/tensorflow_io/core/kernels/elasticsearch_kernels.cc index c092511ee..1b8b4c2d3 100644 --- a/tensorflow_io/core/kernels/elasticsearch_kernels.cc +++ b/tensorflow_io/core/kernels/elasticsearch_kernels.cc @@ -254,7 +254,7 @@ class ElasticsearchReadableResource : public ResourceBase { mutable mutex mu_; Env* env_ TF_GUARDED_BY(mu_); string url_; - CurlHttpRequest::Factory http_request_factory_ = CurlHttpRequest::Factory(); + tsl::CurlHttpRequest::Factory http_request_factory_ = tsl::CurlHttpRequest::Factory(); std::vector base_dtypes_; std::vector base_columns_; From 586dd7c9d5f280e7781f8eee5c8f9fd512371216 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sun, 26 Feb 2023 08:14:02 +0000 Subject: [PATCH 13/24] Fix Signed-off-by: Yong Tang --- tensorflow_io/core/filesystems/BUILD | 1 + tensorflow_io/core/kernels/elasticsearch_kernels.cc | 2 ++ .../gsmemcachedfs/gce_memcached_server_list_provider.cc | 2 ++ .../gsmemcachedfs/gce_memcached_server_list_provider.h | 8 ++++---- .../kernels/gsmemcachedfs/gs_memcached_file_system.h | 7 ++++--- .../kernels/gsmemcachedfs/memcached_file_block_cache.cc | 8 ++++---- .../kernels/gsmemcachedfs/memcached_file_block_cache.h | 4 +++- .../core/kernels/gsmemcachedfs/memcached_file_system.cc | 4 +++- .../core/kernels/gsmemcachedfs/memcached_file_system.h | 4 ++-- third_party/toolchains/tf/BUILD.tpl | 9 +++++++++ third_party/toolchains/tf/tf_configure.bzl | 7 +++++++ 11 files changed, 41 insertions(+), 15 deletions(-) diff --git a/tensorflow_io/core/filesystems/BUILD b/tensorflow_io/core/filesystems/BUILD index 9276208fd..cd5fa018b 100644 --- a/tensorflow_io/core/filesystems/BUILD +++ b/tensorflow_io/core/filesystems/BUILD @@ -23,6 +23,7 @@ cc_library( linkstatic = True, deps = [ "@local_config_tf//:tf_c_header_lib", + "@local_config_tf//:tf_tsl_header_lib", ], alwayslink = 1, ) diff --git a/tensorflow_io/core/kernels/elasticsearch_kernels.cc b/tensorflow_io/core/kernels/elasticsearch_kernels.cc index 1b8b4c2d3..214dbef37 100644 --- a/tensorflow_io/core/kernels/elasticsearch_kernels.cc +++ b/tensorflow_io/core/kernels/elasticsearch_kernels.cc @@ -21,6 +21,8 @@ limitations under the License. #include "tensorflow/core/framework/resource_op_kernel.h" #include "tensorflow/tsl/platform/cloud/curl_http_request.h" +using namespace tsl; + namespace tensorflow { namespace io { namespace { diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.cc b/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.cc index af1fc28ec..54fe68317 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.cc +++ b/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.cc @@ -1,5 +1,7 @@ #include "tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.h" +using namespace tsl; + namespace tensorflow { namespace { diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.h b/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.h index ea7624dc6..43318c970 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.h +++ b/tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.h @@ -7,23 +7,23 @@ #include "absl/memory/memory.h" #include "tensorflow/core/lib/core/status.h" -#include "tensorflow/core/platform/cloud/compute_engine_metadata_client.h" +#include "tensorflow/tsl/platform/cloud/compute_engine_metadata_client.h" namespace tensorflow { class GceMemcachedServerListProvider { public: explicit GceMemcachedServerListProvider( - std::shared_ptr metadata_client); + std::shared_ptr metadata_client); virtual ~GceMemcachedServerListProvider(); Status GetServerList(std::vector* server_list); void SetMetadataClient( - std::shared_ptr metadata_client); + std::shared_ptr metadata_client); private: - std::shared_ptr google_metadata_client_; + std::shared_ptr google_metadata_client_; std::vector cached_list_; TF_DISALLOW_COPY_AND_ASSIGN(GceMemcachedServerListProvider); }; diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/gs_memcached_file_system.h b/tensorflow_io/core/kernels/gsmemcachedfs/gs_memcached_file_system.h index ae368d92b..5fc772b39 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/gs_memcached_file_system.h +++ b/tensorflow_io/core/kernels/gsmemcachedfs/gs_memcached_file_system.h @@ -23,6 +23,7 @@ limitations under the License. #include "absl/memory/memory.h" #include "tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.h" #include "tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.h" +#include "tensorflow/tsl/platform/retrying_file_system.h" namespace tensorflow { @@ -46,11 +47,11 @@ class GsMemcachedFileSystem : public MemcachedGcsFileSystem { /// Google Cloud Storage implementation of a file system with retry on failures. class RetryingGsMemcachedFileSystem - : public RetryingFileSystem { + : public tsl::RetryingFileSystem { public: RetryingGsMemcachedFileSystem() - : RetryingFileSystem(absl::make_unique(), - RetryConfig(100000 /* init_delay_time_us */)) {} + : tsl::RetryingFileSystem(absl::make_unique(), + tsl::RetryConfig(100000 /* init_delay_time_us */)) {} }; } // namespace tensorflow diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.cc b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.cc index 43a9a651d..b735113e6 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.cc +++ b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.cc @@ -86,14 +86,14 @@ int64 ComputeBackoffMicroseconds(int current_retry_attempt, int64 min_delay, namespace { inline void StreamzRecordCacheHitBlockSize( - size_t value, FileBlockCacheStatsInterface* cache_stats) { + size_t value, tsl::FileBlockCacheStatsInterface* cache_stats) { if (cache_stats != nullptr) { cache_stats->RecordCacheHitBlockSize(value); } } inline void StreamzRecordCacheMissBlockSize( - size_t value, FileBlockCacheStatsInterface* cache_stats) { + size_t value, tsl::FileBlockCacheStatsInterface* cache_stats) { if (cache_stats != nullptr) { cache_stats->RecordCacheMissBlockSize(value); } @@ -197,7 +197,7 @@ bool BufferCollator::splice_buffer(T begin, T end, size_t pos, Status block_get(MemcachedDaoInterface* memcached_dao, const string& key, std::vector* value, - FileBlockCacheStatsInterface* cache_stats) { + tsl::FileBlockCacheStatsInterface* cache_stats) { memcached_return rc; size_t value_length; char* retrieved_value; @@ -269,7 +269,7 @@ Status read_with_multi_get( const std::vector& keys, std::map* claim_checks, size_t* total_bytes_transferred, - FileBlockCacheStatsInterface* cache_stats) { + tsl::FileBlockCacheStatsInterface* cache_stats) { VLOG(2) << "Key multi-get of " << claim_checks->size() << " claims"; const auto before = absl::Now(); auto to_claim = claim_checks->size(); diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h index d7a363e42..2efcfa2d2 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h +++ b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h @@ -26,6 +26,8 @@ limitations under the License. #include "absl/container/flat_hash_map.h" #include "tensorflow/tsl/platform/cloud/ram_file_block_cache.h" +#include "tensorflow/core/framework/types.h" +#include "tensorflow/core/platform/env.h" #include "tensorflow_io/core/kernels/gsmemcachedfs/memcached_dao_interface.h" namespace tensorflow { @@ -232,7 +234,7 @@ class MemcachedDao : public MemcachedDaoInterface { // // This class should be shared by read-only random access files on a remote // filesystem (e.g. GCS). -class MemcachedFileBlockCache : public FileBlockCache { +class MemcachedFileBlockCache : public tsl::FileBlockCache { public: MemcachedFileBlockCache( const std::vector& memcached_daos, diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.cc b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.cc index 954d1a76a..643c28df5 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.cc +++ b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.cc @@ -27,7 +27,7 @@ limitations under the License. #ifdef _WIN32 #include // for _mktemp #endif -#include "tensorflow/core/platform/cloud/ram_file_block_cache.h" +#include "tensorflow/tsl/platform/cloud/ram_file_block_cache.h" #ifdef _WIN32 #ifdef DeleteFile @@ -35,6 +35,8 @@ limitations under the License. #endif #endif +using namespace tsl; + namespace tensorflow { namespace { diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.h b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.h index 0ead14242..fe43b5452 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.h +++ b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.h @@ -38,13 +38,13 @@ constexpr size_t kDefaultMemcachedClientPoolSize = 64; // // The clients should use MemcachedGcsFileSystem defined below, // which adds retry logic to GCS operations. -class MemcachedGcsFileSystem : public GcsFileSystem { +class MemcachedGcsFileSystem : public tsl::GcsFileSystem { public: // Main constructor used (via RetryingFileSystem) throughout Tensorflow. MemcachedGcsFileSystem(); protected: - std::unique_ptr MakeFileBlockCache( + std::unique_ptr MakeFileBlockCache( size_t block_size, size_t max_bytes, uint64 max_staleness) override; // If the distributed cache is not specified for use in the env variables diff --git a/third_party/toolchains/tf/BUILD.tpl b/third_party/toolchains/tf/BUILD.tpl index d90c2886a..75e5499a8 100644 --- a/third_party/toolchains/tf/BUILD.tpl +++ b/third_party/toolchains/tf/BUILD.tpl @@ -74,6 +74,14 @@ cc_library( visibility = ["//visibility:public"], ) +cc_library( + name = "tf_tsl_header_lib", + hdrs = [":tf_tsl_header_include"], + include_prefix = "tensorflow/tsl/c", + strip_include_prefix = "include_tsl", + visibility = ["//visibility:public"], +) + cc_library( name = "libtensorflow_framework", srcs = select({ @@ -90,4 +98,5 @@ cc_library( %{TF_HEADER_GENRULE} %{TF_C_HEADER_GENRULE} +%{TF_TSL_HEADER_GENRULE} %{TF_SHARED_LIBRARY_GENRULE} diff --git a/third_party/toolchains/tf/tf_configure.bzl b/third_party/toolchains/tf/tf_configure.bzl index 3565f078b..383908c44 100644 --- a/third_party/toolchains/tf/tf_configure.bzl +++ b/third_party/toolchains/tf/tf_configure.bzl @@ -182,6 +182,12 @@ def _tf_pip_impl(repository_ctx): "include_c", "tf_c_header_include", ) + tf_tsl_header_rule = _symlink_genrule_for_dir( + repository_ctx, + tf_header_dir + "/tensorflow/tsl/c/", + "include_tsl", + "tf_tsl_header_include", + ) tf_shared_library_dir = repository_ctx.os.environ[_TF_SHARED_LIBRARY_DIR] tf_shared_library_name = repository_ctx.os.environ[_TF_SHARED_LIBRARY_NAME] @@ -199,6 +205,7 @@ def _tf_pip_impl(repository_ctx): _tpl(repository_ctx, "BUILD", { "%{TF_HEADER_GENRULE}": tf_header_rule, "%{TF_C_HEADER_GENRULE}": tf_c_header_rule, + "%{TF_TSL_HEADER_GENRULE}": tf_tsl_header_rule, "%{TF_SHARED_LIBRARY_GENRULE}": tf_shared_library_rule, }) From d7c5f4e96db00bfca01acc316beba4da5524dee5 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sun, 26 Feb 2023 08:22:58 +0000 Subject: [PATCH 14/24] Fix Signed-off-by: Yong Tang --- .github/workflows/build.yml | 3 +++ tensorflow_io_gcs_filesystem/core/BUILD | 1 + 2 files changed, 4 insertions(+) diff --git a/.github/workflows/build.yml b/.github/workflows/build.yml index c2ad22240..20bbfb864 100644 --- a/.github/workflows/build.yml +++ b/.github/workflows/build.yml @@ -330,6 +330,9 @@ jobs: cat > service_account_creds.json << EOF ${{ secrets.GCP_CREDS }} EOF + - uses: actions/setup-python@v1 + with: + python-version: "3.8" - name: Bazel on Windows env: BAZEL_VC: "C:/Program Files (x86)/Microsoft Visual Studio/2019/Enterprise/VC/" diff --git a/tensorflow_io_gcs_filesystem/core/BUILD b/tensorflow_io_gcs_filesystem/core/BUILD index c1b0a1b6e..5d20c8129 100644 --- a/tensorflow_io_gcs_filesystem/core/BUILD +++ b/tensorflow_io_gcs_filesystem/core/BUILD @@ -36,6 +36,7 @@ cc_library( "@com_google_absl//absl/synchronization", "@com_google_absl//absl/types:variant", "@local_config_tf//:tf_c_header_lib", + "@local_config_tf//:tf_tsl_header_lib", ], alwayslink = 1, ) From 5c9d0d476ccd4aa0f477f0efa35509510039f0b2 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sun, 26 Feb 2023 18:01:38 +0000 Subject: [PATCH 15/24] Windows-2022 Signed-off-by: Yong Tang --- .github/workflows/build.yml | 5 +---- 1 file changed, 1 insertion(+), 4 deletions(-) diff --git a/.github/workflows/build.yml b/.github/workflows/build.yml index 20bbfb864..79c4ba354 100644 --- a/.github/workflows/build.yml +++ b/.github/workflows/build.yml @@ -318,7 +318,7 @@ jobs: windows-bazel: name: Bazel Windows - runs-on: windows-2019 + runs-on: windows-2022 steps: - uses: actions/checkout@v2 - uses: egor-tensin/vs-shell@v2 @@ -330,9 +330,6 @@ jobs: cat > service_account_creds.json << EOF ${{ secrets.GCP_CREDS }} EOF - - uses: actions/setup-python@v1 - with: - python-version: "3.8" - name: Bazel on Windows env: BAZEL_VC: "C:/Program Files (x86)/Microsoft Visual Studio/2019/Enterprise/VC/" From 5006973e1d57cebbd04a19d66d2aba0b4e7beb28 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sun, 26 Feb 2023 18:29:38 +0000 Subject: [PATCH 16/24] Lint fix Signed-off-by: Yong Tang --- tensorflow_io/core/kernels/elasticsearch_kernels.cc | 3 ++- .../core/kernels/gsmemcachedfs/gs_memcached_file_system.h | 7 ++++--- .../kernels/gsmemcachedfs/memcached_file_block_cache.h | 2 +- 3 files changed, 7 insertions(+), 5 deletions(-) diff --git a/tensorflow_io/core/kernels/elasticsearch_kernels.cc b/tensorflow_io/core/kernels/elasticsearch_kernels.cc index 214dbef37..5ec6e8267 100644 --- a/tensorflow_io/core/kernels/elasticsearch_kernels.cc +++ b/tensorflow_io/core/kernels/elasticsearch_kernels.cc @@ -256,7 +256,8 @@ class ElasticsearchReadableResource : public ResourceBase { mutable mutex mu_; Env* env_ TF_GUARDED_BY(mu_); string url_; - tsl::CurlHttpRequest::Factory http_request_factory_ = tsl::CurlHttpRequest::Factory(); + tsl::CurlHttpRequest::Factory http_request_factory_ = + tsl::CurlHttpRequest::Factory(); std::vector base_dtypes_; std::vector base_columns_; diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/gs_memcached_file_system.h b/tensorflow_io/core/kernels/gsmemcachedfs/gs_memcached_file_system.h index 5fc772b39..b8b5060f0 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/gs_memcached_file_system.h +++ b/tensorflow_io/core/kernels/gsmemcachedfs/gs_memcached_file_system.h @@ -21,9 +21,9 @@ limitations under the License. #include #include "absl/memory/memory.h" +#include "tensorflow/tsl/platform/retrying_file_system.h" #include "tensorflow_io/core/kernels/gsmemcachedfs/gce_memcached_server_list_provider.h" #include "tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_system.h" -#include "tensorflow/tsl/platform/retrying_file_system.h" namespace tensorflow { @@ -50,8 +50,9 @@ class RetryingGsMemcachedFileSystem : public tsl::RetryingFileSystem { public: RetryingGsMemcachedFileSystem() - : tsl::RetryingFileSystem(absl::make_unique(), - tsl::RetryConfig(100000 /* init_delay_time_us */)) {} + : tsl::RetryingFileSystem( + absl::make_unique(), + tsl::RetryConfig(100000 /* init_delay_time_us */)) {} }; } // namespace tensorflow diff --git a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h index 2efcfa2d2..acd3be670 100644 --- a/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h +++ b/tensorflow_io/core/kernels/gsmemcachedfs/memcached_file_block_cache.h @@ -25,9 +25,9 @@ limitations under the License. #include #include "absl/container/flat_hash_map.h" -#include "tensorflow/tsl/platform/cloud/ram_file_block_cache.h" #include "tensorflow/core/framework/types.h" #include "tensorflow/core/platform/env.h" +#include "tensorflow/tsl/platform/cloud/ram_file_block_cache.h" #include "tensorflow_io/core/kernels/gsmemcachedfs/memcached_dao_interface.h" namespace tensorflow { From da8f6eadb683694a5043374e8a379f2b83906605 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sun, 26 Feb 2023 18:30:55 +0000 Subject: [PATCH 17/24] Github actions Signed-off-by: Yong Tang --- .github/workflows/build.yml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/.github/workflows/build.yml b/.github/workflows/build.yml index 79c4ba354..3d13ad46f 100644 --- a/.github/workflows/build.yml +++ b/.github/workflows/build.yml @@ -2,8 +2,8 @@ name: GitHub CI on: push: - # branches: - # - master + branches: + - master pull_request: branches: - master From e10837a33e67de20a708887aa6f3903753b77351 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sun, 26 Feb 2023 21:54:45 +0000 Subject: [PATCH 18/24] Try Signed-off-by: Yong Tang --- tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc | 1 + tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.h | 4 +++- tensorflow_io/core/kernels/io_kernel.h | 2 ++ 3 files changed, 6 insertions(+), 1 deletion(-) diff --git a/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc b/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc index 8e6fdf278..c6a52f2f4 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc @@ -17,6 +17,7 @@ limitations under the License. #include "arrow/io/stdio.h" #include "arrow/ipc/api.h" #include "arrow/result.h" +#include "tensorflow/core/framework/dataset_options.pb.h" #include "tensorflow/core/framework/dataset.h" #include "tensorflow/core/graph/graph.h" #include "tensorflow/core/public/version.h" diff --git a/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.h b/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.h index 46183d8fe..449701ff3 100644 --- a/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.h +++ b/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.h @@ -12,6 +12,8 @@ limitations under the License. #ifndef TENSORFLOW_DATA_CORE_KERNELS_AVRO_RECORD_DATASET_OP_H_ #define TENSORFLOW_DATA_CORE_KERNELS_AVRO_RECORD_DATASET_OP_H_ +#include "tensorflow/core/framework/dataset_metadata.pb.h" +#include "tensorflow/core/framework/dataset_options.pb.h" #include "tensorflow/core/framework/dataset.h" namespace tensorflow { @@ -36,4 +38,4 @@ class AvroRecordDatasetOp : public DatasetOpKernel { } // namespace data } // namespace tensorflow -#endif // TENSORFLOW_DATA_CORE_KERNELS_AVRO_RECORD_DATASET_OP_H_ \ No newline at end of file +#endif // TENSORFLOW_DATA_CORE_KERNELS_AVRO_RECORD_DATASET_OP_H_ diff --git a/tensorflow_io/core/kernels/io_kernel.h b/tensorflow_io/core/kernels/io_kernel.h index 0c86172df..4e374a95c 100644 --- a/tensorflow_io/core/kernels/io_kernel.h +++ b/tensorflow_io/core/kernels/io_kernel.h @@ -13,6 +13,8 @@ See the License for the specific language governing permissions and limitations under the License. ==============================================================================*/ +#include "tensorflow/core/framework/dataset_metadata.pb.h" +#include "tensorflow/core/framework/dataset_options.pb.h" #include "tensorflow/core/framework/op_kernel.h" #include "tensorflow/core/lib/strings/scanner.h" From 86c811804c0510fccad3f7d1dcd10b1469572ac9 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Mon, 27 Feb 2023 02:05:20 +0000 Subject: [PATCH 19/24] Try with ??_7Metadata@data@tensorflow@@6B@ Signed-off-by: Yong Tang --- third_party/toolchains/tf/BUILD.tpl | 1 + 1 file changed, 1 insertion(+) diff --git a/third_party/toolchains/tf/BUILD.tpl b/third_party/toolchains/tf/BUILD.tpl index 75e5499a8..6698d4818 100644 --- a/third_party/toolchains/tf/BUILD.tpl +++ b/third_party/toolchains/tf/BUILD.tpl @@ -55,6 +55,7 @@ genrule( " TF_SetStatusFromIOError", " TF_StartThread", " TF_VLog", + " ??_7Metadata@data@tensorflow@@6B@", "EOF", ]), ) From 0b5505053ff8f63f6cde2834e0069bbe1f5ebaec Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sun, 26 Feb 2023 19:19:11 -0800 Subject: [PATCH 20/24] Try2 Signed-off-by: Yong Tang --- third_party/toolchains/tf/BUILD.tpl | 2 ++ 1 file changed, 2 insertions(+) diff --git a/third_party/toolchains/tf/BUILD.tpl b/third_party/toolchains/tf/BUILD.tpl index 6698d4818..0e4c74db4 100644 --- a/third_party/toolchains/tf/BUILD.tpl +++ b/third_party/toolchains/tf/BUILD.tpl @@ -56,6 +56,8 @@ genrule( " TF_StartThread", " TF_VLog", " ??_7Metadata@data@tensorflow@@6B@", + " ??_7Options@data@tensorflow@@6B@", + " ??_7TensorProto@tensorflow@@6B@", "EOF", ]), ) From fc775cbbb45701c1512db264064724a2c0464937 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Thu, 9 Mar 2023 18:11:50 +0000 Subject: [PATCH 21/24] Bump to 2.12.0rc1 Signed-off-by: Yong Tang --- WORKSPACE | 6 +++--- tensorflow_io/python/ops/version_ops.py | 2 +- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/WORKSPACE b/WORKSPACE index f27922bb0..352bcce0e 100644 --- a/WORKSPACE +++ b/WORKSPACE @@ -115,10 +115,10 @@ pip_install() http_archive( name = "org_tensorflow", - sha256 = "0090d8c11a346f078acdf16ee83a64adf96c43eb1e34b971b90bb9bc5303a60b", - strip_prefix = "tensorflow-2.12.0-rc0", + sha256 = "1aee7a62742aae80bb94fcd83de49bf7317084ce64b24965967382ab082e5f52", + strip_prefix = "tensorflow-2.12.0-rc1", urls = [ - "https://github.com/tensorflow/tensorflow/archive/refs/tags/v2.12.0-rc0.tar.gz", + "https://github.com/tensorflow/tensorflow/archive/refs/tags/v2.12.0-rc1.tar.gz", ], ) diff --git a/tensorflow_io/python/ops/version_ops.py b/tensorflow_io/python/ops/version_ops.py index c50c2e8d5..94aea759b 100644 --- a/tensorflow_io/python/ops/version_ops.py +++ b/tensorflow_io/python/ops/version_ops.py @@ -15,4 +15,4 @@ """version_ops""" version = "0.31.0" -require = "tensorflow>=2.12.0rc0,<2.13.0" +require = "tensorflow>=2.12.0rc1,<2.13.0" From d2df9a42edd36c9a9ba569acb96084c839c2b903 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Fri, 10 Mar 2023 18:22:03 +0000 Subject: [PATCH 22/24] Disable windows build for now Signed-off-by: Yong Tang --- .github/workflows/build.yml | 52 ++++++++++++++++++------------------- 1 file changed, 26 insertions(+), 26 deletions(-) diff --git a/.github/workflows/build.yml b/.github/workflows/build.yml index 3d13ad46f..536554a25 100644 --- a/.github/workflows/build.yml +++ b/.github/workflows/build.yml @@ -449,7 +449,7 @@ jobs: release: name: Release if: github.event_name == 'push' - needs: [linux-wheel, macos-wheel, windows-wheel] + needs: [linux-wheel, macos-wheel] #, windows-wheel] runs-on: ubuntu-18.04 steps: - uses: actions/download-artifact@v1 @@ -492,26 +492,26 @@ jobs: with: name: Linux-3.11-wheel path: Linux-3.11-wheel - - uses: actions/download-artifact@v1 - with: - name: Windows-3.7-wheel - path: Windows-3.7-wheel - - uses: actions/download-artifact@v1 - with: - name: Windows-3.8-wheel - path: Windows-3.8-wheel - - uses: actions/download-artifact@v1 - with: - name: Windows-3.9-wheel - path: Windows-3.9-wheel - - uses: actions/download-artifact@v1 - with: - name: Windows-3.10-wheel - path: Windows-3.10-wheel - - uses: actions/download-artifact@v1 - with: - name: Windows-3.11-wheel - path: Windows-3.11-wheel + #- uses: actions/download-artifact@v1 + # with: + # name: Windows-3.7-wheel + # path: Windows-3.7-wheel + #- uses: actions/download-artifact@v1 + # with: + # name: Windows-3.8-wheel + # path: Windows-3.8-wheel + #- uses: actions/download-artifact@v1 + # with: + # name: Windows-3.9-wheel + # path: Windows-3.9-wheel + #- uses: actions/download-artifact@v1 + # with: + # name: Windows-3.10-wheel + # path: Windows-3.10-wheel + #- uses: actions/download-artifact@v1 + # with: + # name: Windows-3.11-wheel + # path: Windows-3.11-wheel - run: | set -e -x mkdir -p wheelhouse @@ -525,11 +525,11 @@ jobs: cp Linux-3.9-wheel/*.whl wheelhouse/ cp Linux-3.10-wheel/*.whl wheelhouse/ cp Linux-3.11-wheel/*.whl wheelhouse/ - cp Windows-3.7-wheel/*.whl wheelhouse/ - cp Windows-3.8-wheel/*.whl wheelhouse/ - cp Windows-3.9-wheel/*.whl wheelhouse/ - cp Windows-3.10-wheel/*.whl wheelhouse/ - cp Windows-3.11-wheel/*.whl wheelhouse/ + # cp Windows-3.7-wheel/*.whl wheelhouse/ + # cp Windows-3.8-wheel/*.whl wheelhouse/ + # cp Windows-3.9-wheel/*.whl wheelhouse/ + # cp Windows-3.10-wheel/*.whl wheelhouse/ + # cp Windows-3.11-wheel/*.whl wheelhouse/ ls -la wheelhouse/ sha256sum wheelhouse/*.whl - uses: actions/upload-artifact@v2 From 5227447fe9bcc6cb716a2fe4b9a607c1aa542fa6 Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Sat, 11 Mar 2023 22:50:42 +0000 Subject: [PATCH 23/24] Fix lint Signed-off-by: Yong Tang --- tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc b/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc index c6a52f2f4..f6ee8a434 100644 --- a/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc +++ b/tensorflow_io/core/kernels/arrow/arrow_dataset_ops.cc @@ -17,8 +17,8 @@ limitations under the License. #include "arrow/io/stdio.h" #include "arrow/ipc/api.h" #include "arrow/result.h" -#include "tensorflow/core/framework/dataset_options.pb.h" #include "tensorflow/core/framework/dataset.h" +#include "tensorflow/core/framework/dataset_options.pb.h" #include "tensorflow/core/graph/graph.h" #include "tensorflow/core/public/version.h" #include "tensorflow_io/core/kernels/arrow/arrow_kernels.h" From 2f62c56f917d2bff3e98ce417d0907dc81d9c0ff Mon Sep 17 00:00:00 2001 From: Yong Tang Date: Mon, 13 Mar 2023 14:18:34 +0000 Subject: [PATCH 24/24] Fix lint issue Signed-off-by: Yong Tang --- tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.h | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.h b/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.h index 449701ff3..3d1a8668a 100644 --- a/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.h +++ b/tensorflow_io/core/kernels/avro/avro_record_dataset_kernels.h @@ -12,9 +12,9 @@ limitations under the License. #ifndef TENSORFLOW_DATA_CORE_KERNELS_AVRO_RECORD_DATASET_OP_H_ #define TENSORFLOW_DATA_CORE_KERNELS_AVRO_RECORD_DATASET_OP_H_ +#include "tensorflow/core/framework/dataset.h" #include "tensorflow/core/framework/dataset_metadata.pb.h" #include "tensorflow/core/framework/dataset_options.pb.h" -#include "tensorflow/core/framework/dataset.h" namespace tensorflow { namespace data {