Skip to content

Commit

Permalink
Add streamz metric graph_optimization_cache_restore_count.
Browse files Browse the repository at this point in the history
PiperOrigin-RevId: 531383819
  • Loading branch information
MaidouPP authored and tensorflower-gardener committed May 12, 2023
1 parent 5feae80 commit 59ea4ff
Show file tree
Hide file tree
Showing 2 changed files with 29 additions and 6 deletions.
28 changes: 22 additions & 6 deletions tensorflow/core/framework/metrics.cc
Original file line number Diff line number Diff line change
Expand Up @@ -293,6 +293,12 @@ auto* graph_optimization_cache_miss_count = tsl::monitoring::Counter<1>::New(
"source" // graph optimization source
);

auto* graph_optimization_cache_load_count = tsl::monitoring::Counter<1>::New(
"/tensorflow/core/graph_optimization_cache_load_count",
"The number of times loading an optimized function graph to RAM.",
"source" // graph optimization source
);

auto* xla_compilations = tsl::monitoring::Counter<0>::New(
"/tensorflow/core/xla_compilations",
"The number of XLA compilations used to collect "
Expand Down Expand Up @@ -449,8 +455,7 @@ void RecordTFDataServiceWorkerCreated() {
}

void RecordTFDataServiceJobsCreated(
const data::ProcessingModeDef& processing_mode,
bool is_coordinated_read) {
const data::ProcessingModeDef& processing_mode, bool is_coordinated_read) {
const std::string sharding_policy_str =
data::ProcessingModeDef::ShardingPolicy_Name(
processing_mode.sharding_policy());
Expand All @@ -463,8 +468,7 @@ void RecordTFDataServiceJobsCreated(

void RecordTFDataServiceClientIterators(
int64_t worker_uid, data::DeploymentMode deployment_mode,
const data::ProcessingModeDef& processing_mode,
bool is_coordinated_read) {
const data::ProcessingModeDef& processing_mode, bool is_coordinated_read) {
const std::string deployment_mode_str =
data::DeploymentMode_Name(deployment_mode);
const std::string sharding_policy_str =
Expand Down Expand Up @@ -519,8 +523,7 @@ void RecordTFDataFilename(const string& name, const string& filename) {
tf_data_filename_counter->GetCell(name, filename)->IncrementBy(1);
}

void RecordTFDataAutoShard(const string& id,
data::AutoShardPolicy policy,
void RecordTFDataAutoShard(const string& id, data::AutoShardPolicy policy,
int64 num_workers, int64 num_replicas) {
tf_data_auto_shard->GetCell(id, "policy")->Set(static_cast<int64_t>(policy));
tf_data_auto_shard->GetCell(id, "num_workers")->Set(num_workers);
Expand Down Expand Up @@ -671,6 +674,19 @@ int64_t GetFunctionGraphOptimizationCacheMissCount(
return graph_optimization_cache_miss_count->GetCell(mapped_source)->value();
}

void IncrementFunctionGraphOptimizationCacheLoadCount(
int count, GraphOptimizationSource source) {
std::string mapped_source = GraphOptimizationSourceMapping(source);
graph_optimization_cache_load_count->GetCell(mapped_source)
->IncrementBy(count);
}

int64_t GetFunctionGraphOptimizationCacheLoadCount(
GraphOptimizationSource source) {
std::string mapped_source = GraphOptimizationSourceMapping(source);
return graph_optimization_cache_load_count->GetCell(mapped_source)->value();
}

void UpdateTpuVariableDistributionTime(const uint64 distribution_time_usecs) {
if (distribution_time_usecs > 0) {
tpu_variable_distribution_time_usecs->GetCell()->IncrementBy(
Expand Down
7 changes: 7 additions & 0 deletions tensorflow/core/framework/metrics.h
Original file line number Diff line number Diff line change
Expand Up @@ -259,6 +259,13 @@ void IncrementFunctionGraphOptimizationCacheMissCount(
int64_t GetFunctionGraphOptimizationCacheMissCount(
GraphOptimizationSource source);

// Increments the number of restoring function graph optimization cache.
void IncrementFunctionGraphOptimizationCacheLoadCount(
int count, GraphOptimizationSource source);

int64_t GetFunctionGraphOptimizationCacheLoadCount(
GraphOptimizationSource source);

// Records the activity of the first phase of the mlir bridge using the
// tf_metadata.tf_mlir_bridge_first_phase_count metric.
// device_type: tpu, cpu, gpu, etc.
Expand Down

0 comments on commit 59ea4ff

Please sign in to comment.