Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Performance issues in examples/mnist/estimator (by P3) #573

Closed
DLPerf opened this issue Aug 22, 2021 · 3 comments
Closed

Performance issues in examples/mnist/estimator (by P3) #573

DLPerf opened this issue Aug 22, 2021 · 3 comments

Comments

@DLPerf
Copy link

DLPerf commented Aug 22, 2021

Hello! I've found a performance issue in examples/mnist/estimator: batch() should be called before map(), which could make your program more efficient. Here is the tensorflow document to support it.

Detailed description is listed below:

  • in mnist_spark_streaming.py: .batch(BATCH_SIZE)(here) should be called before .map(scale)(here).
  • in mnist_spark.py: .batch(BATCH_SIZE)(here) should be called before .map(scale)(here).
  • in mnist_pipeline.py: .batch(BATCH_SIZE)(here) should be called before .map(scale)(here).

Besides, you need to check the function called in map()(e.g., scale called in .map(scale)) whether to be affected or not to make the changed code work properly. For example, if scale needs data with shape (x, y, z) as its input before fix, it would require data with shape (batch_size, x, y, z).

Looking forward to your reply. Btw, I am very glad to create a PR to fix it if you are too busy.

@leewyang
Copy link
Contributor

closing bot activity.

@DLPerf
Copy link
Author

DLPerf commented Aug 31, 2021

I'm not a robot! @tmielika

@DLPerf
Copy link
Author

DLPerf commented Nov 4, 2021

fuck u!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants