-
Notifications
You must be signed in to change notification settings - Fork 69
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feature: cache the columnToFieldIndex map for each struct type #25
Comments
Thanks for submitting your issue. I will take a deeper look at your idea and the benchmarks and get back to you soon! |
Sorry, I've been pretty busy at work and couldn’t find enough time to work on this issue yet. I think what you are proposing is good and will improve the performance of the library (I actually want to measure it on my own also). I hope I will be able to find the time to work on this during this month. But If you want it to be done earlier or you are just interested in tackling it on your own, I encourage you to propose a PR with the solution. I understand it won’t be that easy though, since it involves some design changes. |
Hey! Now when I introduced a package-level API object we can easily enhance it with caching logic. |
Thanks for your library.
Because you can't update a struct dynamically in go, when scanning a row into a struct there isn't much reason to create a columnToFieldIndex map for the same struct more than once.
I created a test struct with 1024 fields and some quick benchmarks to see if the cache helps, you can see the benchmarks in my fork of scanny here.
To implement a cache I only changed the
getColumnToFieldIndexMap
function which will attempt to get a column to field index map from and return it if it exists. The cache can be implemented using both amap[reflect.Type]
with async.RWMutex
or with async.Map
. I tested both in the benchmarks.The BenchmarkStruct functions reuse the same row scanner for each iteration. The BenchmarkScannerStruct functions create a new row scanner for each iteration. The benchmarks that end in MapCache use a
map[reflect.Type]
with async.RWMutex
, while the SyncMapCache benchmarks use async.Map
to store the column to field index maps. The results of the benchmarks:When reusing a row scanner there isn't much difference to the performance when using a cache. The real benefit happens when you create a new row scanner each iteration. The allocs drop from 3081 to only 4. And both the bytes/op and ns/op drop by over three times.
I think this would be a useful feature to add since even though having 1024 fields in a struct is pretty extreme but from the benchmarks the
getColumnToFieldIndexMap
function is creating 3 allocs per field of a struct which can be avoided after the first call togetColumnToFieldIndexMap
with the same struct without much overhead.The text was updated successfully, but these errors were encountered: