-
Notifications
You must be signed in to change notification settings - Fork 2.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[stdlib][feature request] use Tuple[K,V]
instead of DictEntry[K,V]
#4138
Conversation
Signed-off-by: Illia Sheshyn <[email protected]>
Signed-off-by: Illia Sheshyn <[email protected]>
self.size -= 1 | ||
return entry_value^.reap_value() | ||
return entry_value[1] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't actually think this is right, but I haven't learnt Mojo to the extent of refactoring this bit correctly.
What is the impact of this change in terms of performance? |
I also want to know that. FYI public benchmarking script is currently broken, so you will need to use the trick from #4054 for now |
I'll run the benchmarks 👍 Does the benchmarking script run with the wrong stdlib without the patch @martinvuyk ? Also what's the approximate runtime to execute all 9 benchmarks? For reference, I'm on M1. |
yes It currently uses the packaged stdlib instead of the in-tree one
No idea, never timed the dict benchmarks. You can try it with fewer dict sizes first 🤷♂️ |
After running the benchmarks I no longer think this is a good idea. main:
vs the current branch:
Surfacing the |
I don't think that sacrificing performance for Python compatibility is the way to go for language which aims to squeeze every bit of performance from hardware. I don't see sense of being compatible with Python but at moderate speed (which this compatibility will cause). In such case why to even bother to choose Mojo if there is already C++ for CPU or CUDA for GPU (not speaking about raw assembly) which do not care about Python ballast and will be way faster? |
I also don't think it's worth sacrificing performance yet. Our dict is still slower than other implementations. My hope is that in the future we get to control how a type is unpacked with a dunder method, and that way I also think we could make |
I'd love to see priorities as follow:
|
I'd like to know why there is any performance difference at all. I thought they would generate identical assembly. |
My understanding so far is the change in data locality. This code is doing an additional index for the key's hash, which is now in its own array, while the key and hash are very frequently read together. I'll think about it some more, but the problem that I see would be constructing a tuple for the entry if I were to change how those values are stored within |
I see. I missed the part where the hash is no longer stored in the entry. So this PR is actually two changes combined: moving the hash out of line and converting the struct to a tuple. We need to be careful with the former I think. |
@soraros yeah, that was just an idea I thought was easy to try out. Now that I see the performance impact I don't want to proceed with this change. Hope that the PR assists future discussions in some way. |
Signed-off-by: Illia Sheshyn [email protected]
as per #2554
Changes
Dict
type owns computed key hashesDictEntry
type