Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

optimize Send Message to Groups by changing Client => encryptMessageForDevices #721

Open
wants to merge 3 commits into
base: main
Choose a base branch
from

Conversation

Mahdi-Ostadrahim
Copy link

@Mahdi-Ostadrahim Mahdi-Ostadrahim commented Dec 26, 2024

hello there
thank you for the library
we are using it on a project of ours with currently about 150 clients. Sending message in them had gotten extremely slow, especially in large groups, so I performed some changes on it to improve the speed. I decided to use some of those changes in the main library as well, in case someone was having the same issue. the changes are as following:

increased send message performance by collecting all neccessary device sessions and keys related to the message encryption and storing them into a cache, then storing them all into the db at once to minimize I/O to the database

…e sessions and keys related to the message encryption and storing them into a cache, then storing them all into the db at once to minimize I/O to the database
@HoshenKadosh
Copy link
Contributor

@tulir can this be merged please?

Copy link
Contributor

@devlikepro devlikepro left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The solution likely will solve the issue, but the price we'll have to pay later on maintenance with that code is too much, IMO.

If we could hide all "caches" inside the Store (so the caller encryptMessageForDevices doesn't care about caches and always get the final result - it'd be great! And reducing number of "caches" would also help, right now you move entities from one to another in a bit complex way.

ℹ️ It's not an official maintainer opinion, I was just passing by the PR :)

@@ -126,6 +126,13 @@ type PrivacyTokenStore interface {
GetPrivacyToken(user types.JID) (*PrivacyToken, error)
}

type CacheStore interface {
GetSessions(addresses []string) map[string][]byte
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Instead of using primitives (string/byte) we should try using already defined types, like types.JID (likely there's something for Session/IdentityKeys as well)

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

IdentityKeys and Session types are both []byte in the code which is the reason I chose that

Comment on lines +164 to +171
Cache CacheStore
Container DeviceContainer

DatabaseErrorHandler func(device *Device, action string, attemptIndex int, err error) (retry bool)

//Cache to Temporary save sessions and identity keys for faster group send
SessionsCache map[string][]byte
IdentityKeysCache map[string][32]byte
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks like too many Caches here, I lost the idea when we use SessionCache, when we use IdentityKeysCache and when just Cache.

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hello
Sorry for the delay, I didn't check my github in a while
So the idea of CacheStore is to a have a way that implements the method to get all the required IdentityKeys and Sessions at once
SessionsCache and IdentityCache save the result of said query. While doing it on redis will certainly improve speed, I would be forcing the implementation of the library to be dependent on that DB

}
//clear the cache once the encryption is done to release memory
clear(cli.Store.SessionsCache)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

interesting that we clear cache after every call - why does it need to be available global then?

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Since the I/O to the db is being used inside LoadSession and IsTrustedIdentity in store/signal.go, the check had to be made whether we already have it or not. Since both of these functions are whatsmeows implementation of another library, this was the only suitable place I could find to place it in

@purpshell
Copy link
Contributor

I think that, like @devlikepro said, having multiple ambiguous caches isn't very sustainable.

I'd suggest you implement your own SQLStore layer that hits your own custom cache first. You can also look into other options like Redis at that point.

@Nubuki-all
Copy link

@tulir any chance you or any other maintainer are working on something that fixes this ?

@tulir
Copy link
Owner

tulir commented Feb 27, 2025

The proper solution is updating libsignal-protocol-go as well as the sql store to use contexts, after which the code can use a native db transaction. Not planned in the near future, but will probably be added at some point

@Mahdi-Ostadrahim
Copy link
Author

I think that, like @devlikepro said, having multiple ambiguous caches isn't very sustainable.

I'd suggest you implement your own SQLStore layer that hits your own custom cache first. You can also look into other options like Redis at that point.

Hello
Sorry for the late response
I didn't want to force the library to be dependent on a DB like redis
That said, I like the idea of merging the caches into a single cache entity. beyond that do you have any suggestions that improves this approach?

@Mahdi-Ostadrahim
Copy link
Author

Mahdi-Ostadrahim commented Feb 28, 2025

The proper solution is updating libsignal-protocol-go as well as the sql store to use contexts, after which the code can use a native db transaction. Not planned in the near future, but will probably be added at some point

If I'm not mistaken the approach you're suggesting is wrapping the entire process in a transaction, which wouldn't really solve the main issue which is too many I/Os to the database.
Or did I misinterpret something there?

Thank you for your response

@purpshell
Copy link
Contributor

I think that, like @devlikepro said, having multiple ambiguous caches isn't very sustainable.
I'd suggest you implement your own SQLStore layer that hits your own custom cache first. You can also look into other options like Redis at that point.

Hello Sorry for the late response I didn't want to force the library to be dependent on a DB like redis That said, I like the idea of merging the caches into a single cache entity. beyond that do you have any suggestions that improves this approach?

I meant implementing this on your own outside of the library, in my original approach.

@theeusmartins
Copy link

i need this

@GustavoGSA
Copy link

I try this PR in my fork and o have this error

image

@GustavoGSA
Copy link

Individual mesage @http://s.whatsapp.net no working to with this PR, i send mutiples message and stop send

@Mahdi-Ostadrahim
Copy link
Author

I try this PR in my fork and o have this error

image

You have to share the version you're using for me to understand what's wrong in your fork
The line you have logged is 'queryValues += ","' in my fork which has no relation to the error you are displaying

@GustavoGSA
Copy link

I try this PR in my fork and o have this error
image

You have to share the version you're using for me to understand what's wrong in your fork The line you have logged is 'queryValues += ","' in my fork which has no relation to the error you are displaying

I use the current version of the main branch, I forked it and then applied this PR

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

9 participants