I am trying to come up with a way to measure the entropy of a password that is easy to remember, based on a list of common English words, loosely based on this XKCD comic. I'd like to know if my math is correct or if my assumptions about "easy to remember" are flawed.
I'll consider the number of common English words in the dictionary to be the variable d.
I'll consider the number of words to use in the password to be the variable n.
If the US-English keyboard is considered to contain all the likely characters that would reasonably make up a password that is easy to remember, I count 96 total symbols that can be directly keyed, including uppercase and lowercase letters.
Those are:
TAB SPACE ABCDEFGHIJKLMNOPQRSTUVWXYZ abcdefghijklmnopqrstuvwxyz 0123456789 ~`!@#$%^&*()_-+= {[}]|\ :;"'<,>.?/
I'll consider these characters to be extra complexity which can be added to the dictionary words, or the variable e as 96.
So that they are easier to remember, I'll only consider adding symbols to the beginning or end of a word, or as a word by themselves, but not arbitrarily placed inside of a dictionary word. That should mean there are 2*n+n+1 positions available for each symbol which is added. I'll use the variable s for the number of extra complexity symbols added.
The equation for the total possible combinations in use should then be:
combinations = d^n + e^(s*(2*n+n+1))
Therefore, the number of bits of entropy this kind of password provides should be:
bits = log2(commbinations)
Is my math correct?
Are my assumptions about rules for an easy to remember password flawed?