The Laws (was the principles) of secure information systems design
Peter Fairbrother
peter@m-o-o-t.org
Tue Jul 12 14:19:04 PDT 2016
I've been revising the principles, and came up with this. It's an early
version.
As ever, corrections and suggestions are welcome.
Calling them Laws is perhaps a bit overreaching - but on reflection I
thought that's mostly what they are, break them and the system won't be
secure.
I will put the Laws up on the 'net shortly, hopefully with a link for
suggestions and comments.
The Laws of secure information systems design:
Law 0: It's all about who is in control
Law 1: Someone else is after your data
Law 2: If it isn't stored it can't be stolen
Law 3: Only those you trust can betray you
Law 4: Attack methods are many, varied, ever-changing and eternal
Law 5: The entire system is subject to attack
Law 6: A more complex system has more places to attack
Law 7: Holes for good guys are holes for bad guys too
Law 8: Kerckhoffs's Principle rulez! - usually...
Law 9: A system which is hard to use will be abused or unused
law 10: Design for future threats
Law 11: Security is a Boolean
Law 12: People offering the impossible are lying
Law 13: Nothing ever really goes away
Law 15: "Schneier's law c" [1] holds illimitable dominion over all...
including these laws
-- Peter Fairbrother
[1] "
a: Anyone, from the most clueless amateur to the best cryptographer, can
create an algorithm that he himself can't break. It's not even hard.
b: What is hard is creating an algorithm that no one else can break,
even after years of analysis.
c: And the only way to prove that is to subject the algorithm to years
of analysis by the best cryptographers around."
More information about the cypherpunks
mailing list