A realization that this leaves out something essential Nothing so fundamental lies in the realm of concern to us aggregate humans, where the need is, now, for the study of real complexity, not idealized simplicity. In every field except high-energy physics on one hand, and cosmology on the other, one hears the same. The immense understanding that has come from digging deeper to atomic explanations has been followed by a realization that this leaves out something essential. In its rapid advance, science has had to ignore the fact that a whole is more than the sum of its parts. Matter versus Materials: A Historical View knowledgecomplexityholism
Measured by the number of its features A primary cause of complexity is that software vendors uncritically adopt almost any feature that users want. Any incompatibility with the original system concept is either ignored or passes unrecognized, which renders the design more complicated and its use more cumbersome. When a system's power is measured by the number of its features, quantity becomes more important than quality. Every new release must offer additional features, even if some don't add functionality. Niklaus Wirth, A Plea for Lean Software featuresqualitycomplexity
Features and complexity Niklaus Wirth of Pascal fame wrote a famous paper in 1995 called A Plea for Lean Software. His take is that “a primary cause for the complexity is that software vendors uncritically adopt almost any feature that users want”, and “when a system’s power is measured by the number of its features, quantity becomes more important than quality”. Ben Hoyt, The small web is beautiful benhoyt.com A Plea for Lean SoftwareSpeed is a featureRequirements proliferation featurescomplexity
The multiplicity of living patterns The more living patterns there are in a thing—a room, a building, or a town—the more it comes to life as an entirety, the more it glows, the more it has this self-maintaining fire, which is the quality without a name. Christopher Alexander, The Timeless Way of Building The right overlapEach element performs many functions complexity
The kind of problem a city is Dr. Weaver lists three stages of development in the history of scientific thought: (1) ability to deal with problems of simplicity; (2) ability to deal with problems of disorganized complexity; and (3) ability to deal with problems of organized complexity. The history of modern thought about cities is unfortunately very different from the history of modern thought about the life sciences. The theorists of conventional modern city planning have consistently mistaken cities as problems of simplicity and of disorganized complexity, and have tried to analyze and treat them thus. Jane Jacobs, The Death and Life of Great American Cities Order Out of ChaosOrder Without Design problemscitiescomplexity
The difficulty of designing complexity Designers, limited as they must be by the capacity of the mind to form intuitively accessible structures, cannot achieve the complexity of the semilattice in a single mental act. The mind has an overwhelming predisposition to see trees wherever it looks and cannot escape the tree conception. Experiments suggest strongly that people have an underlying tendency, when faced by a complex organization, to reorganize it mentally in terms of non-overlapping units. The complexity of the semilattice is replaced by the simpler and more easily grasped tree form. Christopher Alexander, A City Is Not a Tree complexityintuitiondesign
Structural complexity The idea of overlap, ambiguity, multiplicity of aspect, and the semilattice are not less orderly than the right tree, but more so. They represent a thicker, tougher, more subtle and more complex view of structure. Christopher Alexander, A City Is Not a Tree structurecomplexity
Losing meaning The people who’ve proven that they can make very good individual products with the radical focus of a spotlight seem to be pushed ever further from making good ecosystems. Products are being made “consistent” with the application of so-called “design patterns,” and rather than bringing coherence to these various touch-points, the painting-on of interface standards and interaction patterns did something far less valuable. Rote consistency, in the way many seem to be going about it (Material Design being just one example), is at odds with making things be good. It simplifies what needs to remain complex. Always, when simplification is underway, meaning is being lost. Dan Klyn, What Good Means complexitysoftware
Notes on the Legibility War An Article by David R. MacIver notebook.drmaciver.com The basic idea of legibility is that the act of making something comprehensible enough to control is itself an act that shapes the thing to be controlled, often with far greater consequences than the control itself. This is because it removes complexity that is deemed as irrelevant that makes it harder to control, and that complexity may be in some way essential to the health of the system. controlsystemscomplexitylegibility
The return of fancy tools An Article by Tom MacWright macwright.com Technology is seeing a little return to complexity. Dreamweaver gave way to hand-coding websites, which is now leading into Webflow, which is a lot like Dreamweaver. Evernote give way to minimal Markdown notes, which are now becoming Notion, Coda, or Craft. Visual Studio was “disrupted” by Sublime Text and TextMate, which are now getting replaced by Visual Studio Code. JIRA was replaced by GitHub issues, which is getting outmoded by Linear. The pendulum swings back and forth, which isn’t a bad thing complexitysimplicitytoolssoftwaretechnologynotetaking
On the other side of complexity A Quote "I would not give a fig for the simplicity this side of complexity, but I would give my life for the simplicity on the other side of complexity." — Oliver Wendell Holmes Jr. Don't Rush to Simplicity simplicitycomplexity
The Fidelity Curve An Article by Ryan Singer m.signalvnoise.com How do we choose which level of fidelity is appropriate for a project? I think about it like this: The purpose of making sketches and mockups before coding is to gain confidence in what we plan to do. I’m trying to remove risk from the decision to build something by somehow “previewing” it in a cheaper form. There’s a trade-off here. The higher the fidelity of the mockup, the more confidence it gives me. But the longer it takes to create that mockup, the more time I’ve wasted on an intermediate step before building the real thing. I like to look at that trade-off economically. Each method reduces risk by letting me preview the outcome at lower fidelity, at the cost of time spent on it. The cost/benefit of each type of mockup is going to vary depending on the fidelity of the simulation and the work involved in building the real thing. Four levels of fidelityTime to build versus confidence gained prototypesinterfaces
Four levels of fidelity Suppose we have four levels of fidelity… Rough sketch (on paper or an iPad) Static mock-up (eg. Photoshop or Sketch) Interactive mock-up (eg. Framer, InVision) Working code prototype (HTML/CSS, iOS views) Depending on the feature you’re working on, these levels of fidelity take different amounts of time to create. If you plot them in terms of time to build versus confidence gained, you could imagine something like a per-feature fidelity curve.
Time to build versus confidence gained Show image 0 Show image 1 Show image 2 Take a simple CRUD web UI, where you’re just navigating between screens. It doesn’t take much more time to build the real version than it does to mock it when the design is simple. If you were to build out an interactive mock first, you would end up spending twice as much time in total without gaining much out of it. Contrast that with a complicated Javascript interaction. Or a native iOS feature that requires programmer time to build out. If it takes substantially more time to build the real code version, then it may be smart to do an interactive mockup first.