Research & Ethnography
What is unspoken
A nested classificatory hierarchy
Unfinished
Record them all
Recommendations for field notes
In Defense of Browsing
The illustrated guide to a Ph.D.
When users never use the features they asked for
An Article by Austin Z. HenleyWe deployed our tool. Almost no one used it.
The handful that did use it, used it once or twice and barely interacted with it. After a few days, zero people were using it.
Why did they tell me they wanted these features?
Why Most Published Research Findings Are False
A Research Paper by John P.A. IoannidisThere is increasing concern that most current published research findings are false. The probability that a research claim is true may depend on study power and bias, the number of other studies on the same question, and, importantly, the ratio of true to no relationships among the relationships probed in each scientific field. In this framework, a research finding is less likely to be true when the studies conducted in a field are smaller; when effect sizes are smaller; when there is a greater number and lesser preselection of tested relationships; where there is greater flexibility in designs, definitions, outcomes, and analytical modes; when there is greater financial and other interest and prejudice; and when more teams are involved in a scientific field in chase of statistical significance.
How can we develop transformative tools for thought?
A Research Paper by Andy Matuschak & Michael NielsenConventional tech industry product practice will not produce deep enough subject matter insights to create transformative tools for thought.
...The aspiration is for any team serious about making transformative tools for thought. It’s to create a culture that combines the best parts of modern product practice with the best parts of the (very different) modern research culture. You need the insight-through-making loop to operate, whereby deep, original insights about the subject feed back to change and improve the system, and changes to the system result in deep, original insights about the subject.
A hypothesis is a liability
A Research Paper by Itai Yanai & Martin LercherThere is a hidden cost to having a hypothesis. It arises from the relationship between night science and day science, the two very distinct modes of activity in which scientific ideas are generated and tested, respectively [1, 2]. With a hypothesis in hand, the impressive strengths of day science are unleashed, guiding us in designing tests, estimating parameters, and throwing out the hypothesis if it fails the tests. But when we analyze the results of an experiment, our mental focus on a specific hypothesis can prevent us from exploring other aspects of the data, effectively blinding us to new ideas.
Keep digging
An Article by Ryan SingerThe hardest thing about customer interviews is knowing where to dig. An effective interview is more like a friendly interrogation. We don’t want to learn what customers think about the product, or what they like or dislike — we want to know what happened and how they chose... To get those answers we can’t just ask surface questions, we have to keep digging back behind the answers to find out what really happened.
Fast Path to a Great UX – Increased Exposure Hours
An Article by Jared SpoolAs we’ve been researching what design teams need to do to create great user experiences, we’ve stumbled across an interesting finding. It’s the closest thing we’ve found to a silver bullet when it comes to reliably improving the designs teams produce.
The solution? Exposure hours. The number of hours each team member is exposed directly to real users interacting with the team’s designs or the team’s competitor’s designs. There is a direct correlation between this exposure and the improvements we see in the designs that team produces.
Weighing up UX
An Article by Jeremy KeithMetrics come up when we’re talking about A/B testing, growth design, and all of the practices that help designers get their seat at the table (to use the well-worn cliché). But while metrics are very useful for measuring design’s benefit to the business, they’re not really cut out for measuring user experience.
Monkeys testing random designs
A Tweet by Jared SpoolA/B testing is an effective approach to use science to design and deliver deeply-frustrating user experiences.
A/B testing without upfront research is just random monkeys testing random designs to see which of those designs do “best” against random criteria.
If drug testing was actually implemented like most A/B tests, you’d give 2 drugs to 2 groups of people and pick the “winner” by whichever group had fewer deaths.
You and Your Research
A Speech by Richard HammingThis talk centered on Hamming's observations and research on the question "Why do so few scientists make significant contributions and so many are forgotten in the long run?"
The vanishing designer
Visionary designers have lost their conceptual integrity to an industrial complex optimized for consensus, predictability, and short-term business gain. The rise of customer-obsession mantra and data-driven culture cultivated a generation of designers who only take risk-free and success-guaranteed steps towards the inevitable local maxima of design monotony.
The same, the same, the same
Look around us. Every business is an app and every app feels the same, because every designer has the same resume, follows the same process, graduates from the same program, uses the same tool, scrolls the same Dribbble feed, reads the same Medium articles, expects the same career outcome, lives in the same ideology bubble.
Design as an engineering problem
The Silicon Valley giants, testifying with their runaway success, claimed to have “solved” design as an engineering problem. The solution substituted the human essence of design — intuition, ingenuity, and taste— with the tangibles, measurables, and deliverables.
Companies say they are “design-driven”, but designers are actually driven by dashboards filled with metrics like CSAT, NPS, CES, DAU, MAU. We rigorously run tests, studies, experiments as if innovative ideas are hidden in spreadsheets, waiting to be extracted by data scientists.
The heat death of design
Losing the design diversity means falling into a singular narrative of how design must be done, which grants unfair and self-reinforcing advantage to the mainstream while discouraging, stifling, or even punishing the idiosyncratic designers who bring unorthodox but remarkably innovative processes to the table. The true opportunity cost is the diverse future that humanity can no longer access.
A future without diversity is fundamentally stagnant: imagine designs so standardized that you can’t tell them apart. While every design is guaranteed to be good, none will be great. New designs are marginally better than previous ones with the rate of improvement eventually approaching zero. We have reached the heat death of design.
Design with courage
- Make a bold decision (that is controversial).
- Make a mistake (as a result of a bold decision).
- Challenge “conventional wisdom”.
- Challenge authority (that preaches conventional wisdom).
- Challenge hierarchy (that perpetuates conventional wisdom).
- Ignore the committee (and the need to converge).
- Decide who your clients are (and aren’t).
- Ignore clients that aren’t (especially those who pay the most).
- Cultivate clients if none exists (instead of compromising your design).
- Be a generalist (and ignore your job title).
- Be a specialist (who specializes in being a generalist).
- Design things from scratch (and build them yourself from scratch).
- Design things that no one wants (yet).
- Design freely (and think freely).