SYDE 261: Culture of Design, Impacts
Estimated study time: 7 minutes
Table of contents
Sources and References
- Norman, The Design of Everyday Things (Basic Books)
- Winner, The Whale and the Reactor: A Search for Limits in an Age of High Technology (University of Chicago)
- Papanek, Design for the Real World (Thames & Hudson)
- Vermaas, Kroes, van de Poel, Franssen, Houkes, A Philosophy of Technology (Morgan & Claypool)
- Online: Stanford d.school resources, Engineering Change Lab Canada reports
Chapter 1: Technology, Society, and Values
1.1 Technology Is Not Neutral
Technologies embed values. Langdon Winner’s argument that “artifacts have politics” highlights how design choices encode power: a bridge too low for buses excludes certain populations; an algorithm tuned on biased data reproduces discrimination; a city wired for cars marginalises pedestrians and transit. Engineers, often trained to view their work as value-free calculation, bear responsibility for the values they embed.
1.2 Users, Publics, and Stakeholders
The user is the person interacting with an artefact; the public is every party touched by its existence. Pesticide users are farmers, but the public includes families drinking downstream water. Useful design conversations distinguish direct users, affected non-users, and ecosystems that cannot speak for themselves. Stakeholder mapping makes these layers visible and informs consultation, consent, and accountability.
Chapter 2: How People Think About Technology
2.1 Mental Models
Users build mental models of how systems work. Norman’s distinctions among designer model, system image, and user model clarify that poor usability often stems from mismatch between the system’s actual behaviour and what users believe it does. Feedback, affordances, constraints, and mappings in the system image shape user models; designers craft these signals deliberately.
2.2 Framing and Narratives
Societies tell stories about technology — progress, disruption, revolution, salvation, threat — that shape adoption, regulation, and expectation. A technology framed as “medicine” accesses different funding and public trust than one framed as “surveillance.” Engineers can influence framings by how they present work: what comparisons are used, what risks foregrounded, what futures imagined.
Chapter 3: Analysing Impacts
3.1 Impact Categories
Impacts of technology fall across environmental, social, economic, health, political, and cultural dimensions. An autonomous vehicle changes energy use, land demand for parking, employment of drivers, injury statistics, mobility access, insurance law, and urban form. Comprehensive analysis must span dimensions even when a project’s direct objective is narrow.
3.2 Evidence-Based Analysis
Evidence-based impact assessment draws on empirical data: peer-reviewed studies, government datasets, field observations, structured interviews. Sources differ in reliability, recency, and bias; triangulation across multiple sources compensates. Strong analyses quantify uncertainty, name gaps, and avoid false precision. Systematic reviews and meta-analyses provide higher-confidence inputs than single studies.
3.3 Life-Cycle and System Analysis
Life-cycle assessment (LCA) tracks environmental impacts from raw-material extraction through manufacturing, use, and end-of-life. Functional units and system boundaries shape conclusions: comparing an electric and a gasoline car per km driven over a 200,000 km lifetime differs from comparing them per unit manufactured. Systems engineering complements LCA with stakeholder analysis, social-impact assessment, and systems dynamics of adoption.
Chapter 4: Ethical Frameworks
4.1 Classical Frameworks
Three dominant frameworks guide engineering ethics discussions. Consequentialism judges actions by outcomes — maximise good, minimise harm; utilitarianism is its best-known form. Deontology judges actions by their adherence to duties and rights — some actions are forbidden regardless of outcome. Virtue ethics asks what a person of good character would do — focusing on dispositions (honesty, courage, prudence) rather than rules or outcomes.
In practice, engineers often combine frameworks: a consequence-heavy calculation checked against inviolable rights, subjected to a gut-check against professional virtue.
4.2 Contemporary Frameworks
Care ethics foregrounds relational responsibilities; capabilities approaches (Sen, Nussbaum) evaluate technologies by whether they expand the real freedoms people have to lead lives they value. Environmental ethics extends moral consideration to ecosystems and future generations. Indigenous ethics grounds responsibilities in relations among humans, non-human beings, and the land.
Chapter 5: Systems-of-Systems and Emerging Technologies
5.1 Systems-of-Systems Engineering
Contemporary challenges — climate, mobility, health systems — require coordinated design of many interacting systems. Systems-of-systems engineering provides methods for managing emergent behaviour, interoperability, and collaborative development of independently managed systems. Each constituent system has its own purpose and governance; the overall architecture must accommodate this distributed authority.
5.2 Emerging and Contested Technologies
Artificial intelligence, genome editing, synthetic biology, autonomous weapons, geoengineering, and surveillance technologies provoke public debate about impacts and permissibility. Responsible innovation frameworks (anticipate, reflect, engage, act) structure consideration of uncertain futures. Precautionary principles urge restraint where potential harms are serious or irreversible; proactionary principles counter that delay itself carries costs.
Chapter 6: Advocacy, Professional Engineering, and Design Practice
6.1 Needs Assessment as Ethical Practice
Good needs assessment is already ethical practice: it surfaces whose needs count, whose expertise informs framing, whose consent matters. Participatory design and co-design methods invite affected communities into the design process; they require engineers to share power, accept uncertainty, and commit to long-term relationships.
6.2 Advocacy
Engineers have voice — individually, through professional associations (PEO, OSPE, IEEE), and through employers. Advocacy can address policy (carbon pricing, safety regulations), organisational practice (worker safety, data governance), or public understanding (science communication). Ethical advocacy distinguishes technical assessment (what the evidence says) from value judgment (what should be done), signalling clearly when moving between.
6.3 Professional Engineering
Professional engineering licensure carries duties to the public, the profession, colleagues, and employer. The PEO Code of Ethics in Ontario articulates these obligations, and an engineer’s seal signifies personal accountability for a design. Duties to report, duty of care, and standards of practice protect the public from design decisions made in private. Increasingly, professional bodies extend these duties to cover equity, environmental stewardship, and truth in communications.
6.4 Designing with Awareness
Every design decision embeds choices: for whom it is optimised, which failure modes it prioritises, what life cycle it assumes, what future it makes more likely. A designer aware of these embedded choices:
- Names the worldviews shaping the framing.
- Identifies direct and indirect stakeholders.
- Uses evidence-based analyses spanning multiple impact dimensions.
- Applies ethical frameworks plurally rather than dogmatically.
- Invites affected communities into the process.
- Documents rationale to support future scrutiny and change.
Culture-of-design awareness does not make engineering less rigorous — it makes it honest. Graduates of a systems design programme who internalise these lessons enter professional practice equipped to see both the technical problem and the human one, and to design artefacts that earn public trust because they deserve it.