Analysis of scholarly documentation on pedagogy of criticism in architecture: A scoping review

Document Type : Review Article

Authors

1 PhD Candidate in Architecture, School of Architecture, College of Fine Arts, University of Tehran, Tehran, Iran

2 Associate Professor, Department of Architecture, School of Architecture, College of Fine Arts, University of Tehran, Tehran, Iran

3 Assistant Professor, Department of Architecture, School of Architecture, College of Fine Arts, University of Tehran, Tehran, Iran

Abstract

Architectural education continues to organize its studio culture around criticism, yet the field still lacks a shared narrative of what architectural criticism is supposed to accomplish, how it is taught, and how its outcomes are assessed in coherent and transparent ways. In both scholarship and curricula, critique is typically seen as a teaching mechanism for improving work in progress—pinups, roundtables, and juries as moments to guide decisions and refine proposals—rather than as a distinct academic competency that can be named, taught, and assessed. This orientation has reinforced a reliance on inherited traditions and individual teaching styles rather than on research-based models that specify learning outcomes, criteria, and methods. The result is a fragmented pedagogy that obscures the multiple functions of critique in higher education and limits the development of structured, cumulative approaches to teaching it. This study undertakes a scoping review of research conducted in the field of architectural criticism pedagogy to map the field, identify recurring gaps, and develop conceptual and pedagogical frameworks that now guide practice. This review does more than simply cataloguing sources. It reads the literature as evidence for the epistemological place of critique in academia: how the discourse of evaluation produces knowledge, structures participation, and is disseminated as a teachable practice in formal programs. In this sense, the project seeks to contribute to theoretical clarification and practical improvement of teaching strategies.
Following the five-step framework of Arce and O’Malley, this review adopts an applied orientation. Targeted searches in Web of Science, Scopus, and ScienceDirect using defined keyword strings yielded 494 records. After removing duplicates and two-level screening against the previous inclusion and exclusion criteria, 24 peer-reviewed studies that directly addressed criticism training were retained for analysis. Data were graphed in Excel and examined through descriptive keyword analysis, primary and secondary theme extraction, conceptual content analysis and thesaurus mapping with VOSviewer to visualize the relationships between terms and themes in the dataset. The findings are remarkably consistent. In most sources, critique is framed as a tool: as a means to improve the quality of design proposals, foster critical thinking, and support decision-making in studio courses. The emphasis is short-term and artifact-focused—what kinds of comments open up better alternatives, how to justify revisions, and where to place judgment in iterative design. In contrast, relatively few studies address critique as an independent, structured, and transferable capability. In those less common narratives, critique involves developing criteria, gathering and weighing evidence, articulating reasoned arguments, managing disagreement, and communicating clear judgments—skills that can be learned, practiced, and evaluated across contexts. This emphasis on critique for design has two predictable consequences. First, without an explicit narrative of criticism as systematic inquiry, students are rarely taught how intuition and taste translate into reasoned judgment. Education often prioritizes providing feedback over cultivating evaluative reasoning and ethical awareness. Second, the lack of stated outcomes and benchmarking guidelines undermines curriculum coherence and hinders comparisons across courses and institutions. Practices remain subject to local custom and individual preferences, limiting transferability and slowing cumulative improvement.
This synthesis suggests several paths for recalibration. Conceptually, theoretical-practical frameworks should present critique as an inquiry based on criteria, evidence, and argument, not as ad hoc interpretation. Pedagogically, programs can stage a clear learning progression—moving from description to analysis, evaluation, and, ultimately, to metacriticism. Early activities can foster careful observation and impartial description of design movements. Intermediate tasks can require analytical comparisons between precedents, alternatives, and constraints. Advanced tasks can require students to make and defend evaluative judgments, accept opposing arguments, and reflect on the norms and values that guide their reasoning. Assessment should align with this path. Rubrics can make criteria visible, separate the quality of reasons from the tone of the presentation, and differentiate design assessment from critical assessment. Short written “critic notes,” annotated design briefs, and short oral defenses can require students to frame claims, present evidence, and clarify reasons. Careful, measured peer review can amplify voices and foster accountability, while instructor feedback can model clear reasoning, responsible dissent, and humane behavior in the studio.
Technology in the reviewed literature appears promising but underdeveloped. Digital platforms can frame asynchronous critique, expand participation, and document how projects and arguments evolve over time; version histories make lines of argument traceable. Learning analytics can help educators regulate engagement and identify patterns of inclusion and silence. Immersive tools—virtual and augmented reality—support shared spatial discussion and situated judgment. AI can help with formative feedback, clustering of themes, and conversational analysis. Used intentionally and transparently, such tools can broaden reach and help align opinions with stated criteria. However, none of these technologies can replace a clear pedagogy or an inclusive studio culture. At the curriculum and faculty development levels, review becomes actionable advice. Programs should articulate learning outcomes for critique alongside design outcomes, map them into sequential courses, and create assignments that make reasoning visible. Formats that distribute authority—rotating juries, conversational critiques, structured rounds, or silent reviews—can moderate power dynamics and reduce unnecessary anxiety. Faculty development should focus on facilitation and questioning strategies, reducing bias, and designing inclusive critique environments that encourage participation, respect diverse forms of knowledge, and support constructive disagreement. Program evaluation should include measures that track the development of critical judgment over time, not just the quality of final artifacts.
The review also identifies research priorities that can generate cumulative insights. Longitudinal studies can examine whether explicit training in criticism leads to sustained gains in evaluative judgment, collaboration, and professional growth among graduates. Experimental and comparative designs can assess the effectiveness of different feedback formats and methods in supporting learning and inclusion. Cross-cultural research can examine how local traditions shape critique and how educational concepts are disseminated across languages and institutions. Ongoing bibliometric monitoring can track shifts in discourse—for example, whether terms like “critical pedagogy” are moving from the margins to the conceptual core as programs formalize education and share outcomes.

Keywords

Main Subjects


آن‌بیچر، مری (۱۴۰۰). طراحی نقد: گنجاندن نقد نوشتاری در آموزش طراحی‌داخلی (مترجم: اکبر دبستانی‌رفسنجانی)، اندیشنامه معماری داخلی، 1(1)، 29–36.  https://doi.org/10.30480/intand.2021.963
‏نامورمطلق، بهمن (۱۳۹۸). درآمدی بر بوطیقای نقد. پژوهشنامه فرهنگستان هنر، 16(2)، 9-26. https://pazhouheshnameh.ir/article-1-27-fa.pdf
 
El-Latif, M.A., Al-Hagla, K.S., & Hasan, A. (2020). Overview on the criticism process in architecture pedagogy. Alexandria Engineering Journal, 59(2), 753–762. https://doi.org/10.1016/j.aej.2020.01.019
Anthony, K.H. (1987). Private reactions to public criticism: students, faculty, and practicing architects state their views on design juries in architectural education. Journal of Architectural Education, 40(3), 2–11. https://doi.org/10.1080/10464883.1987.10758454
Arksey, H., & O’Malley, L. (2005). Scoping studies: towards a methodological framework. International Journal of Social Research Methodology, 8(1), 19–32. https://doi.org/10.1080/1364557032000119616
Attoe, W. (1975). Criticisms and architecture, and, Sometimes towns are toys. [Doctoral dissertation, Union Institute and University]. Ohio, United States.
Bartelse, G., du Preez, H., & Steyn, R. (2024). Exploring landscape architecture education: scoping review of innovations, challenges, and future directions. Journal of Digital Landscape Architecture, 9, 995–1002. https://doi.org/10.14627/537752094
Bassindale, J. (2020). Learning through successful feedback: digital opportunities for effective feedback in project-based architectural education. International Journal of Construction Education and Research, 16(2), 84–101. https://doi.org/10.1080/15578771.2020.1717681
Beecher, M.A. (2006). Designing criticism: integrating written criticism in interior design education. Journal of Interior Design, 31(3), 54–61. https://doi.org/10.1111/j.1939-1668.2006.tb00531.x
Betz, J.A. (2004). Peer review plus: a case for combining architectural design studios earlier. The 2004 American Society for Engineering Education and Annual Conference & Exposition, United States.
Bose, M., Pennypacker, E., & Yahner, T. (2006). Enhancing critical thinking through “independent design decision making” in the studio. Open House International, 31(3), 33–42. https://doi.org/10.1108/OHI-03-2006-B0005
Brookfield, S. (2013). Teaching for critical thinking: Tools and techniques to help students question their assumptions. John Wiley & Sons.
Daudt, H.M. Van Mossel, C., & Scott, S.J. (2013). Enhancing the scoping study methodology: a large, inter-professional team’s experience with Arksey and O’Malley’s framework. BMC Medical Research Methodology, 13(1), 48. https://doi.org/10.1186/1471-2288-13-48
Davis, A.T., & Preiser, W.F.E. (2012). Architectural criticism in practice: from affective to effective experience. Archnet-IJAR, 6(2), 24–42.
Dickson, A.W., & White, A.C. (1997). Interior design criticism: developing a culture of reverence for the interior environment. Journal of Interior Design, 23(1), 4–10. https://doi.org/10.1111/j.1939-1668.1997.tb00237.x
Donthu, N., Kumar, S., Mukherjee, D., Pandey, N., & Lim, W.M. (2021). How to conduct a bibliometric analysis: an overview and guidelines. Journal of Business Research, 133, 285–296. https://doi.org/10.1016/j.jbusres.2021.04.070
Eklund, A.R., Dixon, B., & Wegener, F. (2023). Pre-reflection-in-action: Rethinking Schön’s reflective practice through the “Habits of Design Artistry.” Design Issues, 39(4), 9-20. https://doi.org/10.1162/desi_a_00734
Fasli, M., & Hassanpour, B. (2017). Rotational critique system as a method of culture change in an architecture design studio: urban design studio as case study. Innovations in Education and Teaching International, 54(3), 194–205. https://doi.org/10.1080/14703297.2016.1174142
Geahigan, G. (2002). Art criticism: reflections on the evolution of an educational concept. Journal of Aesthetic Education, 36(2), 84–97. https://doi.org/10.2307/3333759
Goldschmidt, G., Hochman, H., & Dafni, I. (2010). The design studio “crit”: teacher-student communication. AI EDAM -Artificial Intelligence for Engineering Design Analysis and Manufacturing, 24(3), 285–302. https://doi.org/10.1017/S089006041000020X
Gul, C.G.G., & Afacan, Y. (2018). Analysing the effects of critique techniques on the success of interior architecture students. International Journal of Art & Design Education, 37(3), 469–479. https://doi.org/10.1111/jade.12145
Hassanpour, B. (2012). Evaluating critique sessions in architecture education [Doctoral dissertation, Universiti Kebangsaan Malaysia]. Malaysia.
Ibrahim, A.O., Sanad, E.M.M., Abdelhai, N.M.R., & Alsadun, I.S. (2022). Architectural theory as a tool for architectural criticism necessarily employed for the betterment of architectural education. International Journal of Advanced and Applied Sciences, 9(1), 71–83. https://doi.org/10.21833/ijaas.2022.01.009
Janniere, H. (2007). Architectural criticism: identifying an object of study. OASE Journal for Architecture 30(81), 34–54.
Jones, S.H. (1996). Crits - an examination. International Journal of Art and Design Education, 15(2), 132–141. https://doi.org/10.1111/j.1476-8070.1996.tb00660.x
Lymer, G. (2009). Demonstrating professional vision: the work of critique in architectural education. Mind Culture and Activity, 16(2), 145–171. https://doi.org/10.1080/10749030802590580
Lymer, G. (2010). The work of critique in architectural [Doctoral dissertation, University of Gothenburg]. Gothenburg.
Lymer, G., Ivarsson, J., & Lindwall, O. (2009). Contrasting the use of tools for presentation and critique: some cases from architectural education. International Journal of Computer-Supported Collaborative Learning, 4(4), 423–444. https://doi.org/10.1007/s11412-009-9073-9
Lymer, G., Lindwall, O., & Ivarsson, J. (2011). Space and discourse interleaved: intertextuality and interpretation in the education of architects. Social Semiotics, 21(2), 197–217. https://doi.org/10.1080/10350330.2011.548642
Megahed, N. (2018). Reflections on studio-based learning: assessment and critique. Journal of Engineering Design and Technology, 16(1), 63–80. https://doi.org/10.1108/JEDT-08-2017-0079
Moher, D. (2009). Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. Annals of Internal Medicine, 151(4), 264-270. https://doi.org/10.7326/0003-4819-151-4-200908180-00135
Murphy, K.M., Ivarsson, J., & Lymer, G. (2012). Embodied reasoning in architectural critique. Design Studies, 33(6), 530–556. https://doi.org/10.1016/j.destud.2012.06.005
Oh, Y., Ishizaki, S., Gross, M.D., & Yi-Luen Do, E. (2013). A theoretical framework of design critiquing in architecture studios. Design Studies, 34(3), 302–325. https://doi.org/10.1016/j.destud.2012.08.004
Peters, M.D.J., Godfrey, C.M., Khalil, H., McInerney, P., Parker, D., & Soares, C.B. (2015). Guidance for conducting systematic scoping reviews. JBI Evidence Implementation, 13(3), 141–146. https://doi.org/10.1097/xeb.0000000000000050
Scagnetti, G. (2017). A dialogical model for studio critiques in design education. The Design Journal, 20(sup1), S781–S791. https://doi.org/10.1080/14606925.2017.1353024
Seymour, M.W., & Chance, S. (2011). Assessment formats: student preferences and perceptions. International Journal of Learning, 17(10), 137–154. https://doi.org/10.21427/d7bw4s
Tricco, A.C., Lillie, E., Zarin, W., O’Brien, K.K., Colquhoun, H., Levac, D., & Straus, S.E. (2018). PRISMA extension for scoping reviews (PRISMA-ScR): checklist and explanation. Annals of Internal Medicine, 169(7), 1–28. https://doi.org/10.7326/M18-0850
Utaberta, N., Hassanpour, B., & Usman, I. (2010). Redefining critique methods as an assessment tools in architecture design studio. WSEAS Transactions on Advances in Engineering Education, 7(4), 359–364.
Van Eck, N.J., & Waltman, L. (2023). VOSviewer manual. Leiden: Centre for Science and Technology Studies, Leiden University.
Zimring, C., Khan, S., Craig, D., Haq, S., & Guzdial, M. (2001). CoOL studio: using simple tools to expand the discursive space of the design studio. Automation in Construction, 10(6), 675–685. PII: S0926-5805(00)00092-3