Contents

Cover

Half Title page

Copyright

Title page

Foreword: Architecture & Intelligence: Inside Smartgeometry

Brett Steele

Introduction

Brady Peters Terri Peters

The Origins of Smartgeometry

Hugh Whitehead, Lars Hesselgren J Parrish

Hugh Whitehead

Lars Hesselgren

J Parrish

First Build Your Tools

Robert Aish

Parametric Evolution

Branko Kolarevic

Matrix Architecture

Jenny E Sabin

Metrics of Human Experience

Roly Hudson and Michael Westlake

Interacting with the Model

Neil Katz, with Bess Krietemeyer and Tobias Schwinn

Responsive Design: Towards an Ecology of Objects and People

Andre Chaszar, Anton Savov, Peter Liebsch and Sascha Bohnenberger

Design Flow and Tool Flux

Robert Woodbury

The Sound of Smartgeometry

Brady Peters

Design Exploration and Steering of Design

Axel Kilian

Geometry: How Smart Do You Have to Be?

Chris Williams

Generative Components and Smartgeometry: Situated Software Development

Volker Mueller and Makai Smith

From Descriptive Geometry to Smartgeometry: First Steps Towards Digital Architecture

Mark Burry

Exploring Human–Computer Interaction in Design Process

Flora Salim and Przemek Jaworski

Designing Intelligence: Diy Robotics for Responsive Environments

Nick Puckett

Marco Poletto and Claudia Pasquero (Ecologicstudio)

Mariana Ibañez and Simon Kim

Encoding Design

Fabian Scheurer

Working Prototypes, Creating Knowledge

Shane Burger and Xavier De Kestelier

Mind the Gap: Stories of Exchange

Case

Designing Robotic Assemblies

Tobias Bonwetsch

The Practice of Smartgeometry

Hugh Whitehead

Digital Crafting: Performative Thinking for Material Design

Mette Ramsgard Thomsen and Martin Tamke

Design Robotics: New Strategies for Material System Research

Martin Bechthold

Contributor Biographies

INSIDE SMARTGEOMETRY

FOREWORD

ARCHITECTURE & INTELLIGENCE: INSIDE SMARTGEOMETRY

BRETT STEELE

‘Everyone designs who devises courses of action aimed at changing existing situations into preferred ones.’

‘What information consumes is rather obvious: it consumes the attention of its recipients. Hence a world of information creates a poverty of attention.’ Herbert Simon, Sciences of the Artificial

The Nobel Prize-winning polymath Herbert Simon wrote the above observations more than half a century ago at the time the world’s first CAD systems arrived. His freakishly prescient declaration of the extreme consequences of modern, cognitive models captures both the potential as well as the challenge of a) design worlds conceived (enthusiastically) in terms of information-based problem solving; and b) the unexpected consequences (somewhat more tentatively) to time, and not only space, in an era dominated by the production of information.

Long after Simon’s foresight we all know now that the construction of attention is amongst the most difficult of all architectural undertakings, in a world of relentless media, information and access. Surely one of many remarkable accomplishments of Smartgeometry is its least obvious feature: its genuinely sustained, focused attention to a bounded set of architectural, geometric and linked questions, from which immense knowledge has been advanced. Like other computational cultures and not only digital design technologies, Smartgeometry is now a truly global enterprise; a regularly convened, creative space of like-minded architects and others interested in an open experimentation, exploration and invention of new design systems alongside the projects for which these associative systems have been developed.

Conceived a little more than a decade ago by Hugh Whitehead, Lars Hesselgren and J Parrish as a cross-disciplinary series of design workshops, presentations and discussions, and soon joined by Robert Aish and other key collaborators giving from its earliest days both breadth and depth (in an architectural world too frequently describable by means of a surface), Smartgeometry has by now come of age. This fact is the first of many demonstrated in the wonderful volume that follows; a book that is both a document and proposition, not only for recent (rapidly evolving) concepts, tools and techniques, but also for future ones. This is an interest Smartgeometry’s founders share with a cast of thousands that by now have participated (myself included) as presenters, workshop leaders, design gurus, students, teachers and curious observers at the group’s annual gatherings.

The following volume provides an in-depth record of these gatherings, and has been wonderfully edited by Brady Peters and Terri Peters as two dozen chapters containing insightful accounts of both the history of these events and highlights from their various undertakings. What follows is part post-post-modern super-nerd primer (filled with fascinating weird things like ‘augmented composites’, ‘shape grammars’, ‘swarm algorithms’, ‘parametric acoustic surfaces’, or ‘particle spring solvers’, amongst many other things), while also being a remarkably accessible, straightforward demonstration of how contemporary design tools are at work in unexpected ways. They are reconfiguring the concepts emerging alongside the various forms of machinic, digital and physical modelling, prototyping and testing that provide the more visible outcome of Smartgeometry’s remarkably robust, sustained attention. It’s an approach that, like the impetus of contemporary ‘object-oriented’ programming cultures, crucially prefigures as well as continues to inform the generative model-making activities surrounding Smartgeometry. What we see above all else in this book is how contemporary experimentation wrestles ultimately with the most complex architectural projects of all – the cognitive architecture of the architect’s own mind.

 

1 Sean Ahlquist, Bum Suk Ko and Achim Menges, Material Equilibria: Variegated surface structures, ggggallery, Copenhagen, 2012.
Material Equilibria is a part of a larger body of research done by Sean Ahlquist, a professor at the University of Michigan and a tutor for the 2010 Smartgeometry workshop cluster ‘Deep Surfaces’. The project focuses on the development of computational design methodologies and techniques which enable the generation of self-structured spatial forms through the generation of informed material behaviours.

2 Sean Ahlquist, Bum Suk Ko and Achim Menges, Material Equilibria: Variegated surface structures, ggggallery, Copenhagen, 2012.
The parametric model shown here controls the knit of shifting patterns and densities, influencing the structure of the tensile spatial surface. Through computation, the micro-structure of the textile is varied to create particular organisational and structural behaviours. The accumulated material phenomena are calibrated to derive an equilibrium which works with the resistance of an actively bended glass-fibre structure at the boundary.

 

What follows is far more than a demonstration of just how far and fast contemporary experimentation has pushed architecture’s millennial reliance on geometry and advanced mathematics. Writing half a century ago Herbert Simon understood the figure of the modern architect as a perfect demonstration of what it means to design today. ‘The modern architect’, Simon wrote, ‘is the maker of instructions’. The architect’s job isn’t the ‘making’ of things in a conventional sense – it is the recording of design intentions, ideas and ambitions in the form of documents (drawings, sketches, models etc.) whose real purpose is to tell others what to do. Architects, in Simon’s view, ‘make instructions’. In this sense, the artefacts produced in the architect’s studio are hardly anything other than memory structures. In this sense, we can grasp the real values of Smartgeometry’s remarkable, collective, collaborative and sustained focus: not only on the role of information-based approaches to architectural design, but also on the making of the most difficult of all architectures, the architect’s own cognitive structure. It’s a memory structure of a very different sort than most kinds of architecture, and in the hands of the protagonists whose examples follow it is proven no less elastic, and intelligent, than other forms of building.

TEXT

© 2013 John Wiley & Sons, Ltd

IMAGES

© Sean Ahlquist, Achim Menges, Institute for Computational Design, University of Stuttgart

INTRODUCTION

BRADY PETERS TERRI PETERS

Smartgeometry (SG) was founded in 2001 as an informal network of designers interested in harnessing the powers of computation for architectural design. Friends and former colleagues Hugh Whitehead, Lars Hesselgren and J Parrish felt frustrated by the lack of resources and network surrounding computation and architecture and sought to redefine ways that architects could use digital tools. At first, the trio of architects drew on their network of friends and collaborators such as computer scientist Robert Aish, academics Robert Woodbury and Axel Kilian, and experimental practitioners architect Mark Burry and engineer Chris Williams to put together a few modest conferences and workshops. These began with a lecture and workshop in 2003 in Cambridge, UK, then in 2004 at the University of Waterloo, Ontario, Canada, where the focus was on software development, new tools for architects and engaging with ideas outside the boundaries of ‘architecture’. These early workshops provided inspiration and a testing ground for the creation of new parametric software GenerativeComponents (GC) that was introduced to the group by Robert Aish and Bentley Systems. Rather than being concerned solely with software or form-making, SG focuses on the creation and application of digital tools and technologies, and in cross-disciplinary fertilisation of emerging ideas in practice. In workshop groups, designers are able to work on projects ‘off the books’, away from their offices or university settings, creating pure explorations of technique beyond the confines of the design project. SG embodied new ideas and new ways of thinking. The event now spans six days, with a four-day curated workshop and two-day public conference, and attracts more than 300 international participants and attendees each year.

 

1 Design explorations at SG 2006,
Cambridge, UK.
Conceptual geometric design explorations using GenerativeComponents.

 

WHY GEOMETRY?

Architectural design software at the time SG was founded was created by software developers using object-oriented programming that almost literally translated software ‘objects’ as building ‘objects’. SG co-founder Lars Hesselgren has written that they wanted to build new design tools and founded SG as a rejection of these conservative influences that promoted computer-aided design (CAD) solely as the organisation of building components.1 In order to be free of these predefined tools and have a higher-level discussion of building form in terms of first principles, this led to a discussion of geometry and mathematics. As this is a more generic approach, thinking of architecture and form in this way allowed them to share computational tools between disciplines. It allowed architects to design conceptually and create their own custom ‘objects’ rather than use the specified objects provided by their CAD software.

As Robert Aish explains in his chapter, SG explores the ideas of design computation, with the notion that there is a distinction between the generative description of the building, and the resulting generated model of the building. Therefore SG is more about the exploration of design intent and how this is inscribed in the design tools and the design environment, rather than specific technology for the integrated delivery of building projects. It is about designing a system, rather than working on a more detailed 3D model.

 

2 ‘Responsive Acoustic Surfaces’ workshop cluster at SG 2011, Copenhagen, Denmark.
Participants engage in a design discussion.

Digital models from the ‘Responsive Acoustic Surfaces’ workshop cluster at SG 2011, Copenhagen, Denmark.
Participants work on digital models of hyperboloid geometry using a variety of software.
‘Use the Force’ workshop cluster at SG 2011, Copenhagen, Denmark.
Participants discuss design and computation.

 

SG is an agile network. It is purposely structured to be able to react to and reflect ideas in contemporary practice; there is no overriding goal or charter. The idea is to engage with current issues and debates in a collaborative and non-competitive environment. Digital design leads logically to digital fabrication. Over several years, but culminating in 2010 where it was a central feature, the event embraced digital fabrication, interaction and simulation with ‘workshops’ that more equally split experimentation in digital and physical realms. As Xavier De Kestelier and Shane Burger explain in their chapter, the evolving workshop structure is due to shifts in participants and leadership. The earlier events attracted lower numbers of workshop participants and leaders and these were almost exclusively from professional practice. Recent SG events have had multi-day programs with larger audiences and an increased focus on academic and research questions. This shift is discussed in the chapter by CASE, where they identify the move away from the pragmatics of designing for construction of buildings, towards workshops based not only on research and experimentation, which does not necessarily rule out the practical building issues, but also on creative explorations using these same methods. The five current SG Directors are all from architectural practice, but each year attendees and workshop leaders come increasingly from research and academia.

TALKING ABOUT COMPUTATION

‘The most profound technologies are those that disappear. They weave themselves into the fabric of everyday life until they are indistinguishable from them.’2 This statement from computational pioneer Mark Weiser in 1991 is relevant to architectural practice today. Computation is everywhere; should it really be the medium and not the message? Architects desperately need to talk about computation, and over the past decade SG has provided the only experimental workshop-based discussion forum on this topic. It is not enough to say computation is ubiquitous in our field; it is not ‘just’ a tool – there can be no doubt that it is fundamentally changing architecture. Computation is not what architecture is, but if architecture can be understood as a practice, concerned with technique, then computation is a technique intricately connected to designing for meaning and experience in architecture. Even architecture as edifice, separated from any discussion of technique, reveals the tool of the maker. While meaning in architecture can come from symbols and symbolism in the building itself, it also comes from the experience of that building.3 Therefore the better we can simulate the experience of architecture, the better we can design for it. The technologies explored and discussed at SG are still quite visible. However, one hopes they will be customised, tested and engaged with by architects even as they continue to be woven into the fabric of the design environment.

 

3 Prototypes and analysis from the ‘Responsive Acoustic Surfaces’ workshop cluster at SG 2011, Copenhagen, Denmark.
1:10 scale prototypes were tested for their acoustic performance and this data was used to inform the design of the full-scale prototype wall.

4 Workshops at SG 2011, Copenhagen, Denmark.
View of participants and workshop clusters.

 

Of course, the mere fact of using a particular computational technique does not guarantee good architecture, the same way that using the same pen as Norman Foster will not guarantee a great building. But why shouldn’t architects share techniques and tools? It would be pretty silly if architects each had to invent our own pens, drafting boards and drawing conventions. So while it is the building that matters most, rather than focusing on the process of design and making, in the context of design it is critical to acknowledge that design processes are changing, and SG is at the forefront of this change.

CREATING KNOWLEDGE AND TECHNIQUE

The ways that computation and architectural design are explored at SG are unique. The idea is to nourish a collaborative environment where participants feel as though anything can happen. The theme of the event is set in advance and then workshop leaders apply to lead a cluster based on their own research: for example, in 2012, 40 detailed applications were submitted for 10 positions. Participants also apply to join workshops, submitting portfolios and statements of interest, with only 100 selected. Carefully curated by the SG community and directors, the selected workshop themes are developed and in the four days, focused questions of design, digital technique and physical making can be explored. In contrast to many other design workshops with traditional student–teacher dynamics, workshop leaders do not bring work they have done earlier to get ‘students’ to build, participants do not come to learn something they know nothing about, and experts do not arrive with ‘answers’ to disseminate. Research and knowledge is created during the workshops. The challenge is not to construct a research question that can be ‘answered’ in four days but rather to construct a line of thinking that can be investigated intelligently and discussed through experimentation. The SG environment is part research and part professionally focused, which seems to inspire productivity, as participants work long into the night to actually do something as a group within the given time, to produce some results to share with the wider group by the end of the workshop, and to work together. In Robert Woodbury’s chapter, he calls this the ‘flow’ of design computation. There is of course a healthy fear of failure and underlying pressure to make it perfect, or at least beautiful. This is architecture after all.

SG makes no claims to produce ‘architecture’. It is not about form, it is about how we arrive at form. SG is about technique. There are, of course, many valid ways to design and SG celebrates this plurality of concept. It is not the place for design crits. In the four-day workshops, there simply is not time. Instead, techniques and tools are developed and tested. Participants find where a tool hits the wall, then how to mash it up with other tools and make it work better. It is like building a racing car – how fast and how hard can we push this machine – not how nicely can we drive it.

 

5 Aggregate module structure from the ‘Agent Construction’ workshop cluster at SG 2011, Copenhagen, Denmark.
The participants operate as ‘agents’, building and altering the structure without pre-made drawings or plans, instead being guided by rules responding to local conditions – ‘here and now’.

6 Virtual agent model from the ‘Agent Construction’ workshop cluster at SG 2011, Copenhagen, Denmark.
In the computer model, a swarm of virtual agents gradually and collectively build up a structure. Different agents are guided by different rule-sets, and the only communication between them is through the environment which they manipulate and which in turn affects their behaviour.

7 Visualisation of scan data from the ‘Agent Construction’ workshop cluster at SG 2011, Copenhagen, Denmark.
The emerging physical structure is continuously scanned, and the data imported into virtual formats for analysis and further processing. In this format it can be directly analysed beside the virtual agent models, or used as an input in these.

8 Airflow simulation from the ‘Agent Construction’ workshop cluster at SG 2011, Copenhagen, Denmark.
The scan data allows for simulation of the structure’s performance, here through a fluid dynamic simulation in X-Flow.

 

CODING

‘The ability to ‘read’ a medium means you can access materials and tools generated by others. The ability to ‘write’ in a medium means you can generate materials and tools for others. You must have both to be literate. In print writing, the tools you generate are rhetorical; they demonstrate and convince. In computer writing, the tools you generate are processes; they simulate and decide.’ Alan Kay4

When a designer writes a script to solve a problem, the algorithm becomes part of the design and may then be explored in a creative way. But, as Fabian Scheurer explains in his chapter, algorithms are both a description of the problem and the solution. They define the solution space and they are built around the definition of the problem. He argues that design is all about decisions and that delegating these to an algorithm always means following predefined paths. Often the use of existing tools leads to existing solutions. Through the creation of new tools, new ways of thinking and new solutions can be found. Algorithmic thinking means taking on an interpretive role to understand the results of the generating code, and understand how to then modify the code to explore new options, and to speculate on further design potentials. As designers, we are influenced by the tools and techniques that allow us to realise our visions. It has been said that the tools determine the boundaries of art, and that it is the use of the right tools for the thing that one is making, and a deep relationship between the use of the tool and its formal results, that establishes the potentials of what can be made.5 With computation, the boundaries of what can be made just got a lot bigger. Parametric systems and computational tools have enabled the realisation of projects that were previously inconceivable.

Nicholas Negroponte introduced the concept of bits and atoms, arguing that atoms make up physical, tangible objects around us – the architecture that we inhabit – while our design environment and our digital models inhabit the space of the bits – the information that is contained within the computer that we use to design.6 So, how does this relationship affect the architecture we design? This relationship between bits and atoms is becoming blurred. Not only do the experiments undertaken at SG work in between physical and digital realms, but design tools are increasingly used that simulate real-world performance and provide feedback on designs. Computational tools become co-creators in design, extending the intellect of the designer, and so the role of the designer becomes one of tool builder, of interpreter of results, and of a guide through solution spaces. In his chapter, practitioner Neil Katz explains that the technology needs to disappear: it is the design intent and process that is more important than the tool itself.

‘Software modified by the designer through scripting, however, provides a range of possibilities for creative speculation that is simply not possible using the software only as the manufacturers intended it to be used. Because scripting is effectively a computing program overlay, the tool user (designer) becomes the new toolmaker (software engineer).’ Mark Burry7

 

9 Geodesic curves from the ‘Gridshell Digital Tectonics’ workshop cluster at SG 2012, Rensselaer Polytechnic Institute, Troy, New York, USA.
Geodesic curves allow for complex curvature from straight segments.

10 Fabrication layout of laths from the ‘Gridshell Digital Tectonics’ workshop cluster at SG 2012, Rensselaer Polytechnic Institute, Troy, New York, USA.
Geodesic laths are unrolled with precise lengths and spacing of nodes for pin joints.

11 Elevation from the ‘Gridshell Digital Tectonics’ workshop cluster at SG 2012, Rensselaer Polytechnic Institute, Troy, New York, USA.
Diagrams of the lattice gridshell identified the location of each piece in the four-layer lath system.

 

SIMULATING EXPERIENCE

Architecture can be thought of as drawing, but should be thought of as simulation. Architecture is the act of imagining a building at a remove from its construction, and then communicating this concept for others to build. To date, the imagining and communicating has been largely through drawing. However, it is not necessarily drawing that defines architecture, but this ability to create an abstraction of the building through some means. Through the drawing, the architect is able to imagine how light and space and material relate in the creation of architecture. Although largely within the mind of the architect, this simulation of effect and experience is a necessary part of architectural design. The pragmatic aspects of performance can be simulated as well. The digital design environment can be a design partner for this simulation of architecture. Through the adoption of new technologies, the creation of design techniques, the coding of custom design tools and the gaining of critical performance feedback, the abilities of the architect are extended.

SG was founded on the premise that a first-principles exploration of geometry in relation to design intent could benefit architectural design. The development, discussion and dissemination of these explorations of technique have been central to the SG workshops and conferences. The SG community explores these through parametric design, computer programming, digital fabrication, interactive design, simulation and optimisation. The scope of these approaches is enlarged at each yearly event. SG has been, and continues to be, a place where these concepts are not only discussed, tested and critically reflected upon, but critically, a place where this knowledge is created. A place for designing, coding and building.

 

12 Curvature analysis from the ‘Gridshell Digital Tectonics’ workshop cluster at SG 2012, Rensselaer Polytechnic Institute, Troy, New York, USA.
Curvature analysis determines radii of curvature to verify the minimum radius allowable for the bending of lath segments.

13 Completed gridshell from the ‘Gridshell Digital Tectonics’ workshop cluster at SG 2012, Rensselaer Polytechnic Institute, Troy, New York, USA.

 

REFERENCES

1 Lars Hesselgren, ‘Smooth Is Not Enough’, 18 January 2009, http://www.core.form-ula.com/2009/01/18/smooth-is-not-enough/ [accessed 12/09/2012].

2 Mark Weiser, ‘The Computer for the 21st Century’, in Scientific American, September 1991, pp 94–104.

3 Stan Allen, ‘Introduction: Practice vs Project’, in Practice: Architecture, Technique and Representation, G+B Arts International (Amsterdam), 2000, page XIV.

4 Alan Kay, ‘User Interface: A Personal View’, in Brenda Laurel (ed), The Art of Human–Computer Interface Design, Addison-Wesley (Reading), 1990, p 125.

5 Jose-Manuel Berenguer, ‘Interview with Jose Manuel Berenguer’ in Mosaic, 2005, http://mosaic.uoc.edu/2005/09/09/jose-manuel-berenguer/ [accessed 10/02/2010].

6 Nicholas Negroponte, Being Digital, Vintage Books (New York), 1995.

7 Mark Burry, Scripting Cultures, John Wiley & Sons (Chichester), 2011, p 9.

TEXT

© 2013 John Wiley & Sons, Ltd

IMAGES

Figure 1 © Smartgeometry Ltd; Figures 2, 3, 4, 5 © Smartgeometry Ltd, photos by Anders Ingvartsen; Figure 6 © Peter Suen; Figures 7, 8 Samuel Wilkinson; Figures 10, 11, 12 © Andrew Kudless with the Gridshell Digital Tectonics Cluster; Figures 9, 13 © Mark Cabrinha, PhD

THE ORIGINS OF SMARTGEOMETRY

HUGH WHITEHEAD, LARS HESSELGREN J PARRISH

Smartgeometry (SG) was founded in 2001 by former colleagues and friends Hugh Whitehead, Lars Hesselgren and J Parrish as a way to recapture parametric and computational design for architecture. At the time of founding SG they were leaders in the London-based architectural practices Foster + Partners, Kohn Pedersen Fox (KPF) and ArupSport respectively, and were strong proponents of digital design. Each was striving to create architecture through the use of parametric tools and computational methods. Through SG they hoped not only to create new digital tools, but to foster a community that would develop, test and disseminate these ideas of architecture and design to a wider audience. Whitehead founded the Specialist Modelling Group at Foster + Partners in 1998 which has been responsible for a host of innovative buildings and consistently pioneers computational design methods in architectural practice. After years at KPF in London, architect Lars Hesselgren is now the Director of the Computational Design Research Group at PLP Architects. J Parrish is a globally renowned sports stadium designer leading teams to design some of the most iconic stadium projects in the world. After many years at ArupSport, he moved to AECOM where he is currently working on the venues for the Rio Olympic Park. Here each tells their story of the origins of SG, now an international multidisciplinary community of professionals, academics and students in the fields of architecture and engineering.

 

1 Hugh Whitehead, Lars Hesselgren and J Parrish, the original SG founders, at SG 2011, Copenhagen, Denmark.

 

HUGH WHITEHEAD

BACK TO THE FUTURE

Four people were sitting in a car travelling to a Bentley conference. Robert Aish, in the front, was due to host a research seminar; and architects Lars Hesselgren, J Parrish and Hugh Whitehead, in the back, were enjoying the opportunity to tease a captive software developer. It was a familiar formation: we had all worked together with YRM back in the 1980s. So we began the light-hearted banter with a searching question: ‘Why is it that ten years have passed, and we still cannot even get close to the kind of capability that we had then?’

At the end of the 1980s boom, YRM had grown to an international multidisciplinary design consultancy of 600 people, and took the strategic opportunity to acquire Anthony Hunt Associates, the engineering firm of choice for Norman Foster, Richard Rogers, Nicholas Grimshaw and many other leading architects. What Anthony Hunt Associates needed was access to computer modelling expertise, which was already well advanced at YRM. The dialogue between architect and engineer was shifting rapidly from back-of-envelope sketches to digital representations, where 3D geometry became the input to analysis routines and setup cycles of design iterations.

 

2 Foster + Partners (architects), Anthony Hunt Associates (engineers), Faculty of Law, University of Cambridge, UK, 1995.
The geometry of the diagonal panels and offset supporting structure was formed by proportional subdivision of a cylindrical vault. A parametric model was developed by Hugh Whitehead so that changes to the radius of the vault caused the geometry to regenerate the data needed for structural analysis.

 

The acquisition of Anthony Hunt Associates brought exposure to a whole new world of adventurous designers, who were expecting us to provide them with new design technology. Where would we find it? Engineers and product designers always seemed to have far better tools than architects, and we realised that we were looking for something that was generic rather than discipline-specific.

Around this time we saw a presentation by Robert Patience who led the development of the new Intergraph Vehicle Design System (I/VDS). It was a revelation. That rare kind of presentation that seems to come from another time or another place and brings you out in a cold sweat! There, back in the 1980s, we saw a first glimpse of the power of parametrics, associative geometry and relationship modelling, all in full 3D, at a time when leading computer-aided design (CAD) systems of the day were still only trying to mimic and crudely automate flat drawing-board technology. Robert Patience ended his presentation with the throwaway line, ‘Last weekend I did HVAC [heating, ventilation and air conditioning], with automated duct sizing and routing just from a rule-based schematic, all in full 3D with clash detection!’

We invited Robert Patience to visit YRM to discuss the potential for developing his ideas in an architectural context. He brought with him Robert Aish, who was working with him in Paris, helping to implement the new technology for the Gdansk shipyard, where the aim was to directly flame-cut steel from a rule-based 3D design model. Design-to-fabrication was already happening.

The show-and-tell session lasted far into the night, while we explained the design challenges we were facing and the two Roberts talked about the potential of associative systems. At the end we asked, ‘Why label the product as a Vehicle Design System (VDS), when it clearly has the potential to provide generic solutions which could support a far more integrated approach to design?’ Robert Patience replied, ‘I always think in generic terms, but as a software developer I can only get funding from the Marketing Department by pretending to be discipline-specific, so I chose vehicles because at least they include cars, ships and aircraft. All have structure, services, form, space and aesthetic requirements, just like buildings! Perhaps we could describe buildings as very slow-moving vehicles, almost tending to the limit!’ At this moment an idea was born, and we convinced Intergraph to develop an architectural application based on VDS technology. The result was a specification for a product called ‘Master Architect’. Robert Aish joined us at YRM to help develop the brief and explore concepts based on the challenges of live projects. With Robert’s help, Lars Hesselgren produced a fully associative 3D model of London’s Waterloo Station for Grimshaw while Hugh did a similar exercise on the University of Cambridge Faculty of Law for Foster + Partners. We all worked with J Parrish on a modular concept design called ‘Stadium for the ‘90s’. The stadium roof was a tensile membrane structure supported on cantilever beams with a retractable centre section.

 

3 Antoni Gaudí, Sagrada Família Basilica, Barcelona, Spain, 1883–, central crossing of the nave.
The progression from constructive geometry to parametrics and then to scripting and computational design was already mapped out by designers like Gaudí, who worked only with models and raw intellect.

4 Foster + Partners (architects), Buro Happold and Waagner-Biro (engineers), Great Court at the British Museum, London, UK, 2000.
The elegant resolution of complex aesthetic and structural requirements is reflected in the elegance of the mathematics derived by Chris Williams. Three functions describe the transformation from rectangle to circular boundary, maintaining singularity of curvature. The triangulated pattern floats on this surface using dynamic relaxation to achieve continuity of geodesic curvature.

5 Digital model by Mark Burry of the Sagrada Família Basilica, Barcelona, Spain, 1995.
The work of architect Mark Burry helped to decode Gaudí in terms of a language of intersecting helicoid, hyperbolic paraboloid and hyperboloid surfaces. This enabled contemporary design technology to engage with traditional craftsmanship and so to realise a vision that had never been fully described.

 

So what happened in that decade between the introduction of VDS to YRM and the car ride to Exton? How was parametric design lost to architecture in those 10 years? The1980s bubble burst: YRM went down, Intergraph went down, Lars moved to KPF, J moved to ArupSport, Hugh to Foster + Partners, and Robert to Bentley Systems, but the friendship and the shared experience remained. In the car that day, the response from Robert was this: ‘Sometimes I feel as though I have been to the future. I have seen it and I know that it works!’

But the question was, how could we get back to the future?

We held a conference in Cambridge, UK in 2003. The event attracted strong interest with many presentations. Two were particularly inspirational. Mark Burry described 15 years of decoding the designs of Antoni Gaudí (1852–1926) which enabled the completion of the Sagrada Família in Barcelona, and Chris Williams explained the generation of the geometry for the Great Court roof at the British Museum in London. Here were two people who had already delivered the kind of projects that we aspired to. They both combined a background in architecture and engineering with fluency in mathematics and scripting. This expertise was used to give expression to design ideas by developing custom workflows, which engaged a variety of applications. Mark described how he used Excel as a kind of blind CAD system to process data before exporting to graphics. Chris gave a live demonstration in which he showed how to ‘sketch with code’.

We were delighted when Mark and Chris agreed to join us as tutors at the next SG workshop at the University of Waterloo, Ontario, Canada in 2004. With the addition of Axel Kilian from the Massachusetts Institute of Technology (MIT) and Robert Woodbury from Simon Fraser University (SFU), British Columbia, we had an international all-star team. The significance would only appear in retrospect as the community reached critical mass and gained momentum. So we approached the first workshop as a ‘learn by doing’ experiment, not just in design technology but also in design sociology, and this spirit continues.

If the future lay in integrated design then we needed a comprehensive platform that would support disparate activities and promote collaboration between different specialists. To give the workshop an experimental focus, we then invited Robert Aish to join and asked Bentley Systems if they would agree to let us trial their new object-based system, GenerativeComponents (GC). Never before had a group of designers experienced a raw development platform in the early stages of specification while the author of the software looked over their shoulders. It was a rough and bumpy ride, but a rare privilege and everyone was captivated by Robert’s tireless enthusiasm. Often he would work through the night to address issues that arose and emerge in the morning with a ‘crisis build’ which he distributed on a memory stick.

 

6 The SG 2004 workshop briefing at the University of Waterloo, Ontario, Canada.
Tutors gave a brief description of their background and special interests. Participants then formed groups around their chosen tutor to discuss their projects before moving off to rooms to begin the workshop. We became a self-organising community.

7 Design explorations at SG 2004, University of Waterloo, Ontario, Canada.

 

Almost by accident, SG became a design process laboratory with a unique characteristic. It became a community of people drawn from competing organisations, but the bonds that were formed crossed all known boundaries. Instinctively the tutors led from the front and the participants followed their example and competed to share knowledge, ideas and experience, and even tools, techniques and concepts. This special synergy was well appreciated by Bentley, who gave Robert the opportunity to pursue his vision. They have continued to support SG in what became a series of annual events that have spanned the last 10 years. But those early workshops were all high-risk ventures with an edgy quality, and so we would treat each one as if it might be the last.

Not everything went according to plan and ‘the great pizza disaster’ has now passed into SG folklore. We had no experience of event management on this scale and there was so much to organise – tutors, participants, admission, registration, venues, accommodation, travel, facilities, equipment etc. We forgot about food! At the end of a totally exhausting first day we suddenly realised that we had 40 starving people out on a university campus miles away from civilisation. If everyone just left in search of sustenance the event would lose all momentum and we would lose their goodwill. At an emergency meeting the SG Directors agreed to put their hands in their pockets and send out for a pizza delivery, but could we afford a few beers? Fortunately we could, because the pizza was late and cold, but at the sound of popping cans the spirits revived and the intense discussions flowed again.

When we related the story of this near-disaster to Greg Bentley the next day, he kindly offered to pick up the tab and bail us out, but added a stern warning that we had to become more businesslike because ‘altruism is not commercially viable’. However, it could be said that Bentley have spent the last 10 years proving that with the right community it can be. During this period the SG workshops helped to specify and test four different versions of GC, which demanded extreme patience and perseverance from all concerned. The concepts that were being prototyped proved to require a new type of development language with advanced capabilities. Perhaps, if the SG Directors had not sent out for the pizza, GC might never have happened!

 

8 Architect: Grimshaw (architects), YRM Anthony Hunt Associates (engineers), International Terminal, Waterloo Station, London, UK, 1993.

9 Parametric model by Lars Hesselgren of the International Terminal, Waterloo Station, London, UK, 1980s.
The original parametric model of Waterloo Station by Lars Hesselgren while at YRM Anthony Hunt Associates for Grimshaw using I/EMS mechanical modelling software.

10 Parametric model by Lars Hesselgren of the International Terminal, Waterloo Station, London, UK, 1980s.
Rebuilt parametric model of the Waterloo Station arches, architects Grimshaw, structural engineers YRM Anthony Hunt Associates. The original model used Intergraph I/EMS software, the rebuild is done with Bentley GC software.

 

Some participants look back on those early workshops as a life-changing experience, which launched a whole new career – and it is comments like these that provide our motivation. Now, with the addition of a ShopTalk day and a public conference, SG provides not just a forum but also a stage, which many people regard as a window on the future. The original founders can only look back and marvel at the culture that grew from a cold pizza in Cambridge!

LARS HESSELGREN

APPLIED PARAMETRIC DESIGN

In 1982 I was a fully-fledged computer addict with an early-generation Sinclair computer to my name. I was working at Halcrow where we were trying to convince them that CAD was the future. At the time, we spent a lot of time stencilling notes onto contract drawings. I was a port architect and Halcrow was designing all the new ports during this first Middle Eastern building boom: Jubail, Jeddah, Dubai, and Jebel Ali. I made a change after seeing a lecture at the RIBA where YRM showed off its new technology, the fabulous Intergraph system. I joined YRM to play with the new toy; it was so fabulously expensive we worked the machines in shifts. Hugh Whitehead joined YRM soon after as we graduated from 2D in black and white to 3D in colour.

Hugh and I were both sent to the conference in Huntsville, Alabama every year and here we got to know the cutting-edge CAD scene. As a result of connections made there, we had the opportunity to alpha test Intergraph’s ‘Master Architect’ software and one of the developers was Robert Aish. Master Architect was sitting on top of a mechanical modelling system that was very advanced and had parametric modules. When Master Architect was cancelled, Hugh and I persuaded YRM to employ Robert to continue with our parametric work. One summer Grimshaw came along with the Waterloo Station International Terminal project. I grabbed the model, cross-examined the project architect on the setting-out principles behind the design and built a parametric model of it.

A contemporary example of the need for 3D parametric technology in practice was the Pinnacle Tower by KPF. This tower at Bishopsgate in London has a unique cascading rooftop with a viewing gallery and restaurant, and was designed to be the tallest building in London, exactly in line with the Shard. The design team were hard at work but found it difficult to come up with a rational geometry to realise their design goals. The geometry of the tower is based on inclined planes in 3D, and on plan, circular arcs which get smaller to form a sheared cone. The top is mapped as a curve. This is then remapped back onto the inclined planes and sheared cones. The design of the project was carried out on the new GC software developed by Bentley and used at SG.

THE FUTURE OF SOFTWARE

SG emerged at a time of great opportunities in computational tools and developments in computer hardware. This has meant that, far from being an elitist pursuit, anybody could join in. A new product category in architecture, ‘parametric design’, was born. It seems likely that this will soon disappear as it is increasingly clear that parametric design will become embedded into mainstream CAD software, just as it has in the mechanical arena. In order to use parametric software well, a user needs to have an underlying mental model of a process. This is often geometric in nature, but it increasingly encompasses other aspects such as performance, cost and end-user enjoyment. The digital age is changing our perceptions as designers and for myself this is where the future of architectural design lies. How do we use digital technology to inform, survive and entertain ourselves?

 

11 GC model of Kohn Pedersen Fox (architects), The Pinnacle, London, UK, under construction.
The sheared cones concept was used in the Pinnacle building designed by the current PLP partners while at KPF. This image, which illustrates the sheared cones and the overlapping sets of inclined snakeskin facade components, is drawn from a new GC model built by Lars Hesselgren in 2012 while at PLP.

12 Kohn Pedersen Fox (architects), The Pinnacle, London, UK, under construction.
The Pinnacle in Bishopsgate, located between Tower 42 and St Mary Axe, will form a prominent landmark on the London skyline when completed. It is a project of Kohn Pedersen Fox (International) PA in which the following people at PLP Architecture were involved: Lee Polisano was Partner-in-Charge; Karen Cook, Design Partner; Lars Hesselgren, Head – Computational Geometry. (Lee Polisano and Karen Cook are partners of PLP Architecture.)

 

J PARRISH

THE FUTURE SEEMED SO CLEAR

The late 1960s and early 1970s were exciting times – as students we felt we could do anything, and Concorde and the Jumbo’s first flights and the first moon landing helped convince us even the sky wasn’t the limit!

My ‘eureka moment’ came with a final-year visit to architect John Lansdown of Turner, Lansdown & Holt. We’d already been introduced to CAD by our Systematic Design department at Bristol University, but in his dimly lit London office we were amazed by parametric design tools that calculated, designed, modelled and documented stairs, lifts and other building elements in 2D and 3D and provided neatly tabulated performance data. John had even written a program for specifying planting for different climates and locations with an interface that responded to questions typed as normal English sentences, just because it was an interesting challenge.

I’D SEEN THE FUTURE OF ARCHITECTURAL DESIGN!