Normal Accidents: Living with High Risk Technologies et plus d'un million d'autres livres sont disponibles pour le Kindle d'Amazon. En savoir plus

Amazon Rachète votre article
Recevez un chèque-cadeau de EUR 5,86
Amazon Rachète cet article
Vous l'avez déjà ? Vendez votre exemplaire ici
Désolé, cet article n'est pas disponible en
Image non disponible pour la
couleur :
Image non disponible

Commencez à lire Normal Accidents: Living with High Risk Technologies sur votre Kindle en moins d'une minute.

Vous n'avez pas encore de Kindle ? Achetez-le ici ou téléchargez une application de lecture gratuite.

Normal Accidents: Living With High Risk Technologies [Anglais] [Broché]

Charles Perrow
4.0 étoiles sur 5  Voir tous les commentaires (1 commentaire client)

Voir les offres de ces vendeurs.


Prix Amazon Neuf à partir de Occasion à partir de
Format Kindle EUR 23,56  
Broché --  
Broché EUR 31,41  
Broché --  
Vendez cet article - Prix de rachat jusqu'à EUR 5,86
Vendez Normal Accidents: Living With High Risk Technologies contre un chèque-cadeau d'une valeur pouvant aller jusqu'à EUR 5,86, que vous pourrez ensuite utiliser sur tout le site Les valeurs de rachat peuvent varier (voir les critères d'éligibilité des produits). En savoir plus sur notre programme de reprise Amazon Rachète.

Offres spéciales et liens associés

Les clients ayant acheté cet article ont également acheté

Détails sur le produit

  • Broché: 400 pages
  • Editeur : Basic Books; Édition : Reprint (11 septembre 1985)
  • Langue : Anglais
  • ISBN-10: 0465051421
  • ISBN-13: 978-0465051427
  • Dimensions du produit: 22,9 x 15 x 2,8 cm
  • Moyenne des commentaires client : 4.0 étoiles sur 5  Voir tous les commentaires (1 commentaire client)
  • Classement des meilleures ventes d'Amazon: 24.665 en Livres anglais et étrangers (Voir les 100 premiers en Livres anglais et étrangers)
  •  Souhaitez-vous compléter ou améliorer les informations sur ce produit ? Ou faire modifier les images?

En savoir plus sur l'auteur

Découvrez des livres, informez-vous sur les écrivains, lisez des blogs d'auteurs et bien plus encore.

Dans ce livre (En savoir plus)
Parcourir et rechercher une autre édition de ce livre.
Première phrase
Our first example of the accident potential of complex systems is the accident at the Three Mile Island Unit 2 nuclear plant near Harrisburg, Pennsylvania, on March 28, 1979. Lire la première page
En découvrir plus
Parcourir les pages échantillon
Couverture | Copyright | Table des matières | Extrait | Index | Quatrième de couverture
Rechercher dans ce livre:

Commentaires en ligne 

5 étoiles
3 étoiles
2 étoiles
1 étoiles
4.0 étoiles sur 5
4.0 étoiles sur 5
Commentaires client les plus utiles
1 internautes sur 1 ont trouvé ce commentaire utile 
4.0 étoiles sur 5 C'est un classique 19 octobre 2013
Par charvolin
Format:Broché|Achat vérifié
L'ouvrage est un classique pour comprendre le quotidien des systèmes techniques complexes, dans des industries très variées (d'usines chimiques, de centrales nucléaires, à l'aviation ou les OGM).
Avez-vous trouvé ce commentaire utile ?
Commentaires client les plus utiles sur (beta) 4.0 étoiles sur 5  54 commentaires
127 internautes sur 147 ont trouvé ce commentaire utile 
3.0 étoiles sur 5 Living With High-Risk Conclusions 29 janvier 2004
Par Robert I. Hedges - Publié sur
Format:Broché|Achat vérifié
I have been mulling over this review for a while now, and am still undecided on the correct rating to award this book. On the one hand Perrow offers some genuine insight into systems safety, but frequently does not understand the technicalities of the systems (or occasionally their operators) well enough to make informed decisions and recommendations. In more egregious cases he comes to conclusions that are guaranteed to reduce safety (as when he argues that supertankers should be run by committee, and the usefulness of the Captain is no more) or are merely the cherished liberal opinions of an Ivy League sociologist (he teaches at Yale) as when he argues for unilateral nuclear disarmament, government guaranteed income plans, and heroin maintenance (distribution) plans for addicts "to reduce crime." In the case of disarmament, remember this was written during the early 1980s while the Soviet Union was still a huge threat...complete nuclear disarmament would have resulted in fewer US nuclear accidents, but would NOT have made us safer as we would have been totally vulnerable to intentional nuclear attack. He has great personal animosity toward Ronald Reagan, and makes inflammatory statements in the mining section that mining safety regulations would surely be weakened by Reagan, causing many more accidents and deaths. Later in the same section, though, he concludes that mining is inherently dangerous, and no amount of regulation can make it safe. So which is it? Any of this is, at very best, folly, but regardless of political bent (he is a self avowed "leftist liberal") has absolutely no place in a book ostensibly on safety systems. As such I think portions of this book show what is so wrong in American academia today: even genuinely excellent research can be easily spoiled when the conclusions are known before the research is started. This is one of the many reasons that physical scientists scorn the social sciences, and it doesn't have to be this way.
Having said all that there IS a wealth of good information and insight in this book when Perrow sticks to systems and their interactions. The book contains the finest analysis commercially available of the Three Mile Island near-disaster, and his insight about how to improve safety in nuclear plants was timely when the book was written in 1984, though many improvements have been made since then.
Speaking as a commercial airline pilot, I feel his conclusions and observations about aircraft safety were generally true at the time of printing in 1984, but now are miserably out of date. (The same is true of the Air Traffic Control section.) I believe that he generally has a good layman's grasp of aviation, so I am willing to take it as a given that he has a knowledgeable layman's comprehension of the other systems discussed. As an aside, he never gets some of the technicalities quite right. For instance, he constantly uses the term 'coupling' incorrectly in the engineering sense; this is particularly objectionable in the aviation system where it has a very specific meaning to aeronautical engineers and pilots.
The section on maritime accidents and safety is superbly written. Here I am not an expert, but there seems to be a high degree of correlation with the aviation section. His section on "Non Collision Course Collisions" by itself makes this book a worthwhile read. He presents very compelling information and reasoning until the very end of the section, at which point he suggests that since ships are now so big, large ships (especially supertankers) essentially should have no Captain, but should be run by committee. This is an invalid conclusion, and he offers no evidence or substantial argument to support that idea. Clearly, it is an idea hatched in his office and not on a ship (or plane.) There always needs to be a person in a place of ultimate authority in fast moving, dynamic systems, or the potential exists to have crew members begin to work at direct odds with each other, making a marginal situation dangerous. Ironically, in the very same part of the discussion where he concludes that there should be no Captain, he has hit upon the key to the problem. He mentions that he was pleased to see that some European shippers were now training their crews together as a team, and that he expected this to lower accident rates. He is, in fact, exactly right about that. Airlines now have to train crews in Crew Resource Management (CRM) in which each member of the crew has the right and obligation to speak up if they notice anything awry in the operation of their aircraft, and the Captain makes it a priority to listen to the input of others, as everyone has a different set of concerns and knowledge. In this way, the Captain becomes much less dictatorial, and becomes more of a final decision maker after everyone has had their say. It IS critical, though, to maintain someone in command, as there is no time to assemble a staff meeting when a ship is about to run aground, or a mid-air collision is about to occur. Many other well documented studies and books have come to this conclusion, and in the airline industry since CRM was introduced the accident rate has decreased dramatically.
Overall, if you have a desire to understand high risk systems, this book has a lot of good information in it; however it is woefully out of date and for that reason among others, I can only recommend it with reservations. A better and much more contemporary introductory book on the subject is 'Inviting Disaster' by James R. Chiles. Remember, this book was written over twenty years ago, and much has changed since then. There is knowledge to be gleaned here, but you have to be prepared to sort the wheat from the chaff.
25 internautes sur 28 ont trouvé ce commentaire utile 
5.0 étoiles sur 5 Of Lasting Value, Relevant to Today's Technical Maze 27 janvier 2003
Par Robert David STEELE Vivas - Publié sur
Edit of 2 April 2007 to add link and better summary.

I read this book when it was assigned in the 1980's as a mainstream text for graduate courses in public policy and public administration, and I still use it. It is relevant, for example, to the matter of whether we should try to use nuclear bombs on Iraq--most Americans do not realize that there has never (ever) been an operational test of a US nuclear missile from a working missle silo. Everything has been tested by the vendors or by operational test authorities that have a proven track record of falsifying test results or making the tests so unrealistic as to be meaningless.

Edit: my long-standing summary of the author's key point: Simple systems have single points of failure that are easy to diagnose and fix. Complex systems have multiple points of failure that interact in unpredictable and often undetectable ways, and are very difficult to diagnose and fix. We live in a constellation of complex systems (and do not practice the precationary principle!).

This book is also relevant to the world of software. As the Y2K panic suggested, the "maze" of software upon which vital national life support systems depend--including financial, power, communications, and transportation software--has become very obscure as well as vulnerable. Had those creating these softwares been more conscious of the warnings and suggestions that the author provides in this book, America as well as other nations would be much less vulnerable to terrorism and other "acts of man" for which our insurance industry has not planned.

I agree with another review who notes that this book is long overdue for a reprint--it should be updated. I recommended it "as is," but believe an updated version would be 20% more valuable.

Edit: this book is still valuable, but the author has given us the following in 2007:
The Next Catastrophe: Reducing Our Vulnerabilities to Natural, Industrial, and Terrorist Disasters
19 internautes sur 22 ont trouvé ce commentaire utile 
5.0 étoiles sur 5 Altogether a fascinating and informative book 21 mars 2003
Par Atheen M. Wilson - Publié sur
Wow. This is an incredible book. I have to admit, though, that I had some difficulty getting into Normal Accidents. There seemed an overabundance of detail, particularly on the nuclear industry's case history of calamity. This lost me, since I'm not familiar with the particulars of equipment function and malfunction. The book was mentioned, however, by two others of a similar nature and mentioned with such reverence, that after I had finished both, I returned to Perrow's book, this time with more success.
Professor Perrow is a PhD in sociology (1960) who has taught at Yale University Department of Sociology since 1981 and whose research focus has been human/technology interactions and the effects of complexity in organizations. (His most recent publication is the The AIDS disaster : the Failure of Organizations in New York and the Nation, 1990.)
In Normal Accidents, he describes the failures that can arise "normally" in systems, ie. those problems that are expected to arise and can be planned for by engineers, but which by virtue of those planned fail-safe devices, immeasurably complicate and endanger the system they are designed to protect. He describes a variety of these interactions, clarifying his definitions by means of a table (p. 88), and a matrix illustration (p. 97). Examples include systems that are linear vs complex, and loosely vs tightly controlled. These generally arise through the interactive nature of the various components the system itself. According to the matrix, an illustration of a highly linear, tightly controlled system would be a dam. A complex, tightly controlled system would be a nuclear plant, etc.
The degree to which failures may occur varies with each type of organization, as does the degree to which a recovery from such a failure is possible. As illustrations, the author describes failures which have, or could have, arisen in a variety of settings: the nuclear industry, maritime activities, the petrochemical industry, space exploration, DNA research and so on.
The exciting character of the stories themselves are worth the reading; my favorite, and one I had heard before, is the loss of an entire lake into a salt mine. More important still is the knowledge that each imparts. Perrow makes abundantly apparent by his illustrations the ease with which complex systems involving humans can fail catastrophically. (And if Per Bak and others are correct, almost inevitably).
Probably the most significant part of the work is the last chapter. After discussing the fallibility of systems that have grown increasingly complex, he discusses living with high risk systems, particularly why we are and why it should change. In a significant statement he writes, "Above all, I will argue, sensible living with risky systems means keeping the controversies alive, listening to the public, and recognizing the essentially political nature of risk assessment. Unfortunately, the issue is not risk, but power; the power to impose risks on the many for the benefit of the few (p. 306)," and further on, "Risks from risky technologies are not borne equally by the different social classes [and I would add, countries]; risk assessments ignore the social class distribution of risk (p. 310)." How true. "Quo Bono?" as the murder mystery writers might say; "Who benefits?" More to the point, and again with that issue in mind, he writes "The risks that made our country great were not industrial risks such as unsafe coal mines or chemical pollution, but social and political risks associated with democratic institutions, decentralized political structures, religious freedom and plurality, and universal suffrage (p. 311)." Again, very true.
Professor Perrow examines the degrees of potential danger from different types of system and suggests ways of deciding which are worth it to society to support and which might not be. These include categorizing the degree and the extent of danger of a given system to society, defining the way these technologies conflict with the values of that society, determining the likelihood that changes can be made to effectively alter the dangerous factors through technology or training of operators, and the possibility of placing the burden of spill-over costs on the shoulders of the institutions responsible. The latter might conceivably lead to corrective changes, either by the institutions themselves in order to remain profitable or by consumers through purchasing decisions.
The bibliography for the book is quite extensive and includes a variety of sources. These include not only popular books and publications on the topics of individual disasters, but government documents, research journals, and industry reports as well. I did not find any reference to the Johnstown flood, my particular favorite dam burst story, but there are a wide variety of references to chose from should someone wish to do their own research on the topic.
Altogether a fascinating and informative book.
14 internautes sur 16 ont trouvé ce commentaire utile 
5.0 étoiles sur 5 Insightful perspective on serious industrial accidents. 17 juillet 1998
Par Un client - Publié sur
Normal Accidents is the best summary of major industrial accidents in the USA that I have encountered. It is written in a factual and technically complete style that is particularly attractive to anyone with a technical background or interest. I was able to read a borrowed copy from a colleague a few years ago when I was appointed as chairman of the safety committee at a manufacturing facility where workers had potential for exposure to toxic gasses, high voltage, x-radiation, and other more everyday industrial hazards. The author's insight is right on target for achieving a workable understanding of the cause and prevention of disaster events. I wanted to buy copies for all our engineering managers and safety committee members, but the book is out of print. It is my fond hope that the author will write an updated version with analysis of more recent events as well as the well-chosen accidents in the previous edition. For any safety related product or process de! ! signer, this book is a must read! For any technically cognizant reader, this book is a delight to read, even if it is a little scary in its implications. For everyone else, it has some really interesting historical stories.
15 internautes sur 18 ont trouvé ce commentaire utile 
4.0 étoiles sur 5 Cool water for hot-headed analysts of complex systems 7 juillet 1998
Par Un client - Publié sur
I'm dismayed to discover that 'Normal Accidents' is so difficult to find.
Like all voters, I'm sometimes asked to make choices about the use of potentially devastating technology, despite having no training in engineering and only a sketchy idea of statistical risk analysis. 'Normal Accidents' doesn't reduce my reliance on experts, but it does provide a common language for us to discuss the issues.
Perrow's accident descriptions are masterly, and should disturb anyone who lightly dismisses accidents in complex systems as "simple human error", or assumes that all systems can be made safe by a technological fix. I've used Perrow's complexity / coupling matrix as a tool for thinking about and discussing the risks involved in decisions about many systems in addition to those Perrow actually discusses, not least software systems.
I think this book still has a lot to offer anyone interested in public debate about complex technological issues, and I hope it will be reprinted. A new edition would be even better.
Ces commentaires ont-ils été utiles ?   Dites-le-nous
Rechercher des commentaires
Rechercher uniquement parmi les commentaires portant sur ce produit

Discussions entre clients

Le forum concernant ce produit
Discussion Réponses Message le plus récent
Pas de discussions pour l'instant

Posez des questions, partagez votre opinion, gagnez en compréhension
Démarrer une nouvelle discussion
Première publication:
Aller s'identifier

Rechercher parmi les discussions des clients
Rechercher dans toutes les discussions Amazon

Rechercher des articles similaires par rubrique


Souhaitez-vous compléter ou améliorer les informations sur ce produit ? Ou faire modifier les images?