Hassoun provides the first systematic account of artificial neural network paradigms by identifying clearly the fundamental concepts and major methodologies underlying most of the current theory and practice employed by neural network researchers.
As book review editor of the IEEE Transactions on Neural Networks, Mohamad Hassoun has had the opportunity to assess the multitude of books on artificial neural networks that have appeared in recent years. Now, in Fundamentals of Artificial Neural Networks, he provides the first systematic account of artificial neural network paradigms by identifying clearly the fundamental concepts and major methodologies underlying most of the current theory and practice employed by neural network researchers. Such a systematic and unified treatment, although sadly lacking in most recent texts on neural networks, makes the subject more accessible to students and practitioners. Here, important results are integrated in order to more fully explain a wide range of existing empirical observations and commonly used heuristics. There are numerous illustrative examples, over 200 end-of-chapter analytical and computer-based problems that will aid in the development of neural network analysis and design skills, and a bibliography of nearly 700 references. Proceeding in a clear and logical fashion, the first two chapters present the basic building blocks and concepts of artificial neural networks and analyze the computational capabilities of the basic network architectures involved. Supervised, reinforcement, and unsupervised learning rules in simple nets are brought together in a common framework in chapter three. The convergence and solution properties of these learning rules are then treated mathematically in chapter four, using the "average learning equation" analysis approach. This organization of material makes it natural to switch into learning multilayer nets using backprop and its variants, described in chapter five. Chapter six covers most of the major neural network paradigms, while associative memories and energy minimizing nets are given detailed coverage in the next chapter. The final chapter takes up Boltzmann machines and Boltzmann learning along with other global search/optimization algorithms such as stochastic gradient search, simulated annealing, and genetic algorithms.
Le informazioni nella sezione "Riassunto" possono far riferimento a edizioni diverse di questo titolo.
Joshua D. Angrist is Ford Professor of Economics at MIT.
Le informazioni nella sezione "Su questo libro" possono far riferimento a edizioni diverse di questo titolo.
Spese di spedizione:
EUR 3,74
In U.S.A.
Descrizione libro Paperback. Condizione: new. New. Fast Shipping and good customer service. Codice articolo Holz_New_0262514672
Descrizione libro Paperback. Condizione: new. New. Codice articolo Wizard0262514672
Descrizione libro Paperback. Condizione: new. New Copy. Customer Service Guaranteed. Codice articolo think0262514672
Descrizione libro Paperback. Condizione: New. BRAND NEW, Perfect Shape, No Remainder Mark,Fast Shipping With Online Tracking, International Orders shipped Global Priority Air Mail, All orders handled with care and shipped promptly in secure packaging, we ship Mon-Sat and send shipment confirmation emails. Our customer service is friendly, we answer emails fast, accept returns and work hard to deliver 100% Customer Satisfaction!. Codice articolo 9071816
Descrizione libro Paperback. Condizione: Brand New. 537 pages. 9.25x7.50x1.25 inches. In Stock. Codice articolo zk0262514672
Descrizione libro Condizione: New. Book is in NEW condition. Codice articolo 0262514672-2-1
Descrizione libro Condizione: New. New! This book is in the same immaculate condition as when it was published. Codice articolo 353-0262514672-new