CONF Thimm-96.5/IDIAP Neural Network Pruning and Pruning Parameters Thimm, Georg Fiesler, Emile Furuhashi, Takeshi Ed. generalization network size neural network neural network optimization parameters pruning EXTERNAL https://publications.idiap.ch/attachments/papers/1996/prune96.pdf PUBLIC Dept. of Information Electronics Nagoya University - The 1st Workshop on Soft Computing 1996 Furo-cho, Chikusa-ku, Nagoya 464-01, Japan August 1996 Is published as http://www.bioele.nuee.nagoya-u.ac.jp/wsc1/ The default multilayer neural network topology is a fully interlayer connected one. This simplistic choice facilitates the design but it limits the performance of the resulting neural networks. The best-known methods for obtaining partially connected neural networks are the so called pruning methods which are used for optimizing both the size and the generalization capabilities of neural networks. Two of the most promising pruning techniques have therefore been selected for a comparative study. It is shown that these novel techniques are hampered by having numerous user-tunable parameters, which can easily nullify the benefits of these advanced methods. Finally, based on the results, conclusions about the execution of experiments and suggestions for conducting future research on neural network pruning are drawn.