このページのリンク

<電子ブック>
Learning from Data : Artificial Intelligence and Statistics V / edited by Doug Fisher, Hans-J. Lenz
(Lecture Notes in Statistics. ISSN:21977186 ; 112)

1st ed. 1996.
出版者 (New York, NY : Springer New York : Imprint: Springer)
出版年 1996
大きさ 450 p. 14 illus : online resource
著者標目 Fisher, Doug editor
Lenz, Hans-J editor
SpringerLink (Online service)
件 名 LCSH:Statistics 
FREE:Statistics
一般注記 I Causality -- 1 Two Algorithms for Inducing Structural Equation Models from Data -- 2 Using Causal Knowledge to Learn More Useful Decision Rules from Data -- 3 A Causal Calculus for Statistical Research -- 4 Likelihood-based Causal Inference -- II Inference and Decision Making -- 5 Ploxoma: Testbed for Uncertain Inference -- 6 Solving Influence Diagrams Using Gibbs Sampling -- 7 Modeling and Monitoring Dynamic Systems by Chain Graphs -- 8 Propagation of Gaussian Belief Functions -- 9 On Test Selection Strategies for Belief Networks -- 10 Representing and Solving Asymmetric Decision Problems Using Valuation Networks -- 11 A Hill-Climbing Approach for Optimizing Classification Trees -- III Search Control in Model Hunting -- 12 Learning Bayesian Networks is NP-Complete -- 13 Heuristic Search for Model Structure: The Benefits of Restraining Greed -- 14 Learning Possibilistic Networks from Data -- 15 Detecting Imperfect Patterns in Event Streams Using Local Search -- 16 Structure Learning of Bayesian Networks by Hybrid Genetic Algorithms -- 17 An Axiomatization of Loglinear Models with an Application to the Model-Search Problem -- 18 Detecting Complex Dependencies in Categorical Data -- IV Classification -- 19 A Comparative Evaluation of Sequential Feature Selection Algorithms -- 20 Classification Using Bayes Averaging of Multiple, Relational Rule-Based Models -- 21 Picking the Best Expert from a Sequence -- 22 Hierarchical Clustering of Composite Objects with a Variable Number of Components -- 23 Searching for Dependencies in Bayesian Classifiers -- V General Learning Issues -- 24 Statistical Analysis fo Complex Systems in Biomedicine -- 25 Learning in Hybrid Noise Environments Using Statistical Queries -- 26 On the Statistical Comparison of Inductive Learning Methods -- 27 Dynamical Selection of Learning Algorithms -- 28 Learning Bayesian Networks Using Feature Selection -- 29 Data Representations in Learning -- VI EDA: Tools and Methods -- 30 Rule Induction as Exploratory Data Analysis -- 31 Non-Linear Dimensionality Reduction: A Comparative Performance Analysis -- 32 Omega-Stat: An Environment for Implementing Intelligent Modeling Strategies -- 33 Framework for a Generic Knowledge Discovery Toolkit -- 34 Control Representation in an EDA Assistant -- VII Decision and Regression Tree Induction -- 35 A Further Comparison of Simplification Methods for Decision-Tree Induction -- 36 Robust Linear Discriminant Trees -- 37 Tree Structured Interpretable Regression -- 38 An Exact Probability Metric for Decision Tree Splitting -- VIII Natural Language Processing -- 39 Two Applications of Statistical Modelling to Natural Language Processing -- 40 A Model for Part-of-Speech Prediction -- 41 Viewpoint-Based Measurement of Semantic Similarity Between Words -- 42 Part-of-Speech Tagging from “Small” Data Sets
Ten years ago Bill Gale of AT&T Bell Laboratories was primary organizer of the first Workshop on Artificial Intelligence and Statistics. In the early days of the Workshop series it seemed clear that researchers in AI and statistics had common interests, though with different emphases, goals, and vocabularies. In learning and model selection, for example, a historical goal of AI to build autonomous agents probably contributed to a focus on parameter-free learning systems, which relied little on an external analyst's assumptions about the data. This seemed at odds with statistical strategy, which stemmed from a view that model selection methods were tools to augment, not replace, the abilities of a human analyst. Thus, statisticians have traditionally spent considerably more time exploiting prior information of the environment to model data and exploratory data analysis methods tailored to their assumptions. In statistics, special emphasis is placed on model checking, making extensive use of residual analysis, because all models are 'wrong', but some are better than others. It is increasingly recognized that AI researchers and/or AI programs can exploit the same kind of statistical strategies to good effect. Often AI researchers and statisticians emphasized different aspects of what in retrospect we might now regard as the same overriding tasks
HTTP:URL=https://doi.org/10.1007/978-1-4612-2404-4
目次/あらすじ

所蔵情報を非表示

電子ブック オンライン 電子ブック

Springer eBooks 9781461224044
電子リソース
EB00201605

書誌詳細を非表示

データ種別 電子ブック
分 類 LCC:QA276-280
DC23:519.5
書誌ID 4000105492
ISBN 9781461224044

 類似資料