Cantitate/Preț
Produs

Algorithmic Learning Theory: 14th International Conference, ALT 2003, Sapporo, Japan, October 17-19, 2003, Proceedings: Lecture Notes in Computer Science, cartea 2842

Editat de Ricard Gavaldà, Klaus P. Jantke, Eiji Takimoto
en Limba Engleză Paperback – 7 oct 2003

Din seria Lecture Notes in Computer Science

Preț: 33420 lei

Preț vechi: 41775 lei
-20% Nou

Puncte Express: 501

Preț estimativ în valută:
6396 6669$ 5315£

Carte tipărită la comandă

Livrare economică 21 martie-04 aprilie

Preluare comenzi: 021 569.72.76

Specificații

ISBN-13: 9783540202912
ISBN-10: 3540202919
Pagini: 332
Ilustrații: XII, 320 p.
Dimensiuni: 155 x 233 x 20 mm
Greutate: 0.47 kg
Ediția:2003
Editura: Springer Berlin, Heidelberg
Colecția Springer
Seriile Lecture Notes in Computer Science, Lecture Notes in Artificial Intelligence

Locul publicării:Berlin, Heidelberg, Germany

Public țintă

Research

Cuprins

Invited Papers.- Abduction and the Dualization Problem.- Signal Extraction and Knowledge Discovery Based on Statistical Modeling.- Association Computation for Information Access.- Efficient Data Representations That Preserve Information.- Can Learning in the Limit Be Done Efficiently?.- Inductive Inference.- Intrinsic Complexity of Uniform Learning.- On Ordinal VC-Dimension and Some Notions of Complexity.- Learning of Erasing Primitive Formal Systems from Positive Examples.- Changing the Inference Type – Keeping the Hypothesis Space.- Learning and Information Extraction.- Robust Inference of Relevant Attributes.- Efficient Learning of Ordered and Unordered Tree Patterns with Contractible Variables.- Learning with Queries.- On the Learnability of Erasing Pattern Languages in the Query Model.- Learning of Finite Unions of Tree Patterns with Repeated Internal Structured Variables from Queries.- Learning with Non-linear Optimization.- Kernel Trick Embedded Gaussian Mixture Model.- Efficiently Learning the Metric with Side-Information.- Learning Continuous Latent Variable Models with Bregman Divergences.- A Stochastic Gradient Descent Algorithm for Structural Risk Minimisation.- Learning from Random Examples.- On the Complexity of Training a Single Perceptron with Programmable Synaptic Delays.- Learning a Subclass of Regular Patterns in Polynomial Time.- Identification with Probability One of Stochastic Deterministic Linear Languages.- Online Prediction.- Criterion of Calibration for Transductive Confidence Machine with Limited Feedback.- Well-Calibrated Predictions from Online Compression Models.- Transductive Confidence Machine Is Universal.- On the Existence and Convergence of Computable Universal Priors.