Supervised learning

Miquel Noguer i Alonso, Daniel Bloch and David Pacheco Aznar

In the previous chapters, we have presented artificial and recurrent neural networks, discussing the construction of neurons and introducing different networks with their respective algorithms. We shall now give a more formal representation of such networks.

5.1 INTRODUCTION TO LEARNING THEORY

Vapnik (1982) developed a fundamental approach to learning theory called structural risk minimisation. In this section we introduce the main ideas.

5.1.1 Function estimation

Risk minimisation

In a general supervised learning problem, we have two spaces of objects X and Y and would like to learn a function h: X → Y, which outputs an object y ∈ Y, given x ∈ X. The function h is sometimes called a score function or hypothesis. It is an element of some space of possible functions H called the hypothesis space. We assume a given sample {(x1, y1), . . . , (xn, yn)}, where xi ∈ X is an input and yi ∈ Y is the corresponding response that we wish to get from h(xi). We consider probabilistic learning models and assume that there is a joint probability distribution P(x, y) and that the sample is made of independent and identically distributed (iid) variables drawn from that distribution. That is, we model

Sorry, our subscription options are not loading right now

Please try again later. Get in touch with our customer services team if this issue persists.

New to Risk.net? View our subscription options

You need to sign in to use this feature. If you don’t have a Risk.net account, please register for a trial.

Sign in
You are currently on corporate access.

To use this feature you will need an individual account. If you have one already please sign in.

Sign in.

Alternatively you can request an individual account here