Abstract
In this paper, regression problem in learning theory is investigated by least square schemes in polynomial space. Results concerning the estimation of rate of convergence are derived. In particular, it is shown that for one variable smooth regression function, the estimation is able to achieve good rate of convergence. As a main tool in the study, the Jackson operator in approximation theory is used to estimate the rate. Finally, the obtained estimation is illustrated by applying simulated data.
| Original language | English |
|---|---|
| Pages (from-to) | 516-521 |
| Number of pages | 6 |
| Journal | Neurocomputing |
| Volume | 74 |
| Issue number | 4 |
| DOIs | |
| State | Published - Jan 2011 |
Keywords
- Covering number
- Jackson operator
- Learning theory
- Rate of convergence