We present a new approach to the composition
of learning algorithms (in various models) for
classes of constant VC-dimension into learning algorithms for
more complicated classes.
We prove that if a class $\CC$ is learnable
in time $t$ from a hypothesis class $\HH$ of constant VC-dimension
then the class $\C^\star$ of all
functions $F$ of the form $F=f(g_1,\ldots,g_m)$,
where $f$ is any function
and $g_1,\ldots,g_m\in \C$, is
learnable in time polynomial in $t$ and $m$ .
We also use a simple
argument to prove that the composition theorem cannot be extended
to classes with a nonconstant VC-dimension.
A composition theorem for the exact learning model
(with equivalence queries only)
is proven in [BBK97] {\it only} for classes $\C$
of constant VC-dimension
that have {\it constant space} learning algorithms.
Constant space algorithms are hard to find
and have large complexity.
Our algorithm is simple and has a
complexity lower than the algorithm in [BBK97].
We then show how to change a PAC-learning
algorithm of $\CC$ from $\HH$ to an
SQ-learning algorithm and to a PAC-learning
algorithm for $\CC^\star$ with malicious noise
that achieves the optimal error rate $\eta/(1-\eta)+\beta$
for any $\beta$.
This, in particular, shows that if a class of
constant VC-dimension is PAC-learnable from a
class of constant VC-dimension then
it is SQ-learnable and PAC-learnable with malicious noise.
We apply this result for SQ-learning and PAC-learning with malicious
noise a general class of geometric objects.
This class includes the set of all geometric objects
in the constant dimensional space
that are bounded by $m$ algebraic surfaces of constant
degree (for example, hyperplanes, spheres, etc.).
This result generalizes all the results known
from the literature about learning geometric objects
in the SQ-learning and PAC-learning models with malicious noise.