I am trying to fit data of the form $(x_i,y_i)$, $i=1,\ldots,n$, in a curve of the form $y=Ax^B+C$, where $B\in (0,1)$. All three constants $A,B,C$ are to be determined optimally (no particular norm for the moment). Also, the $x_i$'s are positive integers, if that helps.
Is there any standard method to attack such a problem?

The standard method is least-squares, which in your case is non-linear, requiring to resort to the Levenberg-Marquardt algorithm, with an extra difficulty as $B$ is constrained.
You can work out your particular case differently: see $B$ as an independent parameter. For a given value of $B$, you easily solve the linear fitting problem
$$y_i=Ax'_i+B$$ where $x'_i=x_i^B$, and compute the residual error $E(B)$.
Now you can use your preferred 1D minimizer (golden section) to find $\min_{B\in[0,1]}E(B)$.
Make sure to have a look at the curve of $E(B)$. Hopefully it will be unimodal (single maximum). It may turn out that the maximum lies outside the range, in which case the optimum is one of $B=0$ or $B=1$.
If your data set has outliers, you may need to use a robust method such as RANSAC. The latter will require to solve the exact fitting for three points.
You can eliminate $A$ and $C$ by
$$\frac{y_2-y_0}{y_1-y_0}=\frac{x_2^B-x_0^B}{x_1^B-x_0^B}=u$$ which is nonlinear in $B$, and can be solved by regula falsi or possibly Newton. From $B$, you derive $A$ and $C$ easily.
Final note:
If the dataset contains values such that $$\frac{x_2}{x_0}=\left(\frac{x_1}{x_0}\right)^2,$$ that is to say $$x_1=\sqrt{x_0x_2},$$ the previous equation simplifies to
$$\left(\frac{x_1}{x_0}\right)^B+1=u$$ which is directy solvable for $B$.