For non-parametric or parametric quantile estimation.
If parametric quantile esimation estimates $p$ by computing the MLE, then how to get non-parametric $p$?
Related:
https://mathoverflow.net/questions/48223/parametric-vs-non-parametric-estimation-of-quantiles
But I don't understand how to do non-parametric quantile estimation, nor how is $p$ estimated in that case. Does one still have to use the inverse CDF?
They are just looking at the sample quantiles. If you think about "observations" instead of being a collection of values, instead think of them as a discrete approximation to the true distribution. That is, if you have a set of observations $\{x_n\}_{n=1}^N$, you can construct the empirical cdf as a mixture of point masses centered at these locations:
$$\hat{P}[X] = \sum_{n=1}^N \frac 1 N \delta_{x_n}$$
For small $N$ this isn't very effective, and so it's helpful to incorporate some parametric assumptions (like they do in that link), but if you have tons of data $\hat{P}$ can characterize the true distribution very well; it's analogous to Bayesians using Monte Carlo to characterize their posterior instead of explicitly getting an expression for it.
It turns out that the empirical CDF has a bunch of useful properties, like, for example, consistency in moments, consistency in quantiles, etc. If you want to get an approximation to $P[a \leq X \leq b]$, then $\hat{P}[a \leq X \leq b]$ is a very good estimate with large sample sizes.
To specifically estimate quantiles, you just look for points in your support where the proportion of observations on one side or the other matches your goal. For example, say you have 1000 observations and you want to estimate the .25 quantile. Then you would find the first point in your support such that 25% of the sampled values are below that point.