A closed form expression is provided in the following paper.
SV Amari, RB Misra, Closed-form expressions for distribution of sum of exponential random variables, IEEE Transactions on Reliability, 46 (4), 519-522.
As already mentioned in the comments, you appear to have some fundamental misconceptions about Bayesian inference, and this is what I intend to address in my response, since you've already received enough computational guidance on your question.
If you were given a single observation $X_1 \mid \lambda \sim \operatorname{Poisson}(\lambda)$, and you wanted to use a Bayesian approach to make an inference about the posterior distribution of $\lambda$ based on this observation and your prior beliefs, then this is intuitively and obviously different than if you had many observations in your sample. Except in certain (unusual) circumstances, the more data you are able to observe from some parametric model, the "better" your understanding about what parameter(s) generated that data.
After all, this principle is something you should be familiar with from frequentist statistics: if you suppose that your observations are generated from a normal distribution with unknown mean $\mu$ and known standard deviation $\sigma = 1$, the sample mean is an intuitive and reasonable estimator for $\mu$; moreover, the larger your sample, the more likely this estimator will tend toward the true value of $\mu$. The same principle applies in Bayesian inference.
With this in mind, it should become clear that the likelihood function for $\lambda$ given the sample, should reflect the size of the sample in some meaningful way. Your version of the likelihood fails to do that. Now, you're not obligated to consider the entire sample--effectively, by using only a single observation, you're saying "I'm going to ignore the rest of the information available to me when calculating the posterior distribution"--but the posterior you calculate will not be as informative as one that uses all of the data. Another statistician, seeing your calculation, would not hesitate to improve upon it by using the other observations to produce a superior inference on $\lambda$.
Clearly, you have some mathematical competence when working with the calculation itself, but there's a deeper conceptual gap that I would invite you to try to close. Applying formulas is easy. Understanding what the formulas mean, relating it to the underlying concepts, is the essence of statistical thinking.
Best Answer
You leave it to be inferred by us based only on conventions that $\alpha$, rather than $\lambda$ is the shape parameter. There's also the question of whether $\lambda$ is supposed to be the intensity parameter, so that the distribution is $$ \frac{1}{\Gamma(\alpha)}\cdot (\lambda x)^{\alpha-1} e^{-\lambda x}\,(\lambda\,dx)\text{ for }x>0 $$ or its reciprocal, the scale parameter, so that the distribution is $$ \frac{1}{\Gamma(\alpha)}\cdot (x/\lambda)^{\alpha-1} e^{-x/\lambda}\,(dx/\lambda)\text{ for }x>0. $$
I'm going to guess that you mean the first of these two alternatives.
Once you understand the integral that defines the Gamma function, you're almost done. You have $$ \begin{align} \mathbb E(X^r) & = \int_0^\infty x^r \frac{1}{\Gamma(\alpha)}\cdot (\lambda x)^{\alpha-1} e^{-\lambda x}\,(\lambda\,dx) \\[12pt] & = \frac{1}{\Gamma(\alpha)} \cdot\frac{1}{\lambda^r} \int_0^\infty (\lambda x)^{r+\alpha-1} e^{-\lambda x}\,(\lambda\,dx) \\[12pt] & = \frac{1}{\Gamma(\alpha)}\cdot\frac{1}{\lambda^r} \int_0^\infty u^{r+\alpha-1} e^{-u}\,du \\[12pt] & = \frac{1}{\Gamma(\alpha)}\cdot\frac{1}{\lambda^r}\cdot\Gamma(r+\alpha). \tag1 \end{align} $$
This bears simplification. We have $$ \begin{align} \Gamma(r+\alpha) & = (r-1)\Gamma(r-1+\alpha)= (r-1)(r-2)\Gamma(r-2+\alpha)=\cdots \\[12pt] & \cdots=(r+\alpha-1)(r+\alpha-2)(r+\alpha-3)\cdots(r+\alpha-r)\Gamma(\alpha). \end{align} $$ Then you can reduce the fraction in $(1)$.