I think you are using words that you don't understand. The sum of multiple dice does not follow a normal distribution. See my explanation for the sum of two $n$-sided dice case in my answer here.
Supposing you remove the word "normal," you can see how the mean and standard deviation interact by the linearity of expected value.
The expected value of a discrete random variable, $X$, is given as $E[X] = \sum\limits_{x\in\mathcal{M}} x\cdot Pr(X=x)$ (where $\mathcal{M}$ is the sample space)
The expected value operator is linear. I.e., $E[\alpha_1 X_1 + \alpha_2 X_2] = \alpha_1 E[X_1] + \alpha_2 E[X_2]$
The mean of a random variable $X$ is defined to be $E[X]$
Let $X$ be a random variable described by the result of a die throw. If $X$ is a fair $n$-sided die with sides labeled $1,2,\dots,n$, then the mean, $\mu = E[X]$ is equal to $\frac{n+1}{2}$.
The variance of $X$ in this case is $\frac{n^2-1}{12}$.
Let $X$ be a random variable described by the result of a die throw. If $X$ is a fair $n$-sided die with sides labeled $1,2,\dots,n$, then the expected value of $\frac{1}{X}$ is $H_n\cdot \frac{1}{n}$ where $H_n$ is the $n^{th}$ Harmonic Number
Suppose our random variable, $Y$, is a linear combination of random variables. Let $X_1$ be the random variable described by the result of an $a_1$-sided die, $X_2$ be the random variable described by the result of an $a_2$-sided die, ..., $X_k$ be the random variable described by the result of an $a_k$-sided die.
Let $Y=\alpha_1 X_1 + \alpha_2 X_2 + \dots \alpha_k X_k$
Then $E[Y] = E[\alpha_1 X_1 + \dots + \alpha_k X_k] = \alpha_1 E[X_1] + \dots + \alpha_k E[X_k] = \alpha_1\frac{a_1+1}{2} + \dots + \alpha_k\frac{a_k+1}{2}$
Note, that for the case where $X_1, X_2,\dots, X_k$ all describe the result of rolling an $n$-sided die, you get $E[X_1+X_2+\dots+X_k] = E[X_1]+\dots+E[X_k] = \frac{n+1}{2}+\dots + \frac{n+1}{2} = k\cdot \frac{n+1}{2}$, matching your intuition.
When you use multiplication, $E[XY] = E[X]E[Y]-Cov(X,Y)$, however noting that when $X$ and $Y$ are independent (which they will be when $X$ and $Y$ are for different dice [not necessarily different types of dice]) it implies that $Cov(X,Y)=0$ and that $E[X,Y] = E[X]E[Y]$
This gives us a way to calculate the mean of any combination of results of dice and $+,-,\times$. Noting that division is just multiplication by the multiplicative inverse of the other, we get that $E[X/Y] = E[X\cdot \frac{1}{Y}] = E[X]E[\frac{1}{Y}]$ and the term on the right will rely on the $n^{th}$ harmonic number.
I see no easy way of giving a general rule to use when using exponentiation of one random variable to the power of another apart from referring you to calculate it directly using the definition of expected value. For finding the expected value of a random variable to an integer power, I refer you to the definition of Moments and to Generalized Harmonic Numbers as well as for the specific case of $E[X^2]$ being the $n^{th}$ Square Pyramidal Number divided by $n$.
For standard deviation, $\sigma$, of a random variable $X$ note that $\sigma^2=Var[X] = E[(X-E[X])^2] = E[X^2]- E[X]^2$
With all of this information, you should have the tools to calculate whatever you are looking for.
Example
You have three fair dice, a d4, a d6, and a d8. Let $x,y,z$ represent the results of throwing the d4, d6, and d8 respectively. Suppose that you are playing a game where you get to attack and deal $z+xy-2$ damage. What is the expected amount of damage and what is the standard deviation?
The average amount of damage is: $E[Z+XY-2] = E[Z]+E[X]E[Y] - 2 = \frac{8+1}{2} + \frac{4+1}{2}\cdot\frac{6+1}{2} - 2 =4.5 + 3.5\cdot 2.5 - 2 = 11.25$
The variance is: $Var[Z+XY-2] = E[(Z+XY-2)^2] - (E[Z+XY-2])^2 = E[Z^2 + X^2Y^2 + 4 + 2XYZ - 4Z -4XY] - 11.25^2$
$= E[Z^2] +E[X^2]E[Y^2]+4+2E[X]E[Y]E[Z] - 4E[Z] - 4E[X]E[Y] - 11.25^2$
$=\frac{(8+1)(16+1)}{6} + \frac{(4+1)(8+1)}{6}\frac{(6+1)(12+1)}{6}+4+2\cdot 2.5\cdot 3.5\cdot 4.5 - 4\cdot 4.5 - 4\cdot 2.5\cdot 3.5 - 11.25^2$
$=42.4375$
The standard deviation then is $\sqrt{42.4375}\approx 6.51$.