Wikipedia article
Lindemann–Weierstrass Theorem (Baker's reformulation).
If $a_1,\ldots,a_n$ are algebraic numbers, and $α_1,\ldots,α_n$ are distinct algebraic numbers, then$$a_{1}e^{\alpha _{1}}+a_{2}e^{\alpha _{2}}+\cdots +a_{n}e^{\alpha _{n}}=0$$has only the trivial solution $a_1 e^{\alpha_1} + a_2 e^{\alpha_2} + \cdots + a_n e^{\alpha_n} = 0$ has only the trivial solution $a_i = 0$ for all $i = 1, \dots, n.$

Proof

The proof relies on two preliminary lemmas. Notice that Lemma B itself is already sufficient to deduce the original statement of Lindemann–Weierstrass theorem.

Preliminary lemmas

Let $c(1),\ldots,c(r)$ be integers and, for every $k$ between 1 and $r$, let $γ(k)_1,\ldots,γ(k)_{m(k)}$ be the roots of a non-zero polynomial with integer coefficients $T_k(x)$. If $γ(k)_i≠γ(u)_v$ whenever $(k,i)≠(u,v)$, then $$c(1)\left (e^{\gamma(1)_1}+\cdots+ e^{\gamma(1)_{m(1)}} \right ) + \cdots + c(r) \left (e^{\gamma(r)_1}+\cdots+ e^{\gamma(r)_{m(r)}} \right) = 0$$ has only the trivial solution $c(i)=0$ for all $i = 1, \dots, r.$
Proof of Lemma A. To simplify the notation set: \begin{align} & n_0 =0, & & \\ & n_i =\sum\nolimits_{k=1}^i m(k), & & i=1,\ldots,r \\ & n=n_r, & & \\ & \alpha_{n_{i-1}+j} =\gamma(i)_j, & & 1\leq i\leq r,\ 1\leq j\leq m(i) \\ & \beta_{n_{i-1}+j} =c(i). \end{align} Then the statement becomes $$\sum_{k=1}^n \beta_k e^{\alpha_k}\neq 0.$$ Let $p$ be a prime number and define the following polynomials: $$f_i(x) = \frac {\ell^{np} (x-\alpha_1)^p \cdots (x-\alpha_n)^p}{(x-\alpha_i)},$$ where $\ell$ is a non-zero integer such that $\ell\alpha_1,\ldots,\ell\alpha_n$ are all algebraic integers. DefineUp to a factor, this is the same integral appearing in the proof that $e$ is a transcendental number, where $β_1=1, ...,β_m=m$. The rest of the proof of the Lemma is analog to that proof. $$I_i(s) = \int^s_0 e^{s-x} f_i(x) \, dx.$$ Using integration by parts we arrive at $$I_i(s) = e^s \sum_{j=0}^{np-1} f_i^{(j)}(0) - \sum_{j=0}^{np-1} f_i^{(j)}(s),$$ where $np-1$ is the degree of $f_i$, and $f_i^{(j)}$ is the $j$-th derivative of $f_i$. This also holds for $s$ complex (in this case the integral has to be intended as a contour integral, for example along the straight segment from 0 to $s$) because $$-e^{s-x} \sum_{j=0}^{np-1} f_i^{(j)}(x)$$ is a primitive of $e^{s-x} f_i(x)$. Consider the following sum: \begin{align} J_i &=\sum_{k=1}^n\beta_k I_i(\alpha_k)\\[5pt] &= \sum_{k=1}^n\beta_k \left ( e^{\alpha_k} \sum_{j=0}^{np-1} f_i^{(j)}(0) - \sum_{j=0}^{np-1} f_i^{(j)}(\alpha_k)\right ) \\[5pt] &=\left(\sum_{j=0}^{np-1}f_i^{(j)}(0)\right)\left(\sum_{k=1}^n \beta_k e^{\alpha_k}\right)-\sum_{k=1}^n\sum_{j=0}^{np-1} \beta_kf_i^{(j)}(\alpha_k)\\[5pt] &= -\sum_{k=1}^n \sum_{j=0}^{np-1} \beta_kf_i^{(j)}(\alpha_k) \end{align} In the last line we assumed that the conclusion of the Lemma is false. In order to complete the proof we need to reach a contradiction. We will do so by estimating $|J_1\cdots J_n|$ in two different ways. First $f_i^{(j)}(\alpha_k)$ is an algebraic integer which is divisible by $p$! for $j\geq p$ and vanishes for $j\lt p$ unless $j=p-1$ and $k=i$, in which case it equals $\ell^{np}(p-1)!\prod_{k\neq i}(\alpha_i-\alpha_k)^p.$ This is not divisible by $p$ when $p$ is large enough because otherwise, putting $\delta_i=\prod_{k\neq i}(\ell\alpha_i-\ell\alpha_k)$ (which is a non-zero algebraic integer) and calling $d_i\in\mathbb Z$ the product of its conjugates (which is still non-zero), we would get that $p$ divides $\ell^p(p-1)!d_i^p$, which is false. So $J_i$ is a non-zero algebraic integer divisible by ($p$ − 1)!. Now $J_i=-\sum_{j=0}^{np-1}\sum_{t=1}^r c(t)\left(f_i^{(j)}(\alpha_{n_{t-1}+1}) + \cdots + f_i^{(j)}(\alpha_{n_t})\right).$ Since each $f_i(x)$ is obtained by dividing a fixed polynomial with integer coefficients by $(x-\alpha_i)$, it is of the form $f_i(x)=\sum_{m=0}^{np-1}g_m(\alpha_i)x^m, $ where $g_m$ is a polynomial (with integer coefficients) independent of $i$. The same holds for the derivatives $f_i^{(j)}(x)$. Hence, by the fundamental theorem of symmetric polynomials, $$f_i^{(j)}(\alpha_{n_{t-1}+1})+\cdots+f_i^{(j)}(\alpha_{n_t})$$ is a fixed polynomial with rational coefficients evaluated in $\alpha_i$ (this is seen by grouping the same powers of $\alpha_{n_{t-1}+1},\dots,\alpha_{n_t}$ appearing in the expansion and using the fact that these algebraic numbers are a complete set of conjugates). So the same is true of $J_i$, i.e. it equals $G(\alpha_i)$, where $G$ is a polynomial with rational coefficients independent of $i$. Finally $J_1\cdots J_n=G(\alpha_1)\cdots G(\alpha_n)$ is rational (again by the fundamental theorem of symmetric polynomials) and is a non-zero algebraic integer divisible by $(p-1)!^n$ (since the $J_i$'s are algebraic integers divisible by $(p-1)!$). Therefore $$|J_1\cdots J_n|\geq (p-1)!^n.$$ However one clearly has: $$|I_i(\alpha_k)| \leq |\alpha_k|e^{|\alpha_k|}F_i(|\alpha_k|),$$ where $F_i$ is the polynomial whose coefficients are the absolute values of those of $f_i$ (this follows directly from the definition of $I_i(s)$). Thus $$|J_i|\leq \sum_{k=1}^n \left |\beta_k\alpha_k \right |e^{|\alpha_k|}F_i \left ( \left |\alpha_k \right| \right )$$ and so by the construction of the $f_i$'s we have $|J_1\cdots J_n|\le C^p$ for a sufficiently large $C$ independent of $p$, which contradicts the previous inequality. This proves Lemma A. ∎
Lemma B.
If $b(1),\ldots,b(n)$ are integers and $γ(1),\ldots,γ(n)$, are distinct algebraic numbers, then$$b(1)e^{\gamma(1)}+\cdots+ b(n)e^{\gamma(n)} = 0$$ has only the trivial solution $b(i)=0$ for all $i = 1, \dots, n.$
Proof of Lemma B: Assuming $b(1)e^{\gamma(1)}+\cdots+ b(n)e^{\gamma(n)}= 0,$ we will derive a contradiction, thus proving Lemma B.
Let us choose a polynomial with integer coefficients which vanishes on all the $\gamma(k)$'s and let $\gamma(1),\ldots,\gamma(n),\gamma(n+1),\ldots,\gamma(N)$ be all its distinct roots. Let $b(n+1)=\ldots=b(N)=0$.
The polynomial$$P(x_1,\dots,x_N)=\prod_{\sigma\in S_N}(b(1) x_{\sigma(1)}+\cdots+b(N) x_{\sigma(N)})$$vanishes at $(e^{\gamma(1)},\dots,e^{\gamma(N)})$ by assumption. Since the product is symmetric, for any $\tau\in S_N$ the monomials $x_{\tau(1)}^{h_1}\cdots x_{\tau(N)}^{h_N}$ and $x_1^{h_1}\cdots x_N^{h_N}$ have the same coefficient in the expansion of $P$.
Thus, expanding $P(e^{\gamma(1)},\dots,e^{\gamma(N)})$ accordingly and grouping the terms with the same exponent, we see that the resulting exponents $h_1\gamma(1)+\dots+h_N\gamma(N)$ form a complete set of conjugates and, if two terms have conjugate exponents, they are multiplied by the same coefficient.
So we are in the situation of Lemma A. To reach a contradiction it suffices to see that at least one of the coefficients is non-zero. This is seen by equipping $C$ with the lexicographic order and by choosing for each factor in the product the term with non-zero coefficient which has maximum exponent according to this ordering: the product of these terms has non-zero coefficient in the expansion and does not get simplified by any other term. This proves Lemma B. ∎

Final step

We turn now to prove the theorem: Let $a(1),\ldots,a(n)$ be non-zero algebraic numbers, and $α(1),\ldots,α(n)$ distinct algebraic numbers. Then let us assume that: $$a(1)e^{\alpha(1)}+\cdots + a(n)e^{\alpha(n)} = 0.$$ We will show that this leads to contradiction and thus prove the theorem. The proof is very similar to that of Lemma B, except that this time the choices are made over the $a(i)$'s:
For every $i∈\{1,\ldots,n\}$, $a(i)$ is algebraic, so it is a root of an irreducible polynomial with integer coefficients of degree $d(i)$. Let us denote the distinct roots of this polynomial $a(i)_1,\ldots,a(i)_{d(i)}$, with $a(i)_1=a(i)$.
Let $S$ be the functions $σ$ which choose one element from each of the sequences $(1,\ldots,d(1)),(1,\ldots,d(2)),\ldots,(1,\ldots,d(n))$, so that for every $1≤i≤n,σ(i)$ is an integer between 1 and $d(i)$. We form the polynomial in the variables $x_{11},\dots,x_{1d(1)},\dots,x_{n1},\dots,x_{nd(n)},y_1,\dots,y_n$ $$Q(x_{11},\dots,x_{nd(n)},y_1,\dots,y_n)=\prod\nolimits_{\sigma\in S}\left(x_{1\sigma(1)}y_1+\dots+x_{n\sigma(n)}y_n\right).$$ Since the product is over all the possible choice functions σ, $Q$ is symmetric in $x_{i1},\dots,x_{id(i)}$ for every $i$. Therefore $Q$ is a polynomial with integer coefficients in elementary symmetric polynomials of the above variables, for every $i$, and in the variables $y_i$. Each of the latter symmetric polynomials is a rational number when evaluated in $a(i)_1,\dots,a(i)_{d(i)}$. The evaluated polynomial $Q(a(1)_1,\dots,a(n)_{d(n)},e^{\alpha(1)},\dots,e^{\alpha(n)})$ vanishes because one of the choices is just $σ(i)=1$ for all $i$, for which the corresponding factor vanishes according to our assumption above. Thus, the evaluated polynomial is a sum of the form $$b(1)e^{\beta(1)}+ b(2)e^{\beta(2)}+ \cdots + b(N)e^{\beta(N)}= 0,$$ where we already grouped the terms with the same exponent. So in the left-hand side we have distinct values $β(1), ..., β(N)$, each of which is still algebraic (being a sum of algebraic numbers) and coefficients $b(1),\dots,b(N)\in\mathbb Q$. The sum is nontrivial: if $\alpha(i)$ is maximal in the lexicographic order, the coefficient of $e^{|S|\alpha(i)}$ is just a product of $a(i)_j$'s (with possible repetitions), which is non-zero. By multiplying the equation with an appropriate integer factor, we get an identical equation except that now $b(1),\ldots,b(N)$ are all integers. Therefore, according to Lemma B, the equality cannot hold, and we are led to a contradiction which completes the proof. ∎
Note that Lemma A is sufficient to prove that $e$ is irrational, since otherwise we may write $e=p/q$, where both $p$ and $q$ are non-zero integers, but by Lemma A we would have $qe−p$ ≠ 0, which is a contradiction. Lemma A also suffices to prove that $\pi$ is irrational, since otherwise we may write $\pi=kn$, where both $k$ and $n$ are integers) and then $±i\pi$ are the roots of $n^2x^2+k^2=0$; thus $2−1−1=2e^0+e^{i\pi}+e^{−i\pi}≠0$; but this is false.
Similarly, Lemma B is sufficient to prove that $e$ is transcendental, since Lemma B says that if $a_0,\ldots,a_n$ are integers not all of which are zero, then $$a_ne^n+\cdots+a_0e^0\ne 0.$$ Lemma B also suffices to prove that $\pi$ is transcendental, since otherwise we would have $1+e^{i\pi}≠0$.

Equivalence of the two statements

Baker's formulation of the theorem clearly implies the first formulation. Indeed, if $\alpha(1),\ldots,\alpha(n)$ are algebraic numbers that are linearly independent over $\mathbb Q$, and $$P(x_1, \ldots, x_n)= \sum b_{i_1,\ldots, i_n} x_1^{i_1}\cdots x_n^{i_n}$$ is a polynomial with rational coefficients, then we have $$P\left(e^{\alpha(1)},\dots,e^{\alpha(n)}\right)= \sum b_{i_1,\dots,i_n} e^{i_1 \alpha(1) + \cdots + i_n \alpha(n)},$$ and since $\alpha(1),\ldots,\alpha(n)$ are algebraic numbers which are linearly independent over the rationals, the numbers $i_1 \alpha(1) + \cdots + i_n \alpha(n)$ are algebraic and they are distinct for distinct $n$-tuples $(i_1,\dots,i_n)$. So from Baker's formulation of the theorem we get $ b_{i_1,\ldots,i_n}=0$ for all $n$-tuples $(i_1,\dots,i_n)$. Now assume that the first formulation of the theorem holds. For $n=1$ Baker's formulation is trivial, so let us assume that $n>1$, and let $a(1),\ldots,a(n)$ be non-zero algebraic numbers, and $\alpha(1),\ldots,\alpha(n)$ distinct algebraic numbers such that: $$a(1)e^{\alpha(1)} + \cdots + a(n)e^{\alpha(n)} = 0.$$ As seen in the previous section, and with the same notation used there, the value of the polynomial $$Q(x_{11},\ldots,x_{nd(n)},y_1,\dots,y_n)=\prod\nolimits_{\sigma\in S}\left(x_{1\sigma(1)}y_1+\dots+x_{n\sigma(n)}y_n\right),$$ at $$\left (a(1)_1,\ldots,a(n)_{d(n)},e^{\alpha(1)},\ldots,e^{\alpha(n)} \right)$$ has an expression of the form $$b(1)e^{\beta(1)}+ b(2)e^{\beta(2)}+ \cdots + b(M)e^{\beta(M)}= 0,$$ where we have grouped the exponentials having the same exponent. Here, as proved above, $b(1),\ldots, b(M)$ are rational numbers, not all equal to zero, and each exponent $\beta(m)$ is a linear combination of $\alpha(i)$ with integer coefficients. Then, since $n>1$ and $\alpha(1),\ldots,\alpha(n)$ are pairwise distinct, the $\mathbb Q$-vector subspace $V$ of $\mathbb C$ generated by $\alpha(1),\ldots,\alpha(n)$ is not trivial and we can pick $\alpha(i_1),\ldots,\alpha(i_k)$ to form a basis for $V.$ For each $m=1,\dots,M$, we have \begin{align} \beta(m) = q_{m,1} \alpha(i_1) + \cdots + q_{m,k} \alpha(i_k), && q_{m,j} = \frac{c_{m,j}}{d_{m,j}}; \qquad c_{m,j}, d_{m,j} \in \mathbb Z. \end{align} For each $j=1,\ldots, k,$ let $d_j$ be the least common multiple of all the $d_{m,j}$ for $m=1,\ldots, M$, and put $v_j = \tfrac{1}{d_j} \alpha(i_j)$. Then $v_1,\ldots,v_k $ are algebraic numbers, they form a basis of $V$, and each $\beta(m)$ is a linear combination of the $v_j$ with integer coefficients. By multiplying the relation $b(1)e^{\beta(1)}+ b(2)e^{\beta(2)}+ \cdots + b(M)e^{\beta(M)}= 0,$ by $e^{N(v_1+ \cdots + v_k)}$, where $N$ is a large enough positive integer, we get a non-trivial algebraic relation with rational coefficients connecting $e^{v_1},\cdots,e^{v_k}$, against the first formulation of the theorem.