Interesting question. Shannon entropy is not always representative of brute force strength, but it is representative in this case.
TL;DR: Either generate all strings with ratio of 0's to 1's approximately 1/3 and try those, and then other strings "near them". This is because these sequences have the maximal probability of occurence under the binomial distribution with $Pr[bit=1]=3/4:=p.$ When picking sequences near these flipping a $0$ to a $1$ has a smaller probability penalty than doing the opposite so try those first. Scroll to the end for a description of the attack.
Edited to add reference: See the Asymptotic Equipartition Property which states in its simplest and weakest form that for any finite alphabet (here {0,1}) the set defined in my answer has probability $1-\epsilon$. Cover and Thomas, Elements of Information Theory, 2nd Edition, Theorem 3.1.2 is a good place to start. Summary on Page 64. This book can be found online. There are also stronger versions proved where convergence is not just in probability but almost sure. Keywords: "Method of Types", under more assumptions (bounded moments, e.g., which apply here) which work for conditional distributions as well.
Background: Note that a better general measure of the strength of a key against brute force attacks is the so-called guessing entropy or Renyi entropy $H_{1/2}$ of order $1/2.$ See this question and this one for some details: in summary for certain (not necessarily fractal) type of distributions (see this paper for details) the Shannon entropy is $H$ but the key can be brute forced in much smaller than $2^H$ guesses on average. One would want to use reasonable key distributions, where $p_{max}$ was upper bounded (equivalently the min-entropy was lower bounded). In an excellent comment one key value is given as $1/2$ but this is not a good distribution to draw from for the cryptographer, but still the point about high Shannon entropy not being enough is valid.
Detailed Answer: But for your specific case independent biased bits this phenomenon is not present, and ou are on the right track, by trying random keys with the same biased distribution. Given the distribution $P=(1/4,3/4)$ almost all the probable sequences (keys) which occur under this distribution are from the typical set a concept from Shannon entropy based information theory. These are all the sequences whose empirical distribution of $0$ and $1$ are "close enough" to the ratio $(1/4,3/4).$
Let your distribution be denoted Bernoulli$-1/4$ or $B(0.25)$ for $n=49$ independent bits. The definition of typical set of $n$ i.i.d variables along with some $\epsilon > 0$ is basically $$A_{\epsilon}^{(n)} = \{(x_1, ..., x_n): 2^{-n[H(B(0.25)) + \epsilon]} \leq p(x_1, ..., x_n) \leq 2^{-n[H(B(0.25)) - \epsilon]}\}$$ and Shannon tells us that with overwhelming probability all sequences from this biased source have probability in the interval given in the inequalities, i.e., belong to the typical set. While $n=49$ is not that large, but you get probability 0.834 or so under AEP if you sum probabilities from 7 1's to 15 1's: $$ \sum_{k=7}^{15} \binom{49}{k} (1/4)^k(3/4)^{49-k}\approx 0.834 $$
Here $H(B(0.25))$ is the entropy of a Bernoulli variable with parameter $0.25$ which is $\approx0.81$. So, $2^{-n(0.81 - \epsilon)} \approx 0.57^{n}$. Take $n=49$ to get $0.57^{49}\approx 1.1\times 10^{-12}.$ This is the probability of $p(x_1, ..., x_n)$ of a typical sequence and thus the number of typical sequences is approximately its reciprocal which is approximately $2^{39.7}.$
Attack: All computations approximate
So you need to first collect all the strings with about a quarter of coordinates (say 12, or 13) equal to zeroes and the rest equal to ones. Try those first, call this set $S_0$. Since those have the highest probability of occuring. Then you could try strings at Hamming distance 1 from those, etc. Let's compute how many such sequences are there in $S_0.$ There are $$ \binom{49}{12}+\binom{49}{13}\approx 3.54 \times 10^{11}\approx 2^{38.36}, $$ in this set $S_0.$ This is off by a multiplicative factor $2^{38.36-39.7}=2^{-1.34}$ from covering the typical set which contains approximately $2^{39.7}$ strings. Just going one Hamming distance beyond will cover almost all sequences in the typical set.
If this deterministic attack is too expensive, just randomly generate strings with the same Bernoulli distribution and you will on average discover the key with not much worse performance.