I loved the first few seasons, but around the time that Negan came, it went downhill. The main thing I liked about the show was the emphasis on humans vs walkers. But now it's mainly humans vs humans.
I feel like the show got sidetracked, and there are so many unexplored avenues. e.g. I would be interested to see if they could connect with people overseas. If they can build things up again and create new towns etc. And gradually restore humanity etc.
I couldn't be MORE biased on this show. TWD is my favorite comic book, I read the books, and LOVE this show. As a Resident Evil game fan I am frightened to play TWD games that I will become addicted! Great acting, tense atmosphere, this is a creepy, gritty and utterly fantastic show in every way, and similar enough to feel identical to the other mediums, but different characters and stories make each seem fresh and independent of each other.
If you like a show where anything can happen like Game Of Thrones (without the swearing and nudity) and aren't afraid to say goodbye to your favorite characters on a regular basis, strap in for survival horror at it's finest.
I believe that The Walking Dead has most certainly gone downhill. These past few seasons have been awful! I looked forward to watching TWD every Sunday night! Not knowing what is going to happen! But now it's been all about PEOPLE! NOT WALKERS! I mean, it is called "The Walking Dead" for a reason! I really think that they should stop filming TWD and just merge it with FEAR THE WALKING DEAD! At least FTWD has better actors/actresses and a much better plot!
When this show begun it was fantastic.Now it´s just the ashes of something who was great...after season 6 it became the same old stuff over and over again, and for me it´s time to receive a bite and let this show RIP.It was good ,while it lasted.