There is Only “One Faith” – Who will “Fight the Good Fight” for “The Faith” in these Evil Last Days?
What does the Bible teach about Christianity? The Bible doesn't teach anything at all about Christianity. As I have previously written, the Christian Religion was defined and adopted in the 4th Century by the Roman State, under the Emperor Constantine. The Romans were mostly a polytheistic culture believing in most of the named Greek gods, until they made Christianity the official State Religion, and converted all of their Greek, "pagan" gods and holidays, into "Christian" ones. What the New Testament writers referred to, instead of the misconception and false belief that they were referring to a new religion called "Christianity," is what they referred to as "The Faith." What is "The faith"?