First, the concept of (Shannon) entropy applies only to random sources... not to given "files" (i.e., deterministic data). Broadly speking, we often consider such a file a realization of some ideal random source. In this case, we might consider that the first one was generated by a white noise source, and the second by a source that outputs zeroes with probability (almost?) one.
Assuming the above, then, yes, the first one has the highest entropy rate ( 8 bits per byte), while the second has the lowest (0 bits per byte).
less entropy which means more information. Is that correct?
No, of course the second one has less information (actually, no information at all).
Low entropy (in the extreme, a deterministic source) has low information content.
If you have trouble digesting the idea that a "pure noise" has maximum information content, you might read this question.