Unless I'm mistaken I would say that it's the other way around, Morse code is more like a human readable machine language expressed in binary because the 26 character alphabet is expressed in different binary values, much like ASCII.
For those that need help in deciphering the title text, there is a super easy trick:
Take the last 5 bits of each byte, convert them from binary to decimal, and that letter of the alphabet (starting from one) is the letter represented by the ASCII code.
To convert binary numbers to decimal, just add the place values of a the 1s together. For 5 bits, the place values are 16, 8, 4, 2, 1. So 10110 becomes 16+4+2 which is equal to 22.
01001101: last 5 bits are 01101, in decimal, that is 8+4+1=13, and the 13th letter of the alphabet is "m"
01100101, becomes 00101, then 4+1=5, and the 5th letter of the alphabet is 'e'
The last 2 are just repeats of those, so the post body reads "meme" (if the third bit is zero, the letter is capital, so it should really be "Meme")
This is a weird meme to me. Have you ever made something like a simple accumulator machine out of logic gates, OP? You literally just program them in binary, although usually the instructions are expressed in hexadecimal. You make your own instruction set. When we did ours in Compsci foundations I just decided that 0x06 was going to be my jump at negative instruction. I could have wired the logic so that instruction was at a different value, it's literally possible to make your own instruction set, then your own assembly language, and then your own compiler, and your own programming language. People, mostly women at first, did this for every new computer their institution built at first.