I am having some difficulties in producing the correct code for a simple
binary to decimal conversion program. The arithmetic I want to use is the
doubling method - so if I wanted to get the decimal equivalent of 1001, I
would start with multiplying 0 * 2 and adding the left most digit. I would
then take the sum of 1 and multiply it by 2 and add the next digit to that
product and so on and so forth until I come to an end sum of 9.
I seem to come up with something like this:
binnum = raw_input("Please enter a binary number: ")
for i in range(0, len(binum), 1):
item = "0"
if i < len(binum) - 1:
item = binum[i + 1]
binsum = binsum * int(item) * 2 + binsum + int(binum[i])
print "\nThe binary number ", binum, " you entered converts to", binsum, "
in decimal."
I can't really figure out what is going wrong here. Please help me - this
has been driving me insane.
_______________________________________________
Tutor maillist - [email protected]
http://mail.python.org/mailman/listinfo/tutor