Decimals are hard.
What would we even want the student to do here if he’s working in decimal? Like, how do standard multiplication algorithms handle something like a repeating digit?
That’s what I’m getting out of this mistake right now: the deviousness of decimal representation, and the way it can obscure numerical properties.
How about you? What do you make of all this?