Why do you think Americans are so "tight" about men's penises, bodies and male interaction. If you watch television shows from other countries you see guys in reality shows walking around with erect dicks, athletes exposing themselves or having wardrobe malfunctions, nude male actors, etc. Sometimes in Latin countries they even have male strippers on TV during the day, guys walk around in thongs, and you can even catch them laying on the beach in the nude sometimes. In some countries men can be buddy buddy hug, kiss, grab each, play around and be perfectly straight and no one bothers them or really gives it a second thought. But over here if someone sees a guy in public with no underwear on it's like a huge traffic stopping event. When black track stars are shown on TV they even pan the camera up for front shots so that you won't see the good bits in motion. I didn't know that their legs were on their shoulders. Very rarely do you see soccer matches on regular television where guys are notorious for wardrobe malfunctions and showing bouncing bulges or wrestling (real wrestling not that fake shit) where guys are always popping boners. And if a male actor shows frontal nudity in a movie (no matter how brief) people debate about whether to leave it in the film for months. But yet we see endless ads for ED drugs and plenty of implied or simulated sex in television and movie scenes. And also why is it now that if two men spend too much time together or seem the least bit chummy, everyone thinks that they are gay. Two men can have a close relationship and not be banging each other. The comradery between two male friends can be one of the most rewarding experiences. Yet many guys are missing out on that now because they don't want people thinking they are gay. Why do you think we are so close minded and up tight over here while large portions of the rest of the world seem to careless?