"Now we’ve found nature brings out more social feelings, more value for community and close relationships. People are more caring when they’re around nature."
Seems like they're telling us that next time we feel like telling someone to "take a hike!" we should take one ourselves.
I typically experience nature with a day in the park or a weekend trip out of the city for some hiking or camping. I've always felt that getting outside was a way to be nice to myself, since I feel calmer and more relaxed after. But apparently it makes me nicer to other people too! Do you spend time in nature? Has it made you feel like a nicer person?