Thursday, October 17, 2019

Long-married couples recall their wedding day: the influence of collaboration and gender on autobiographical memory recall

Long-married couples recall their wedding day: the influence of collaboration and gender on autobiographical memory recall. Azriel Grysman et al. Memory, Oct 15 2019. https://doi.org/10.1080/09658211.2019.1673428

ABSTRACT: The current study examined the influence of collaboration, expertise, and communication on autobiographical memory, by considering gender differences in recall and how they may influence the products and processes of remembering when male-female couples recall events together. Thirty-nine long-married, male-female couples recalled their memories of their wedding day. In Session 1, they recalled it individually for an experimenter. One week later, in Session 2, they recalled the same event jointly as a collaborative pair. Women reported more details, especially episodic details, than men across both sessions. Notably, collaborative recall included many new details that neither spouse had recalled individually. Exploratory analyses suggest that women were less influenced by collaboration than were men: women’s communication behaviours influenced men’s recall, but the reverse was not found for men’s communication. Additionally, when couples’ individual recall was more similar in content, men were more likely to decrease their contribution to the collaborative session. We consider these findings in light of transactive memory theory, in which joint meta-memory and the distribution of expertise influence the processes and products of recall in the interdependent system of a couple who extensively share their autobiographical memories.

KEYWORDS: Autobiographical memory, memory collaboration, gender, transactive memory

Wednesday, October 16, 2019

In a cohort study of 82 232 participants, personality traits in adolescence—a time when dementia pathology is unlikely to be present—were a factor associated with incident dementia almost 5 decades later

Association Between High School Personality Phenotype and Dementia 54 Years Later in Results From a National US Sample. Benjamin P. Chapman et al. JAMA Psychiatry, October 16, 2019. doi:10.1001/jamapsychiatry.2019.3120.

Key Points
Question  Are maladaptive personality traits true risk factors for dementia or merely early expressions of underlying neuropathologic changes?

Findings  In a cohort study of 82 232 participants, personality traits in adolescence—a time when dementia pathology is unlikely to be present—were a factor associated with incident dementia almost 5 decades later in a national US cohort. Calm and mature adolescents were less likely to develop dementia, and this risk reduction was significantly more pronounced with higher socioeconomic status.

Meaning  This study’s findings suggest that maladaptive personality traits decades earlier may be independent risk factors for dementia by age 70 years.

Abstract
Importance  Personality phenotype has been associated with subsequent dementia in studies of older adults. However, neuropathologic changes often precede cognitive symptoms by many years and may affect personality itself. Therefore, it is unclear whether supposed dementia-prone personality profiles (high neuroticism and low conscientiousness) are true risk factors or merely reflections of preexisting disease.

Objectives  To examine whether personality during adolescence—a time when preclinical dementia pathology is unlikely to be present—confers risk of dementia in later life and to test whether associations could be accounted for by health factors in adolescence or differed across socioeconomic status (SES).

Design, Setting, and Participants  Cohort study in the United States. Participants were members of Project Talent, a national sample of high school students in 1960. Individuals were identified who received a dementia-associated International Classification of Diseases, Ninth Revision (ICD-9) diagnosis code during any year between 2011 and 2013. The dates of our analysis were March 2018 to May 2019.

Exposures  Ten personality traits were measured by the 150-item Project Talent Personality Inventory. Socioeconomic status was measured by a composite based on parental educational level, income, occupation, and property ownership. Participants were also surveyed on demographic factors and height and weight.

Main Outcomes and Measures  Medicare records were collected, with dementia diagnoses in the period of 2011 to 2013 classified according to the US Centers for Medicare & Medicaid Services ICD-9–based algorithm. Cox proportional hazards regression models estimated the relative risk of dementia based on the 10 personality traits, testing interactions with SES and adjusting for demographic confounders.

Results  The sample of 82 232 participants was 50.1% female, with a mean (SD) age of 15.8 (1.7) years at baseline and 69.5 (1.2) years at follow-up. Lower risk of dementia was associated with higher levels of vigor (hazard ratio for 1 SD, 0.93; 95% CI, 0.90-0.97; P < .001). Calm and maturity showed protective associations with later dementia that increased with SES. At 1 SD of SES, calm showed a hazard ratio of 0.89 (95% CI, 0.84-0.95; P < .001 for the interaction) and maturity showed a hazard ratio of 0.90 (95% CI, 0.85-0.96; P = .001 for the interaction).

Conclusions and Relevance  This study’s findings suggest that the adolescent personality traits associated with later-life dementia are similar to those observed in studies of older persons. Moreover, the reduction in dementia risk associated with a calm and mature adolescent phenotype may be greater at higher levels of SES. Personality phenotype may be a true independent risk factor for dementia by age 70 years, preceding it by almost 5 decades and interacting with adolescent socioeconomic conditions.

Wild bonobos consume meat at higher rates than previously thought; the individual controlling the carcass frequently resisted sharing, and aggressive attempts to take the carcass were observed

New Observations of Meat Eating and Sharing in Wild Bonobos (Pan paniscus) at Iyema, Lomako Forest Reserve, Democratic Republic of the Congo. Wakefield M.L. · Hickmott A.J. · Brand C.M. · Takaoka I.Y. · Meador L.M. · Waller M.T. · White F. Folia Primatol 2019;90:179–189. https://doi.org/10.1159/000496026

Abstract: Bonobos (Pan paniscus) consume a variety of vertebrates, although direct observations remain relatively rare compared to chimpanzees (Pan troglodytes). We report the first direct observations of meat eating and sharing among bonobos at Iyema, Lomako Forest, Democratic Republic of Congo. We collected meat consumption data ad libitum from June to November 2017 over 176.5 observation hours and conducted monthly censuses to measure the abundance of potential prey species. We observed 3 occasions of duiker consumption and found indirect evidence of meat consumption twice (n = 5). We identified the prey species as Weyn’s duiker (Cephalophus weynsi) in all 4 cases that we saw the carcass. This species was the most abundant duiker species at Iyema, but other potential prey species were also available. Meat sharing was observed or inferred during all 3 observations. However, the individual controlling the carcass frequently resisted sharing, and aggressive attempts to take the carcass were observed. This report contributes to a growing body of data suggesting that wild bonobos consume meat at higher rates than previously thought, female control of carcasses is frequent but not exclusive, and meat sharing in bonobos is primarily passive but not without aggression.

Keywords: Behavioral diversityFaunivoryFood sharingPrey preferenceFemale control

Check also Faux-nobo: “Naked Bonobo” demolishes myth of sexy, egalitarian bonobos. Edward Clint. Oct 9, 2017. https://www.bipartisanalliance.com/2017/10/naked-bonobo-demolishes-myth-of-sexy.html

And Bonobos Prefer Individuals that Hinder Others over Those that Help. Christopher Krupenye, Brian Hare. Current Biology, https://www.bipartisanalliance.com/2018/01/whereas-humans-already-prefer-helpers.html

Studies in Repetitive Transcranial Magnetic Stimulation: The somewhat high frequency of “positive” results seems spurious and may reflect bias; caution is warranted in accepting rTMS as an established treatment

Amad A, Jardri R, Rousseau C, Larochelle Y, Ioannidis J, P, A, Naudet F: Excess Significance Bias in Repetitive Transcranial Magnetic Stimulation Literature for Neuropsychiatric Disorders. Psychother Psychosom 2019. https://doi.org/10.1159/000502805

Abstract
Introduction: Repetitive transcranial magnetic stimulation (rTMS) has been widely tested and promoted for use in multiple neuropsychiatric conditions, but as for many other medical devices, some gaps may exist in the literature and the evidence base for the clinical efficacy of rTMS remains under debate.
Objective: We aimed to test for an excess number of statistically significant results in the literature on the therapeutic efficacy of rTMS across a wide range of meta-analyses and to characterize the power of studies included in these meta-analyses.
Methods: Based on power calculations, we computed the expected number of “positive” datasets for a medium effect size (standardized mean difference, SMD = 0.30) and compared it with the number of observed “positive” datasets. Sensitivity analyses considered small (SMD = 0.20), modest (SMD = 0.50), and large (SMD = 0.80) effect sizes.
Results: A total of 14 meta-analyses with 228 datasets (110 for neurological disorders and 118 for psychiatric disorders) were assessed. For SMD = 0.3, the number of observed “positive” studies (n = 94) was larger than expected (n = 35). We found evidence for an excess of significant findings overall (p < 0.0001) and in 8/14 meta-analyses. Evidence for an excess of significant findings was also observed for SMD = 0.5 for neurological disorders. Of the 228 datasets, 0 (0%), 0 (0%), 3 (1%), and 53 (23%) had a power >0.80, respectively, for SMDs of 0.30, 0.20, 0.50, and 0.80.
Conclusion: Most studies in the rTMS literature are underpowered. This results in fragmentation and waste of research efforts. The somewhat high frequency of “positive” results seems spurious and may reflect bias. Caution is warranted in accepting rTMS as an established treatment for neuropsychiatric conditions.

Keywords: Repetitive transcranial magnetic stimulationRandomized controlled trialMeta-analysisExcess significance

The intensity of polarization on Twitter varies greatly from one country to another

Context matters: political polarization on Twitter from a comparative perspective. Aleksandra Urman. Media, Culture & Society, October 15, 2019. https://doi.org/10.1177/0163443719876541

Abstract: This article explores the issue of political polarization on social media. It shows that the intensity of polarization on Twitter varies greatly from one country to another. The analysis is performed using network-analytic audience duplication approach and is based on the data about the followers of the political parties’ Twitter accounts in 16 democratic countries. Based on the topology of the audience duplication graphs, the political Twitterspheres of the countries are classified as perfectly integrated, integrated, mixed, polarized and perfectly polarized. Explorative analysis shows that polarization is the highest in two-party systems with plurality electoral rules and the lowest in multi-party systems with proportional voting. The findings help explain the discrepancies in the results of previous studies into polarization on social media. The results of the study indicate that extrapolation of the findings from single-case studies on the topic is impossible in most cases, suggesting that more comparative studies on the matter are necessary to better understand the subject and get generalizable results.

Keywords audience duplication, comparative analysis, network analysis, political polarization, social media, Twitter


Beliefs About Human Intelligence in a Sample of Teachers and Non-teachers: There are conflicts between currently accepted intelligence theory and the subjects' beliefs

Warne, Russell T., and Jared Z. Burton. 2019. “Beliefs About Human Intelligence in a Sample of Teachers and Non-teachers.” PsyArXiv. July 17. doi:10.31234/osf.io/uctxp

Abstract: Research in educational psychology consistently finds a relationship between intelligence and academic performance. However, in recent decades, educational fields, including gifted education, have resisted intelligence research, and there are some experts who argue that intelligence tests should not be used in identifying giftedness. Hoping to better understand this resistance to intelligence research, we created a survey of beliefs about intelligence and administered it online to a sample of the general public and a sample of teachers. We found that there are conflicts between currently accepted intelligence theory and beliefs from the American public and teachers, which has important consequences on gifted education, educational policy and the effectiveness of interventions.

Empathic contagious pain and consolation in laboratory rodents: Rodents also have the ability to feel, recognize, understand and share the other’s distressing states, and allolick injuries and allogroom bodies

Empathic contagious pain and consolation in laboratory rodents: species and sex comparisons. Rui Du et al. bioRxiv, October 15, 2019. https://doi.org/10.1101/745299

Abstract: Laboratory rodents are gregarious in nature and have a feeling of empathy when witnessing a familiar conspecific in pain. The rodent observers express two levels of empathic responses: observational contagious pain (OCP) and consolation. Here we examined the sex and species difference of OCP and consolation in male and female mice and rats. We observed no species difference in both OCP and consolation, but significant species difference in general social (allo-mouth and/or allo-tail sniffing) and non-social (self-grooming) behaviors. For sex difference, male mouse observers showed more allolicking and allogrooming behaviors toward a familiar conspecific in pain during and longer time increase in pain sensitivity after the PDSI than female mouse observers. However, no sex difference was observed in rats. Our results highlighted an evolutionary view of empathy that social animals including rodents also have the ability to feel, recognize, understand and share the other’s distressing states.

Online Vigilance and Affective Well-being in Everyday Life: Thinking about smartphone-mediated social interactions (i.e., the salience dimension of online vigilance) was negatively related to affective well-being

Johannes, Niklas, Adrian Meier, Leonard Reinecke, Saara Ehlert, Dinda N. Setiawan, Nicole Walasek, Tobias Dienlin, et al. 2019. “The Relationship Between Online Vigilance and Affective Well-being in Everyday Life: Combining Smartphone Logging with Experience Sampling.” PsyArXiv. October 15. doi:10.31234/osf.io/t3wc2

Abstract: Through communication technology, users find themselves constantly connected to others to such an extent that they routinely develop a mindset of connectedness. This mindset has been defined as online vigilance. Although there is a large body of research on media use and well-being, the question of how online vigilance impacts well-being remains unanswered. In this preregistered study, we combine experience sampling and smartphone logging to address the relation of online vigilance and affective well-being in everyday life. Seventy-five Android users answered eight daily surveys over five days (N = 1615) whilst having their smartphone use logged. Thinking about smartphone-mediated social interactions (i.e., the salience dimension of online vigilance) was negatively related to affective well-being. However, it was far more important whether those thoughts were positive or negative. No other dimension of online vigilance was robustly related to affective well-being. Taken together, our results suggest that online vigilance does not pose a serious threat to affective well-being in everyday life.

The pay premium for high‐potential women: They don't report higher levels of pay satisfaction, suggesting that high‐potential women did not perceive their pay premium to be an inequitable advantage

The pay premium for high‐potential women: A constructive replication and refinement. George F. Dreher, Nancy M. Carter, Terry Dworkin. Personnel Psychology, September 10 2019. https://doi.org/10.1111/peps.12357

Abstract> In this constructive replication, we revisit a provocative study by Leslie, Manchester, and Dahm (2017). They found that gender and being designated a high‐potential employee interacted in accounting for pay and that this resulted in a reversal in the commonly observed gender pay gap favoring men. Our primary aim was to examine important boundary conditions associated with their work by (a) conducting a study using a sample that would better generalize across industries and to individuals who aspire to reach senior management, (b) adding critical control variables to the statistical models used in the pay equation, and (c) by introducing a different conceptualization of the high‐potential construct. Also, to better understand the consequences of their study, we considered an additional dependent variable that addressed pay satisfaction. Even after making these model additions, the gender by high‐potential interaction term was significant—ruling out four plausible third‐variable explanations for the Leslie et al. finding. Moreover, these confirming results were observed using a sample that represented individuals employed in a wide range of industries, who had the educational backgrounds, career histories, and motivational states typically required of candidates competing for senior executive roles. Furthermore, high‐potential women did not report higher levels of pay satisfaction, suggesting that high‐potential women did not perceive their pay premium to be an inequitable advantage and that there may be limited positive return associated with using a pay premium to retain high‐potential talent.

Tuesday, October 15, 2019

Guppies: Brain size affects responsiveness in mating behavior to variation in predation pressure and sex‐ratio

Brain size affects responsiveness in mating behavior to variation in predation pressure and sex‐ratio. Alberto Corral‐López  Maksym Romensky  Alexander Kotrschal  Severine D. Buechel  Niclas Kolm. Journal of Evolutionary Biology, October 14 2019. https://doi.org/10.1111/jeb.13556

Abstract: Despite ongoing advances in sexual selection theory, the evolution of mating decisions remains enigmatic. Cognitive processes often require simultaneous processing of multiple sources of information from environmental and social cues. However, little experimental data exist on how cognitive ability affects such fitness‐associated aspects of behavior. Using advanced tracking techniques, we studied mating behaviors of guppies artificially selected for divergence in relative brain size, with known differences in cognitive ability, when predation threat and sex‐ratio was varied. In females, we found a general increase in copulation behavior in when the sex‐ratio was female biased, but only large‐brained females responded with greater willingness to copulate under a low predation threat. In males, we found that small‐brained individuals courted more intensively and displayed more aggressive behaviors than large‐brained individuals. However, there were no differences in female response to males with different brain size. These results provide further evidence of a role for female brain size in optimal decision‐making in a mating context. In addition, our results indicate that brain size may affect mating display skill in male guppies. We suggest that it is important to consider the association between brain size, cognitive ability and sexual behavior when studying how morphological and behavioral traits evolve in wild populations.

7 types of sugaring: Sugar prostitution, compensated dating, compensated companionship, sugar dating, sugar friendships, sugar friendships with benefits, & pragmatic love

“It’s Its Own Thing”: A Typology of Interpersonal Sugar Relationship Scripts. Maren T. Scull. Sociological Perspectives, September 16, 2019. https://doi.org/10.1177/0731121419875115

Abstract: Although academics have focused on sugaring in various parts of the globe, sugar relationships in the United States have largely been ignored. The few studies that address these arrangements in the United States often frame them as a form of prostitution. Drawing from 48 in-depth interviews with women in the United States who have been in sugar relationships, I adopt a connected lives approach to explore the structure of these arrangements and to assess the extent to which they are a form of prostitution. Overall, I found that, although there is a dominant, subcultural relationship script that serves as a blueprint for sugar arrangements, they comprise their own unique relational package and take a variety of forms when enacted on an interpersonal level. Specifically, I identified seven types of sugar relationships, only one of which can be considered prostitution. These included sugar prostitution, compensated dating, compensated companionship, sugar dating, sugar friendships, sugar friendships with benefits, and pragmatic love.

Keywords script theory, prostitution, transactional sex, qualitative methods

Popular version: The 7 types of sugar daddy relationships. Sarah Erickson. Univ of Colorado at Denver, Oct 15 2019. https://www.eurekalert.org/pub_releases/2019-10/uocd-t7t101219.php

Red-colored male sticklebacks carry more oxidative DNA damage in muscle, testis & sperm in the peak breeding season, but the females find them more attractive

Attractive male sticklebacks carry more oxidative DNA damage in the soma and germline. Sin‐Yeon Kim  Alberto Velando. Journal of Evolutionary Biology, October 14 2019. https://doi.org/10.1111/jeb.13552

Abstract: Trade‐offs between the expression of sexual signals and the maintenance of somatic and germline tissues are expected when these depend upon the same resources. Despite the importance of sperm DNA integrity, its trade‐off with sexual signalling has rarely been explored. We experimentally tested the trade‐off between carotenoid‐based sexual colouration and oxidative DNA damage in skeletal muscle, testis and sperm by manipulating reproductive schedule (early vs. late onset of breeding) in male three‐spined sticklebacks. Oxidative DNA damage was measured as the amount of 8‐hydroxy‐2‐deoxy‐Guanosine in genomic DNA. Irrespective of the experimentally manipulated reproductive schedule, individuals investing more in red colouration showed higher levels of oxidative DNA damage in muscle, testis and sperm during the peak breeding season. Our results show that the expression of red colouration traded off against the level of oxidative DNA damage possibly due to the competing functions of carotenoids as colorants and antioxidants. Thus, female sticklebacks may risk fertility and viability of offspring by choosing redder, more deteriorated partners with decreased sperm DNA integrity. The evolution of sexual signal may be constrained by oxidative DNA damage in the soma and germline.

The Negative Intelligence–Religiosity Relation: New and Confirming Evidence

The Negative Intelligence–Religiosity Relation: New and Confirming Evidence. Miron Zuckerman at al. Personality and Social Psychology Bulletin, October 15, 2019. https://doi.org/10.1177/0146167219879122

Abstract: Zuckerman et al. (2013) conducted a meta-analysis of 63 studies that showed a negative intelligence–religiosity relation (IRR). As more studies have become available and because some of Zuckerman et al.’s (2013) conclusions have been challenged, we conducted a new meta-analysis with an updated data set of 83 studies. Confirming previous conclusions, the new analysis showed that the correlation between intelligence and religious beliefs in college and noncollege samples ranged from −.20 to −.23. There was no support for mediation of the IRR by education but there was support for partial mediation by analytic cognitive style. Thus, one possible interpretation for the IRR is that intelligent people are more likely to use analytic style (i.e., approach problems more rationally). An alternative (and less interesting) reason for the mediation is that tests of both intelligence and analytic style assess cognitive ability. Additional empirical and theoretical work is needed to resolve this issue.

Keywords intelligence, religiosity, meta-analysis, analytic thinking


Check also The Myth of the Stupid Believer: The Negative Religiousness–IQ Nexus is Not on General Intelligence (g) and is Likely a Product of the Relations Between IQ and Autism Spectrum. Edward Dutton et al. Journal of Religion and Health, Oct 5 2019. https://www.bipartisanalliance.com/2019/10/the-myth-of-stupid-believer-negative.html

The size of the stereotype threat effect that can be experienced on tests of cognitive ability in operational (real-world) scenarios such as college admissions tests and employment testing may range from negligible to small

Shewach, O. R., Sackett, P. R., & Quint, S. (2019). Stereotype threat effects in settings with features likely versus unlikely in operational test settings: A meta-analysis. Journal of Applied Psychology, OCt 2019. http://dx.doi.org/10.1037/apl0000420

Abstract: The stereotype threat literature primarily comprises lab studies, many of which involve features that would not be present in high-stakes testing settings. We meta-analyze the effect of stereotype threat on cognitive ability tests, focusing on both laboratory and operational studies with features likely to be present in high stakes settings. First, we examine the features of cognitive ability test metric, stereotype threat cue activation strength, and type of nonthreat control group, and conduct a focal analysis removing conditions that would not be present in high stakes settings. We also take into account a previously unrecognized methodological error in how data are analyzed in studies that control for scores on a prior cognitive ability test, which resulted in a biased estimate of stereotype threat. The focal sample, restricting the database to samples utilizing operational testing-relevant conditions, displayed a threat effect of d = −.14 (k = 45, N = 3,532, SDδ = .31). Second, we present a comprehensive meta-analysis of stereotype threat. Third, we examine a small subset of studies in operational test settings and studies utilizing motivational incentives, which yielded d-values ranging from .00 to −.14. Fourth, the meta-analytic database is subjected to tests of publication bias, finding nontrivial evidence for publication bias. Overall, results indicate that the size of the stereotype threat effect that can be experienced on tests of cognitive ability in operational scenarios such as college admissions tests and employment testing may range from negligible to small.

Lifespan Cognitive Reserve—A Secret to Coping With Neurodegenerative Pathology

Lifespan Cognitive Reserve—A Secret to Coping With Neurodegenerative Pathology. Sylvia Villeneuve. JAMA Neurol. 2019;76(10):1145-1146. October 2019, doi:10.1001/jamaneurol.2019.2899


Given the limited success of therapeutic interventions for Alzheimer disease, there is increased interest in understanding whether modifiable factors can help cope with or postpone the appearance of brain pathology. It is estimated that about 35% of Alzheimer risk is modifiable.1,2 Epidemiologic studies have shown that lifetime exposures to higher education, higher occupational attainment, and cognitively stimulating activities are associated with reduced risk of Alzheimer dementia.3 Autopsy studies have shown interindividual differences in the amount of brain pathology people can tolerate before manifesting cognitive impairments, and autopsied brains of about one-third of individuals who are cognitively normal meet neuropathological criteria for Alzheimer disease.4 About a decade ago, the concept of cognitive reserve was proposed to account for the discrepancy between brain pathology and cognitive status.5 The broad hypothesis was that individuals with enriched lifelong exposures would be able to better tolerate with brain pathology in late life. Many studies have investigated how one can cope with brain damage using proxies of neurodegeneration or synaptic integrity. However, the gold standard for testing the reserve hypothesis is the direct measurement of brain pathology at autopsy.

In this issue of JAMA Neurology, Xu and colleagues6 used the resources of the Rush Memory and Aging Project for an empirical test of the reserve hypothesis. They first assessed whether individuals with lifelong protective factors have a reduced risk of dementia. More importantly, they examined the influence of Alzheimer pathology on this association, thereby providing a direct test of the reserve hypothesis. The Rush Memory and Aging Project is among the largest longitudinal studies that include both comprehensive in vivo and autopsy data. The study had 1602 participants without dementia, of whom 611 had died during the study follow-up period and had autopsy data available. Of interest, the authors6 derived the reserve score by combining weighted measures of education, lifelong cognitive activity, and late-life social activity. In theory, the use of a composite score as a proxy for reserve instead of a single factor score (eg, education) should be more accurate if reserve is built on diverse lifelong experiences, as is commonly thought. Over a mean 6-year study follow-up (range, 1-20 years), one-quarter of these participants developed dementia. As expected, a higher reserve composite was associated with reduced incidence of dementia. Thus, participants in the highest tertile of reserve had about a 40% reduction in occurrence of dementia compared with those in the lowest tertile. A dose-effect association was also evident, in that individuals in the middle tertile showed about a 20% reduction in risk of dementia. Stated another way, individuals with high reserve scores experienced a delay in dementia onset of more than 7 years. Targeting lifestyle factors that enhance reserve could therefore reduce the incidence of new cases, while providing more than half a decade of cognitive health to individuals who would ultimately develop dementia.

The contribution of individual factors to dementia risk was assessed in a secondary set of analyses. Here it was interesting to note that lifestyle behaviors in midlife and late life were the main protective factors of dementia, suggesting that preventive interventions should probably be started in midlife but may also be successful if started in late life. In 2015 the results of Finnish Geriatric Intervention Study to Prevent Cognitive Impairment and Disability (FINGER),7 which involved physical exercise, nutrition, cognitive stimulation, and self-monitoring of heart health, suggested the possibility of preventing cognitive decline using a multidomain intervention among older individuals who were at risk. The FINGER model is now being adapted and implemented in North America (the Alzheimer's Association US Study to Protect Brain Health Through Lifestyle Intervention to Reduce Risk [US POINTER]), Asia (Singapore Intervention Study to Prevent Cognitive Impairment and Disability [SINGER] and the Multimodal Interventions to Delay Dementia and Disability in Rural China [MIND-CHINA]), Australia (the Maintain Your Brain [MYB] trial), and Europe (Multimodal Preventive Trial for Alzheimer’s Disease [MIND-AD]). It is estimated that a multidomain lifestyle intervention that achieves a 25% reduction in Alzheimer risk could prevent more than 3 million cases worldwide.1

Focusing on the autopsy findings, the positive association of reserve with reducing the risk of dementia onset was present even in individuals with high degrees of brain pathology. This was true for Alzheimer disease pathology, but also for gross infarcts and microscopic infarcts. No sex difference was reported for this association. A recent in vivo positron emission tomography (PET) study found that women exhibit higher tau burden than men in the preclinical phase of the disease,8 which could suggest that women can tolerate more pathology before developing cognitive impairments. Since women are at increased risk of developing Alzheimer dementia, particularly in late life, the role of sex differences in lifestyle-protective factors and associated resilience to the pathology will need further investigation.

The protective effect of lifespan cognitive reserve on the incidence of dementia in individuals with high pathology was also found when removing individuals with baseline mild cognitive impairments from the analysis. This last point is important, because individuals who are cognitively normal are the targets of preventive trials and therefore increasing effort is devoted toward a better understating of the factors that could protect this group from developing cognitive impairment. Additionally, it is possible that some reserve-associated phenomena are only quantifiable at a certain stage of the disease, stressing the need to explore reserve-associated phenomena at different disease stages. For instance, a new study9 has shown that, while high reserve was associated with a slower cognitive decline in individuals without dementia, it was associated with a faster decline in individuals with dementia. One interpretation for this unexpected finding is that when individuals with high reserve can no longer cope with the pathology, their clinical decline is faster than individuals with lower reserve because the disease stage is more advanced.5

Evidence from PET research further suggests that some lifestyle factors that can enhance cognitive reserve might also postpone manifestation of brain pathology. This new concept has been called resistance.10 Thus, while the term reserve refers to the ability to cope with AD pathology, resistance refers to the ability to avoid the pathology in the first place. For instance, enriched cognitive and physical activities have been associated with a reduced amyloid burden in individuals who are cognitively normal.11-13 It has even been proposed that protective lifestyle factors could partially offset the negative influence of the APOE ε4 allele on the risk of developing Alzheimer pathology.14,15 In their study, Xu and colleagues6 found no evidence supporting the resistance hypothesis, with individuals with high and low reserve scores exhibiting similar amounts of pathology. The results were similar when removing individuals with baseline mild cognitive impairment from the analysis or using cognitive status as an interactive term. The association between protective lifestyle factors and brain pathology, if it exists, might not be linear but rather closer to a U-shaped distribution and therefore extremely difficult to detect in cross-sectional studies. Plausibly, as soon as individuals with high reserve develop Alzheimer pathology, they are no longer resisting the pathology but are limited to coping with it.

This last point may serve as a reminder that cognitive reserve is a dynamic and therefore complex concept for study. Amyloid and tau PET imaging, even if they only give estimates of pathology, should be extremely powerful tools to quantify reserve-associated processes in real time, since they allow tracking of evolution of pathology in vivo. They can also contribute to simultaneous measurement of protective lifestyle factors, Alzheimer pathology, and cognitive performance. Even more interestingly, PET imaging can permit concurrent quantitation of the effects of preventive lifestyle interventions on Alzheimer pathology and associated cognitive decline.

The article by Dr Xu and colleagues6 provides strong evidence that individuals with higher reserve can tolerate more brain pathology and therefore experienced a reduced risk of dementia compared with others. While autopsy studies will probably remain the gold standard for testing the reserve hypothesis, their limitations in testing the dynamic of this phenomenon are obvious. New fluid, PET, and magnetic resonance imaging techniques for tracking Alzheimer and vascular pathology in vivo will certainly add to comprehension of reserve phenomena in the years to come.


Check also Association of Lifespan Cognitive Reserve Indicator With Dementia Risk in the Presence of Brain Pathologies. Hui Xu et al. JAMA Neurol. 2019;76(10):1184-1191, July 14, 2019, doi:10.1001/jamaneurol.2019.2455
Key Points

Question  Is high lifespan cognitive reserve (CR) indicator associated with a reduction in dementia risk, and how strong is this association in the presence of high brain pathologies?

Findings  In this cohort study including 1602 dementia-free older adults, high lifespan CR was associated with a decreased risk of dementia. This association was present in people with high Alzheimer disease and vascular pathologies.

Meaning  Accumulative educational and mentally stimulating activities enhancing CR throughout life might be a feasible strategy to prevent dementia, even for people with high brain pathologies.

Abstract

Importance  Evidence on the association of lifespan cognitive reserve (CR) with dementia is limited, and the strength of this association in the presence of brain pathologies is unknown.

Objective  To examine the association of lifespan CR with dementia risk, taking brain pathologies into account.

Design, Setting, and Participants  This study used data from 2022 participants in the Rush Memory and Aging Project, an ongoing community-based cohort study with annual follow-up from 1997 to 2018 (mean follow-up, 6 years; maximum follow-up, 20 years). After excluding 420 individuals who had prevalent dementia, missing data on CR, or dropped out, 1602 dementia-free adults were identified at baseline and evaluated to detect incident dementia. During follow-up, 611 died and underwent autopsies. Data were analyzed from May to September 2018.

Exposures  Information on CR factors (education; early-life, midlife, and late-life cognitive activities; and social activities in late life) was obtained at baseline. Based on these factors, lifespan CR scores were captured using a latent variable from a structural equation model and was divided into tertiles (lowest, middle, and highest).

Main Outcomes and Measures  Dementia was diagnosed following international criteria. Neuropathologic evaluations for Alzheimer disease and other brain pathologies were performed in autopsied participants. The association of lifespan CR with dementia or brain pathologies was estimated using Cox regression models or logistic regression.

Results  Of the 1602 included participants, 1216 (75.9%) were women, and the mean (SD) age was 79.6 (7.5) years. During follow-up, 386 participants developed dementia (24.1%), including 357 participants with Alzheimer disease–related dementia (22.3%). The multiadjusted hazards ratios (HRs) of dementia were 0.77 (95% CI, 0.59-0.99) for participants in the middle CR score tertile and 0.61 (95% CI, 0.47-0.81) for those in the highest CR score tertile compared with those in the lowest CR score tertile. In autopsied participants, CR was not associated with most brain pathologies, and the association of CR with dementia remained significant after additional adjustment for brain pathologies (HR, 0.60; 95% CI, 0.42-0.86). The highest CR score tertile was associated with a reduction in dementia risk, even among participants with high Alzheimer disease pathology (HR, 0.57; 95% CI, 0.37-0.87) and any gross infarcts (HR, 0.34; 95% CI, 0.18-0.62).

Conclusions and Relevance  High lifespan CR is associated with a reduction in dementia risk, even in the presence of high brain pathologies. Our findings highlight the importance of lifespan CR accumulation in dementia prevention.

The dolichofacial individual was associated with security agents as the most prone to commit crimes, was seen as less trustworthy

Does the facial pattern give individuals a profile of a crime suspect? Nathalia de Lima SANTOS et al. Biosci. J., Uberlândia, v. 35, n. 5, p. 1614-1621, Sep./Oct. 2019. http://dx.doi.org/10.14393/BJ-v35n5a2019-46326

ABSTRACT: To evaluate the influence of mesofacial, brachyfacial and dolichofacial facial patterns on giving an individual the profile of a crime suspect in the eyes of public security agents. This study had a crosssectional design, conducted with public security agents of both sexes (n=100), where images of facial composites (police sketches) of individuals with different facial patterns (mesofacial, brachyfacial and dolichofacial) were used. With these images in hand, a questionnaire was created, divided into three parts: the first in which all the images were presented together, allowing comparison among them; the second, in which each image was evaluated separately followed by questions and the third that consisted on a visual analog scale that presented a bar with marks going from 0 to 100, where 0 represented the untrustworthy individual, 50 the individual who could be trusted, and 100 a very trustworthy individual. When all the data had been obtained statistical analyses were performed using the Chi-square and Friedman tests. The level of significance adopted was 5% (α=0.05). The dolichofacial individual was associated with security agents as the most prone to commit crimes and became more insecure and distrustful when compared to the mesofacial and brachyfacial individuals (p <0.001). The dolichofacial profile had a negative influence on the judgment of security agents who attributed to it, a character suspected of a crime and a low level of trustworthiness.

KEYWORDS: Face. Social perception. Judgment. Crime.

Monday, October 14, 2019

The Rare Sides of Twin Research: Representation of Self-Image; Twins With Kleine–Levin Syndrome; Heteropaternal Lemur Twins; Risk of Dental Caries/In the Media

The Rare Sides of Twin Research: Important to Remember/Twin Research Reviews: Representation of Self-Image; Twins With Kleine–Levin Syndrome; Heteropaternal Lemur Twins; Risk of Dental Caries/In the Media: High-Society Models; ‘Winkelevii’ Super Bowl Twins; Multiple Birth × Three; Twin Sister Surrogate; A Presidential Twin? Nancy L. Segal. Twin Research and Human Genetics, October 14 2019. https://doi.org/10.1017/thg.2019.84

Abstract: This article explores some rare sides of twin research. The focus of this article is the sad plight of the Dionne quintuplets, born in Canada in 1934. However, several other studies belong in this category, such as Dr Josef Mengele’s horrifying twin research conducted at the Auschwitz concentration camp, Dr John Money’s misguided attempt to turn an accidentally castrated male twin into a female, Russian scientists’ cruel medical study of conjoined female twins and Dr Peter Neubauer’s secret project that tracked the development of separated twins. Reviews of current twin research span twins’ representation of self-image, twins with Kleine–Levin Syndrome, heteropaternal twinning in lemurs and factors affecting risk of dental caries. Media coverage includes a pair of high-society models, a book about the ‘Winkelevii’ twins, Super Bowl twin teammates, a family with three sets of fraternal twins, a twin sister surrogate and a near presidential twin.

Male Qualities and Likelihood of Orgasm in Women: Consistent findings regarding males who elicit orgasm more frequently pertain to their sexual behavior (being attentive, patient, & receptive to instruction) rather than their traits

Male Qualities and Likelihood of Orgasm. James M. Sherlock and Morgan J. Sidari. In T.K. Shackelford, V.A. Weekes-Shackelford (eds.), Springer Encyclopedia of Evolutionary Psychological Science, 2020, https://doi.org/10.1007/978-3-319-16999-6_278-1

Definition
Male traits and sexual behaviors that predict the likelihood of female orgasm.

Introduction
In contrast to male orgasm, the female orgasm is enormously variable in frequency during penetrative sex. While men almost always ejaculate during penile-vaginal intercourse, women orgasm far less frequently from penetrative sex alone and experience significant variation in orgasm frequency with different partners (Lloyd 2005).  While variation in women.s orgasm frequency is poorly understood, evolutionary psychology tends to view variation in behavior as adaptive and responsive to environmental conditions.  Current evolutionary theories regarding the female orgasm can be broadly divided into two positions: those that focus on selection on male sexual function and those that focus on selection on female sexual function. The by-product hypothesis concerns the former and posits that the capacity of women to experience orgasm is a consequence of strong selection pressure on males' capacity to reach orgasm. This position is based on similarities between male and female orgasm as well as the observation that the male glans penis and female clitoris arise from homologous tissue during development. In contrast, the mate choice hypothesis argues that variation in female orgasm frequency during sexual intercourse is reflective of varying quality of their male partners. [...]

Mate-Choice Hypotheses of Female Orgasm
Given the high gestational cost of rearing human offspring women should be driven to select mates of high quality. Under the sire-choice hypothesis, quality tends to be defined as the ability to pass favorable genetic traits onto offspring (i.e., heritable traits) that will contribute to offspring fitness.  In contrast, the pair-bonding hypothesis postulates that male traits that are likely to benefit the woman through increased care and investment in offspring ought to promote orgasm. Traits that have been identified as theoretically important to each theory can be seen in Table 1 below.

Male Qualities and Likelihood of Orgasm, Table 1
Partner traits distinguishing mate-choice hypotheses (For review see Sherlock et al. 2016)

> Sire choice
Physical attractiveness - Height - Athleticism - Muscularity - Voice depth - Physical fitness - Humor - Creativity - Dominance - Body odor pleasantness

> Pair-bonding
Faithfulness - Warmth - Earning potential - Kindness


Mate-Choice Traits that Predict Orgasm Likelihood

Sire-Choice
The sire-choice hypothesis has been more extensively tested than the pair-bonding hypothesis and some evidence does suggest that women may be more likely to orgasm when their partner possesses traits putatively associated with genetic quality. [...]

Pair-Bonding
Few studies to date have thoroughly investigated pair-bonding traits in the context of female orgasm frequency. However, Costa and Brody (2007) have observed that greater orgasm frequency is associated with overall relationship quality. [...]

Alternative Considerations
While some male traits have been reported to covary with orgasm frequency, this may not represent a causal relationship. Firstly, it is possible that women who orgasm more frequently may be more likely to select particular types of partners than women who orgasm infrequently. For example, women who report a higher orgasm frequency may choose to have sex with more physically attractive men. These men are likely to have more experience in short-term sexual relationships and therefore may be more effective at eliciting orgasm in their partners. [...]

Only one study to date has investigated how women.s orgasm frequencies have varied with different sexual partners. Sherlock et al. (2016) had single women with more than two male sexual partners report on a range of characteristics of the man with whom orgasm was the easiest and the man with whom orgasm was the most difficult (or did not occur at all). By comparing high- and low-orgasm men, several traits emerged as important in predicting ease of orgasm. High-orgasm men tended to be higher in humor, attractiveness, creativity, emotional warmth, faithfulness, and body odor pleasantness, consistent with both sire-choice and pair-bonding hypotheses.

Male Sexual Behavior
Importantly, Sherlock et al. (2016) also observed that a number of sexual behaviors differed between high- and low-orgasm males. Specifically, high-orgasm males were more likely to be focused on their partner.s pleasure, engage in oral sex, use sex toys, spend more time on foreplay, and stimulate their partner's clitoris during sex.  Women were also more likely to communicate their sexual position preferences and stimulate their clitoris when having sex with high-orgasm partners. [...] [...] Across all three studies, orgasm was more likely to occur with manual stimulation of the clitoris during intercourse (Frederick et al. 2017; Richters et al. 2006; Sherlock et al. 2016). Consequently, any contribution of male traits (e.g., attractiveness) to female orgasm needs to be considered in the context of variation in male sexual behavior. Further complicating these results is the likely association between male traits and sexual behaviors. [...]

Conclusion
Despite some consistency in male traits associated with orgasm across several studies, there is reason to be cautious interpreting these results without first accounting for male sexual behavior. The most prominent trait associated with increases in the likelihood of female orgasm is attractiveness (Andersson 1994; Gallup et al. 2014; Grammer et al. 2003; Shackelford et al. 2000; Sherlock et al.  2016), yet the causal pathway between attractiveness and female orgasm could be inverse to the current theorizing. That is, women could come to view their partners as more attractive if they are more frequent benefactors of orgasms. In sum, the most consistent findings regarding males who elicit orgasm more frequently pertain to their sexual behavior rather than their traits. Women are more likely to achieve orgasm with a partner who is attentive, patient, and receptive to instruction.

Female Mate Choice; Mate selection; Orgasm; Pair-bonding; The Evolution of Genitalia

Effects of the over-the-counter pain reliever acetaminophen, which can alter consumers’ emotional experiences and their economic behavior well beyond soothing their aches and pains, has also memory effects

Drug influences on consumer judgments: emerging insights and research opportunities from the intersection of pharmacology and psychology. Geoffrey R. O. Durso, Kelly L. Haws, Baldwin M. Way. Marketing Letters, October 10 2019. https://link.springer.com/article/10.1007/s11002-019-09500-z

Abstract: Recent evidence at the intersection of pharmacology and psychology suggests that pharmaceutical products and other drugs can exert previously unrecognized effects on consumers’ judgments, emotions, and behavior. We highlight the importance of a wider perspective for marketing science by proposing novel questions about how drugs might influence consumers. As a model for this framework, we review recently discovered effects of the over-the-counter pain reliever acetaminophen, which can alter consumers’ emotional experiences and their economic behavior well beyond soothing their aches and pains, and also present novel data on its memory effects. Observing effects of putatively benign over-the-counter medicines that extend beyond their originally approved usages suggests that many other drugs are also likely to influence processes relevant for consumers. The ubiquity of drug consumption—medical or recreational, legal or otherwise—underscores the importance of considering several novel research directions for understanding pharmacological-psychological interactions on consumer judgments, emotions, and behaviors.

Keywords: Decision making Emotion Memory Pharmaceuticals Substances Acetaminophen





Check also Conference Talk—SMPC 2019: Effect of Acetaminophen on Emotional Sounds. Lindsay A. Warrenburg. 2019. https://osf.io/79f4d/

Description: The capacity of listeners to perceive or experience emotions in response to music, speech, and natural sounds depends on many factors including dispositional traits, empathy, and enculturation. Emotional responses are also known to be mediated by pharmacological factors, including both legal and illegal drugs. Existing research has established that acetaminophen, a common over-the-counter pain medication, blunts emotional responses to visual stimuli (e.g., Durso, Luttrell, & Way, 2015). The current study extends this research by examining possible effects of acetaminophen on both perceived and felt responses to emotionally-charged sound stimuli. Additionally, it tests whether the effects of acetaminophen are specific for particular emotions (e.g., sadness, fear) or whether acetaminophen blunts emotional responses in general. Finally, the study tests whether acetaminophen has similar or differential effects on three categories of sound: music, speech, and natural sounds. The experiment employs a randomized, double-blind, parallel-group, placebo-controlled design. Participants are randomly assigned to ingest acetaminophen or a placebo. Then, the listeners are asked to complete two experimental blocks regarding musical and non-musical sounds. The first block asks participants to judge the extent to which a sound conveys a certain affect (on a Likert scale). The second block aims to examine a listener’s emotional responses to sound stimuli (also on a Likert scale). In light of the fact that some 50 million Americans take acetaminophen each week, this study suggests that future studies in music and emotion might consider controlling for the pharmacological state of participants.


Sunday, October 13, 2019

Some people share knowledge online, often without tangible compensation; they are motivated to signal general intelligence g; observers infer g from contributions' quality

The quality of online knowledge sharing signals general intelligence. Christian N.Yoder, Scott A.Reid. Personality and Individual Differences, Volume 148, 1 October 2019, Pages 90-94. https://doi.org/10.1016/j.paid.2019.05.013

Abstract: Some people share knowledge online, often without tangible compensation. Who does this, when, and why? According to costly signaling theory people use behavioral displays to provide observers with useful information about traits or states in exchange for fitness benefits. We tested whether individuals higher in general intelligence, g, provided better quality contributions to an information pool under high than low identifiability, and whether observers could infer signaler g from contribution quality. Using a putative online wiki (N = 98) we found that as individuals' scores on Ravens Progressive Matrices (RPM) increased, participants were judged to have written better quality articles, but only when identifiable and not when anonymous. Further, the effect of RPM scores on inferred intelligence was mediated by article quality, but only when signalers were identifiable. Consistent with costly signaling theory, signalers are extrinsically motivated and observers act as “naive psychometricians.” We discuss the implications for understanding online information pools and altruism.

Brain Syndrome Can Make Owners Think Pets Are Impostors

Herzog, Harold, "Brain Syndrome Can Make Owners Think Pets Are Impostors" (2018). 'Animals and Us' Blog Posts. 103, Apr 25, 2018. https://animalstudiesrepository.org/aniubpos/103

Mary was 40 years old when she became convinced that Sarah, her 9 year old daughter, was an impostor.  The real Sarah, she told relatives, had been taken away and placed into a foster home. She claimed social workers had replaced her actual child with an identical-looking impostor.  Mary was so convinced of this substitution that she would sometimes refuse to pick her daughter up at school. Mary would scream to the teachers, “Give me my real daughter back, I know what you have done!”
To no avail, her family and health care providers tried to convince Mary that no substitution had occurred, that Sarah was indeed her real daughter.  But even after Mary was treated with risperiodone, a powerful anti-psychotic drug, she held on to the delusion. The local Department of Social Services became concerned about her ability to raise a child. And when it became apparent that Mary could no longer provide care for the daughter who she believed was an impostor, they successfully sought legal guardianship for Sarah. At one point during the hearing, Sarah told the court, “I love my mother, except when she doesn’t believe I’m me.”

As described in an article by Drs. Jeremy Matuszak and Matthew Parra in the journal Psychiatric Times, Mary was suffering from Capgras Syndrome. This is a rare variant of a group of neuropsychiatric conditions called delusional misidentification disorders.  First identified in 1923 by the French psychiatrists Joseph Capgras and Reboul-Lachaux, individuals with the Capgras delusion come to believe that a person they know has been replaced by an identical-looking impostor. Usually the target of the delusion is a family member or a loved one. In Mary’s case, it was her young daughter.

[...]

What Causes Animal Capgras Delusions?

I admit that nine cases is a small sample, but some interesting patterns did emerge among this group of patients. For example, twice as many women as men thought they were living with impostor animals.  And, as group, the patients tended to be on the old side. Six of eight individuals were over 50, and half were in their late 60’s or older. While only two of the patients had suffered identifiable brain damage, nearly all the patients had been diagnosed with a functional psychosis, usually a form of schizophrenia.  Finally, all seven of the patients which there was information on treatments were given anti-psychotic drugs. In nearly all of these cases, their pet impostor delusions diminished, and in several cases, they seemed to have disappeared.

Speculations about the causes of Capras syndrome abound. The some researchers argue that impostor delusions are a way of subconsciously dealing with love-hate conflicts. V. S. Ramachandran believes that imposter delusions result from disconnections between emotional and face recognition centers in the brain.  Others argue it is usually the result of degenerative diseases such as Alzheimer’s and Lewy body disease. Indeed, impostor delusions have been associated with a wide array of conditions including psychiatric disorders, strokes, tumors, epilepsy, and even vitamin deficiencies and drug use.

Targeted Memory Reactivation During Sleep Improves Next-Day Problem Solving

Targeted Memory Reactivation During Sleep Improves Next-Day Problem Solving. Kristin E. G. Sanders et al. Psychological Science, October 11, 2019. https://doi.org/10.1177/0956797619873344

Abstract: Many people have claimed that sleep has helped them solve a difficult problem, but empirical support for this assertion remains tentative. The current experiment tested whether manipulating information processing during sleep impacts problem incubation and solving. In memory studies, delivering learning-associated sound cues during sleep can reactivate memories. We therefore predicted that reactivating previously unsolved problems could help people solve them. In the evening, we presented 57 participants with puzzles, each arbitrarily associated with a different sound. While participants slept overnight, half of the sounds associated with the puzzles they had not solved were surreptitiously presented. The next morning, participants solved 31.7% of cued puzzles, compared with 20.5% of uncued puzzles (a 55% improvement). Moreover, cued-puzzle solving correlated with cued-puzzle memory. Overall, these results demonstrate that cuing puzzle information during sleep can facilitate solving, thus supporting sleep’s role in problem incubation and establishing a new technique to advance understanding of problem solving and sleep cognition.

Keywords problem solving, incubation, sleep, targeted memory reactivation, restructuring, creative cognition

When recipients open a gift from a friend, they like it less when the giver has wrapped it neatly as opposed to sloppily and we draw on expectation disconfirmation theory to explain the effect

Presentation Matters: The Effect of Wrapping Neatness on Gift Attitudes. Jessica M. Rixom  Erick M. Mas  Brett A. Rixom. Journal of Consumer Psychology, October 11 2019. https://doi.org/10.1002/jcpy.1140

Abstract: While gift‐givers typically wrap gifts prior to presenting them, little is known about the effect of how the gift is wrapped on recipients’ expectations and attitudes toward the gift inside. We propose that when recipients open a gift from a friend, they like it less when the giver has wrapped it neatly as opposed to sloppily and we draw on expectation disconfirmation theory to explain the effect. Specifically, recipients set higher (lower) expectations for neatly (sloppily)‐wrapped gifts, making it harder (easier) for the gifts to meet these expectations, resulting in contrast effects that lead to less (more) positive attitudes toward the gifts once unwrapped. However, when the gift‐giver is an acquaintance, there is ambiguity in the relationship status and wrapping neatness serves as a cue about the relationship rather than the gift itself. This leads to assimilation effects where the recipient likes the gift more when neatly wrapped. We assess these effects across three studies and find that they hold for desirable, neutral, and undesirable gifts, as well as with both hypothetical and real gifts.


Saturday, October 12, 2019

Random effects meta-analyses showed that social media use was significantly & positively related to affective empathy; effects were generally small in size and do not establish causality

Social Media Use and Empathy: A Mini Meta-Analysis. Shu-Sha Angie Guan, Sophia Hain, Jennifer Cabrera, Andrea Rodarte. Computer Science & Communications, Vol.8 No.4, October 2019, pp. 147-157. DOI: 10.4236/sn.2019.84010

ABSTRACT: Concerns about the effects of social media or social networking site (SNS) use on prosocial development are increasing. The aim of the current study is to meta-analytically summarize the research to date (k = 5) about the relationship between general SNS use and two components of empathy (i.e., empathic concern and perspective-taking). Random effects meta-analyses showed that SNS use was significantly and positively related to affective empathy though only marginally related to cognitive empathy. These effects were generally small in size and do not establish causality. Future research should explore how specific behaviors are related to different forms of empathy.

KEYWORDS: Social Media, Empathic Concern, Perspective-Taking

4. Discussion
Despite the decreases in empathy coupled with increases in media use at the societal level [13] , individual social media use in terms of frequency or time spent per day appears to be related to higher levels of empathy, particularly affective empathy. Even though the associations were small, they trended positive. However, there may be some online behaviors that cultivate empathy (e.g., sharing emotions, expressing support [21] ) more than others (e.g., updating profile photos [20] ). In combination with emerging longitudinal evidence that social media use at one time point is predictive of higher levels of cognitive and affective empathy one year later among adolescents [42] and experimental work that shows that interdependent Facebook use can promote relational orientation [37] , this study contributes to the growing literature on how social media can facilitate positive psychosocial development.
Although promising, there are limitations of the current meta-analysis to consider. This study aimed to look only at global measures of social media use in everyday life and, because of this inclusion parameter, includes a small sample of studies and effect sizes. This likely limits the generalizability of the results and our ability to detect differences by moderators (gender, age). Also, the results are correlational and do not establish causality. Previous research suggests that individuals who are prosocial offline are often prosocial online [29]. Despite our attempts to narrow the scope, there remained variability in the measures of media use and study parameters as indicated by the heterogeneity index. Given the wide range of online activities, future studies should explore how specific behaviors are related to different forms of empathy (e.g., helping strangers vs. family or friends [25] ). Additionally, the social media landscape is constantly evolving and this study captures media use as assessed by recent studies in one moment in time. Cultural psychologists suggest that changes in technology use, as part of larger shifting sociodemographic and ecological changes, can shape cultural values and learning environments in ways that directly affect human development across time [43].
It is also important to note that all of the studies included, and much of media research in general, have been conducted in industrialized, individualistic countries like the United States. This limited our ability to detect cultural differences. On the one hand, the most popular SNSs are often developed in Western cultures and can reflect the highly individualistic values of their developers and users [37] [44]. On the other hand, the Internet is a “global village” of individuals from various nationalities and cultural backgrounds with nearly 60% of the online population residing outside of the U.S. [44]. These diverse offline cultural values can be reflected in the online [45] - [52]. Additionally, there may be values and goals specific to the SNS context outside of the values that users bring with them [53]. Previous meta-analyses suggest that the effects of media use may be stronger in non-Western countries [26]. Future research should explore how cultural values in the online and offline interact in shaping development.
Although limited, this meta-analysis provides useful insights into the media-empathy paradox [13]. Additionally, it may be informative in better understanding growing generations of adolescents and young adults who have become the first generations to have grown up fully immersed in digital media (i.e., “digital natives”) having been born around or after the 1990s when the Internet was first commercially launched. This may mean that psychosocial development for these “digital natives” differs from prior generations of “digital immigrants” [9]. For example, greater face-to-face communication with family members, close friends, and acquaintances was associated with higher levels of psychological well-being (e.g., life meaning, relationship quality) for older adults age 35 - 54 but not for young adults age 18 - 34 [54]. As technology transforms society, social relationships, and media landscapes, it will become ever important to track how these changes affect individuals and their development.

Is intense pleasure necessary for intense beauty? If so, the inability to experience pleasure (anhedonia) should prevent the experience of intense beauty

Intense beauty requires intense pleasure. Aenne A. Brielmann1 and Denis Pelli. Front. Psychol., Oct 11 2019 (provisionally accepted, no full text available). doi: 10.3389/fpsyg.2019.02420

Abstract: At the beginning of psychology, Fechner (1876) claimed that beauty is immediate pleasure, and that an object’s pleasure determines its value. In our earlier work, we found that intense pleasure always results in intense beauty. Here, we focus on the complementary question: Is intense pleasure necessary for intense beauty? If so, the inability to experience pleasure (anhedonia) should prevent the experience of intense beauty. We asked 757 participants to rate how intensely they felt beauty from each image. We used 900 OASIS images along with their available valence (pleasure vs. displeasure) and arousal ratings. We then obtained self-reports of anhedonia (TEPS), mood, and depression (PHQ-9). Across images, beauty ratings were closely related to pleasure ratings (r = 0.75), yet unrelated to arousal ratings. Only images with an average pleasure rating above 4 (of a possible 7) often achieved (> 10%) beauty averages exceeding the overall median beauty. For normally beautiful images (average rating > 4.5), the beauty ratings were correlated with anhedonia (r ~ -0.3) and mood (r ~ 0.3), yet unrelated to depression. Comparing each participant’s average beauty rating to the overall median, none of the most anhedonic participants exceeded the median, whereas 50% of the remaining participants did. Thus, both general and anhedonic results support the claim that intense beauty requires intense pleasure. In addition, follow-up repeated measures showed that shared taste contributed 19% to beauty-rating variance, only one third as much as personal taste (58%). Addressing age-old questions, these results indicate that beauty is a kind of pleasure, and that beauty is more personal than universal, i.e., 1.7 times more correlated with individual than with shared taste.

Keywords: beauty, aesthetics, Pleasure, Anhedonia, Depression

Effects of heroin on rat prosocial behavior: They stopped freeing the trapped cagemate, continued to self-administer the drug

From 2018... Effects of heroin on rat prosocial behavior. Seven E. Tomek  Gabriela M. Stegmann  M. Foster Olive. Addiction Biology, May 4 2018. https://doi.org/10.1111/adb.12633

Abstract: Opioid use disorders are characterized in part by impairments in social functioning. Previous research indicates that laboratory rats, which are frequently used as animal models of addiction‐related behaviors, are capable of prosocial behavior. For example, under normal conditions, when a ‘free’ rat is placed in the vicinity of rat trapped in a plastic restrainer, the rat will release or ‘rescue’ the other rat from confinement. The present study was conducted to determine the effects of heroin on prosocial behavior in rats. For 2 weeks, rats were given the opportunity to rescue their cagemate from confinement, and the occurrence of and latency to free the confined rat was recorded. After baseline rescuing behavior was established, rats were randomly selected to self‐administer heroin (0.06 mg/kg/infusion i.v.) or sucrose pellets (orally) for 14 days. Next, rats were retested for rescuing behavior once daily for 3 days, during which they were provided with a choice between freeing the trapped cagemate and continuing to self‐administer their respective reinforcer. Our results indicate that rats self‐administering sucrose continued to rescue their cagemate, whereas heroin rats chose to self‐administer heroin and not rescue their cagemate. These findings suggest that rats with a history of heroin self‐administration show deficits in prosocial behavior, consistent with specific diagnostic criteria for opioid use disorder. Behavioral paradigms providing a choice between engaging in prosocial behavior and continuing drug use may be useful in modeling and investigating the neural basis of social functioning deficits in opioid addiction.

From 2001... Need to hire a private certified arborist at a cost of $500-$2,000 to take pictures, prepare a report & perhaps to recommend protective pruning or other measures before a tree can be cut

From 2001... The (Almost) Untouchables of California. Todd S. Purdum. , Aug. 29, 2001, Section A, Page 1. https://www.nytimes.com/2001/08/29/us/the-almost-untouchables-of-california.html

'Anything that's going to happen under this tree has to be addressed,' said Mr. Sartain, a third-generation arborist, surveying the tree's 90-foot canopy with the cheerful, clinical detachment of your favorite pediatrician. 'There's a lot of issues.'

Indeed, Mr. Sartain's visit is only the first step in a process that will require the homeowner, who asked not to be named, to hire a private certified arborist at a cost of $500 to $2,000 to take pictures, prepare a report and perhaps to recommend protective pruning or other measures before a permit is issued and construction can proceed. Penalties for removing a tree like this, worth perhaps $100,000 under city guidelines because of its size and age, could force an offender to plant trees worth an equivalent amount.

Santa Clarita is not alone.

In the past 30 years, as development pressures increased, scores of California cities and counties from Thousand Oaks in the south to Santa Rosa in the north have passed ordinances protecting not only various species and sizes of oaks, but also sycamores, walnuts, eucalyptus and other trees with a zeal that might make the poet Joyce Kilmer blush.

The specifics vary widely, but the ordinances have one goal in common: protecting trees that are almost as storied in California as its redwoods and that have long been threatened by ranching, wine-making, suburban sprawl and, more recently, mysterious diseases.

Children from Namibia and Germany: Being observed increases overimitation in three diverse cultures

Stengelin, R., Hepach, R., & Haun, D. B. M. (2019). Being observed increases overimitation in three diverse cultures. Developmental Psychology, Oct 2019. http://dx.doi.org/10.1037/dev0000832

Abstract: From a young age, children in Western, industrialized societies overimitate others’ actions. However, the underlying motivation and cultural specificity of this behavior have remained unclear. Here, 3- to 8-year-old children (N = 125) from two rural Namibian populations (Hai||om and Ovambo) and one urban German population were tested in two versions of an overimitation paradigm. Across cultures, children selectively imitated more actions when the adult model was present compared to being absent, denoting a social motivation underlying overimitation. At the same time, children’s imitation was not linked to their tendency to reengage the adult in a second, independent measure of social motivation. These results suggest that, across diverse cultures, children’s imitative behavior is actuated by the attentive state of the model.

Friday, October 11, 2019

Age is the most important risk factor for loneliness, which peaks just prior to the peak age of onset for psychotic disorders; contrary to expectations, it declines thereafter over the lifespan

Shovestul, Bridget, Jiayin Han, Laura Germine, and David Dodell-Feder. 2019. “Risk Factors for Loneliness: The High Relative Importance of Age Versus Other Factors.” PsyArXiv. October 11. doi:10.31234/osf.io/t8n2h

Abstract
Background: Loneliness is a potent predictor of negative health outcomes making it important to identify risk factors for loneliness. Though extant studies have identified characteristics that are associated with loneliness, less is known about the cumulative and relative importance of these factors, and how their interaction may impact loneliness. Thus, here, we investigate risk factors for loneliness.
Methods: 4,885 individuals ages 10-97 years from the US completed the three-item UCLA Loneliness Survey on TestMyBrain.org. Using census data, we calculated the population and community household income of participants’ census area, and the proportion of individuals in the participant’s census area that shared the participant’s demographic characteristics (i.e., sociodemographic density). We evaluated the relative importance of three classes of variables for loneliness risk: those related to the person (e.g., age), place (e.g., community household income), and the interaction of person X place (sociodemographic density).
Results: We find that loneliness is highly prevalent and best explained by person (age) and place (community household income) characteristics. Of the variance in loneliness accounted for, the overwhelming majority was explained by age. On age, loneliness peaks at 19 years, and declines thereafter. The congruence between one’s sociodemographic characteristics and that of one’s neighborhood had no impact on loneliness.
Conclusions: Age appears to be the most important risk factor for loneliness, which peaks just prior to the peak age of onset for psychotic disorders, and, contrary to popular belief, declines thereafter over the lifespan. These data may have important implications for public health interventions.


Youth marijuana use can have adverse health outcomes; however, reports from Colorado, Oregon, & Washington indicate no statewide increase in youth marijuana use following retail legalization for adults


Ta M, Greto L, Bolt K. Trends and Characteristics in Marijuana Use Among Public School Students — King County, Washington, 2004–2016. MMWR Morb Mortal Wkly Rep 2019;68:845–850. https://www.cdc.gov/mmwr/volumes/68/wr/mm6839a3.htm

Summary

What is already known about this topic?
Youth marijuana use can have adverse health outcomes. However, reports from Colorado, Oregon, and Washington indicate no statewide increase in youth marijuana use following retail legalization for adults.

What is added by this report?
Following 2012 legalization of retail marijuana sale to adults in Washington, past 30–day marijuana use decreased or remained stable through 2016 among King County students in grades 6, 8, 10, and 12. Among grade 10 students, the decline in use occurred among males while the rate among females remained steady. Use of alcohol or other substances was four times as frequent among marijuana users as among nonusers.

What are the implications for public health practice?
Understanding reasons for youth marijuana use, particularly among females, might help inform policy, strategies, and educational campaigns.

Use of marijuana at an early age can affect memory, school performance, attention, and learning; conclusions have been mixed regarding its impact on mental health conditions, including psychosis, depression, and anxiety (1–3). Medical marijuana has been legal in Washington since 1998, and in 2012, voters approved the retail sale of marijuana for recreational use to persons aged ≥21 years. The first retail stores opened for business in July 2014. As more states legalize marijuana use by adults aged ≥21 years, the effect of legalization on use by youths will be important to monitor. To guide planning of activities aimed at reducing marijuana use by youths and to inform ongoing policy development, Public Health—Seattle & King County assessed trends and characteristics of past 30–day marijuana use among King County, Washington, public school students in grades 6, 8, 10, and 12. This report used biennial data for 2004–2016 from the Washington State Healthy Youth Survey. Among grade 6 students there was a decreasing trend in self-reported past 30–day marijuana use from 2004 to 2016, while the percentage of grade 8 students who had used marijuana during the past 30 days did not change during that period. Among students in grades 10 and 12, self-reported past 30–day use of marijuana increased from 2004 to 2012, then declined from 2012 to 2016. In 2016, the percentage of students with past 30–day marijuana use in King County was 0.6% among grade 6, 4.1% among grade 8, 13.9% among grade 10, and 25.5% among grade 12 students. Among grade 10 students, 24.0% of past 30–day marijuana users also smoked cigarettes, compared with 1.3% of nonusers. From 2004 to 2016 the prevalence of perception of great risk of harm from regular marijuana use decreased across all grades. Continued surveillance using consistent measures is needed to monitor the impact of marijuana legalization and emerging public health issues, given variable legislation approaches among jurisdictions.

The Healthy Youth Survey is a school-based, anonymous, self-administered, cross-sectional survey conducted in the fall of even-numbered years in Washington public schools.* Schools with grades 6, 8, 10, and 12 are randomly selected using a clustered sampling design. Schools not selected for the state sample also can choose to participate in the survey. The survey measures risk behaviors, attitudes, and factors that contribute to youth health and safety, including alcohol, marijuana, tobacco, and other drug use; behaviors that result in unintentional and intentional injuries (e.g., violence); dietary behaviors, and physical activity.

This analysis used data from all participating schools, both sampled and nonsampled, representing all 19 King County school districts for biennial survey years 2004 through 2016 (the most currently available year of data at the time of analysis). King County is the largest metropolitan county in the state. Local jurisdictions have authority to regulate land uses and can impose additional time, place, and manner-of-use restrictions on state licensed businesses; thus, considerable variation in the availability of and restrictions on retail marijuana exists across the 39 cities in King County, including Seattle.

Survey response rates varied by grade and survey year, with higher rates in more recent surveys.† During 2004–2016, King County response rates ranged from 60%–80% for grades 6 and 8; 50%–70% for grade 10; and 40%–50% for grade 12. For the 2016 survey, response rates for King County were 80% for grades 6 and 8, 70% for grade 10, 40% for grade 12, and 67% for all grades combined.

Data representing substance use, perception of great risk of harm, risky behaviors, and factors associated with marijuana use were categorized dichotomously. Past 30–day marijuana use was considered use on 1 or more days during the past 30 days. Perceived great risk of harm associated with regular marijuana use (more than one or two times per week) was categorized dichotomously as great risk versus all other options combined (moderate, slight, and no perceived risk). Past 30–day use of alcohol, cigarettes, and electronic cigarettes/vape pens was considered use on 1 or more days in the past 30 days, past 30–day risky driving and riding behaviors,§ were considered one or more occurrences during the past 30 days and past binge drinking¶ was over a 2-week period.

Dichotomous factors generally reported to be associated with other substance use (4) were examined for marijuana use; these factors included whether students’ parents had talked about not using marijuana, use by one or more best friends or by a member in the youth’s household, and having been bullied one or more times in the past month.** Stata survey software (version 13; StataCorp) was used to generate percentage estimates and corresponding 95% confidence intervals (CIs). To account for differential participation among school districts across survey years, percentage estimates were weighted to the school district total enrollment by grade and sex, with the final weights adjusted to sum to the county total public school enrollment by grade and sex. Joinpoint trend analysis software (https://surveillance.cancer.gov/joinpoint/external icon) was used to evaluate statistical significance of trends in survey-weighted percentage estimates by grade and sex. Analyses of trends by sex and examination of factors associated with past 30–day use were restricted to grade 10 students as a result of grade-specific sampling and the need for adequate response rates to accommodate a robust analysis.

During 2004–2016, the prevalence of reported past 30–day marijuana use was lowest among students in grade 6 and increased with school grade level (Figure 1). In 2016, past 30–day marijuana use was reported by 0.6% (CI = 0.4–0.7) of grade 6 students, 4.1% (CI = 3.5–4.8) of grade 8 students, 13.9% (CI = 12.6–15.3) of grade 10 students, and 25.5% (CI = 23.7–27.4) of grade 12 students in King County. Among students in grade 6, past 30–day marijuana use declined significantly, from 1.3% in 2004 to 0.6% in 2016. There was no statistically significant trend among students in grade 8; however, among students in grades 10 and 12, past 30–day use increased from 2004 to 2012, and then declined. Across all grades, the percentage of students reporting great risk of harm from regular marijuana use declined over the survey period, with the lowest perceived great risk of harm reported among older students in all years. In 2016, 26.7% (CI = 25.0–28.5) of students in grade 12 perceived great risk of harm from regular marijuana use, whereas 53.3% (CI = 50.5–56.1) reported this perception in 2004.

Among male students in grade 10, past 30–day marijuana use increased from 17.6% in 2004 to 21.4% in 2010 and subsequently declined to 13.5% in 2016 (Figure 2). Among female students in grade 10, there was no change in the prevalence of past 30–day use, which remained approximately 16% during this period. In 2016, there was no significant difference in past 30–day marijuana use between male and female students in grade 10.

Among past 30–day marijuana users in grade 10, 42.8% reported living with someone who uses marijuana, 88.5% reported having at least one best friend who used marijuana, and 26.3% reported having been bullied at least once in the past 30 days; these prevalences were higher than those among grade 10 nonusers (12.8%, 28.3%, and 16.5%, respectively) (Table). Among grade 10 marijuana users, 92.5% reported that it was not very hard to obtain marijuana, compared with 56.7% of nonusers. No parental discussion about marijuana during the past year was reported by similar percentages of past 30–day marijuana users (39.2%) and nonusers (39.8%).

Among grade 10 students, prevalence of past 30–day use of other substances was four times higher among those who had used marijuana in the past 30 days than among those who had not. Among marijuana users, the prevalences of past 30–day use of other substances were as follows; alcohol (67.0%), cigarettes (24.0%), e-cigarettes or vape pens (43.0%), and of binge drinking (43.5%), compared with 10.3%, 1.3%, 4.0%, and 3.7% among nonusers, respectively. Among grade 10 marijuana users, 36% reported driving within 3 hours of using marijuana at least once in the past month.

Discussion

Despite legalization of the retail sale of marijuana to adults in Washington in 2012, evidence from the biennial Washington State Healthy Youth Survey indicates that the prevalence of past 30–day marijuana use among students in grades 10 and 12 began to decline that year. The decline continued in 2016 among grade 10 students and did not change significantly among grade 12 students. This decline or absence of change in youth marijuana use after legalization of retail sales to adults is consistent with trends reported in Colorado and Oregon,†† states that legalized adult retail sales of marijuana in 2013 and 2014, respectively. However, causality of the observed decrease in youth use following retail sale legalization cannot be inferred, because effects might be delayed and this report does not include data from the timeframe that would capture the more recent surge in e-cigarette use by youth and the use of tetrahydrocannabinol (THC) within electronic cigarette (e-cigarette) devices. Although the relationship between legal adult recreational use and youth use is not well understood, two possible reasons for the observed decline in youth use include reduction of illicit market supply through competition§§ and loss of novelty appeal among youths. Furthermore, it would be important to monitor the long-term role legalization might play to foster a permissive use environment given observed strong associations with use and individual and family factors that influence youth use.

Before initiation of retail marijuana sales in Washington in 2014, the statewide prevalence of use among grade 10 students had not changed significantly since 2002, although reported statewide use prevalence in 2016 was higher among students identifying as non-Hispanic American Indian/Alaska Native and Hispanic than among non-Hispanic white and non-Hispanic Asian students (5). Among grade 10 King County students, past 30–day marijuana use by male students has been decreasing since 2010, while the prevalence among female students has not changed. Continued monitoring is necessary to observe how local trends among males change over time. The narrowing of the sex difference gap reflects national trends (6) and suggests that female users might benefit from tailored prevention messages informed by an understanding of reasons for use.

Although overall youth rates of smoking and alcohol are declining nationally (7), the prevalence of any substance use, including alcohol, cigarettes, or vape pens, was four times higher among grade 10 past 30–day marijuana users than among nonusers. Statewide data from 2016 also show similar higher prevalence of household, peer and individual factors associated with youth substance use among grade 10 marijuana users than nonusers (https://www.askhys.net/library/2016/RecentMarijaunaUseGr10.pdfpdf iconexternal icon). Findings from a 2017 survey of Canadian residents aged 15–24 years found that marijuana users were significantly more likely to be past 30-day e-cigarette users, compared with nonusers (8). Polysubstance use and driving after using marijuana or riding in a car driven by someone who had used marijuana recently are public health issues that are important to monitor. Educational campaigns conveying health risk of marijuana use should also address impaired driving, in light of experimental data showing deteriorating control with increasing task complexity and increased risk for involvement in a motor vehicle crash (9).

The findings in this report are subject to at least six limitations. First, these data predate the recent reported increase in youth e-cigarette use and the use of THC in the newest generation of e-cigarette devices. The marijuana use question does not explicitly define use by method and estimates of youth marijuana use might be underestimated if respondents did not consider vaping or edible consumption of marijuana products when responding to the question. Second, data are from public school students only and might not be generalizable to all youths in this age group. Students who might be at higher risk might not be in school; it is estimated that 95.3% of King County residents aged 14–18 years are in school.¶¶ Third, survey participation is voluntary, and responses are based on self-report, which can be subject to recall or response bias. Fourth, these estimates might differ from other state or nationally representative youth health–surveillance systems, in part because of survey methods, age of participants, survey setting, and period during the year the survey was conducted. Fifth, local historical data for youth marijuana use before 2004 are not available, and the effects of medical marijuana legalization, which occurred in 1998, on use by youths is unknown. Finally, binge drinking is framed as five or more drinks in a row during the preceding 2 weeks for both males and females and would likely underestimate excessive alcohol consumption among females compared with using a sex-specific four-drink threshold (10).

The national goals for substance use set by Healthy People 2020*** include a target of 6% for youths aged 12–17 years with past 30–day marijuana use, and progress toward this target requires evidence-based interventions and policies for preventing and treating substance use and abuse among youths. Although some cross-cutting interventions addressing adolescent health are presented in the Community Preventive Task Force’s Community Guide,††† there currently is no specific category for marijuana use, as there is for alcohol and tobacco. The National Registry of Evidence-based Programs and Practices,§§§ a project of the federal Substance Abuse and Mental Health Services Administration, might be a potential alternative source for strategies that reduce marijuana use and prevent associated harms, but these strategies might not be sufficient for states with newly legalized retail marketplaces. In light of the limited evidence base, there is a need to identify individual, relationship, community, and societal determinants of youth substance use that would allow development of broad-based risk-reduction strategies. Continued surveillance would benefit from having a set of standard measures across jurisdictions to monitor the health impacts of retail marijuana sale legalization among states.

Language has emerged in no other species than humans, suggesting a profound obstacle to its evolution; maybe quite specific social conditions were prerequisite for the evolution of language- and symbol-ready hominins

The Role of Egalitarianism and Gender Ritual in the Evolution of Symbolic Cognition. Camilla Power. August 2019. Chp 19 in Handbook of Cognitive Archaeology, Routledge. https://www.researchgate.net/publication/335062001

Abstract: Are there constraints on the social conditions that could have given rise to language and symbolic cognition? Language has emerged in no other species than humans, suggesting a profound obstacle to its evolution. If language is seen as an aspect of cognition, limitations can be expected in terms of computational capacity. But if it is seen it as fundamentally for communication, then the problems will be found in terms of social relationships. Below a certain threshold of cooperation and trust, no language or symbolic communication could evolve (Knight & Lewis, 2017a); this has been termed a “platform of trust” (Wacewicz, 2017).... In this chapter, I argue that quite specific social conditions were prerequisite for the evolution of language- and symbol-ready hominins. One of the requirements differentiating our ancestors from other African apes was a switch to mainly female philopatry – females living with their relatives, rather than dispersing at sexual maturity – coevolving with an increasing tendency to egalitarianism....How did increasing egalitarianism affect males and potentially “feminize” male behavior for cooperative offspring care? How were male and female relations affected in the evolution of genus Homo and Homo sapiens?

We are capable of making accurate personality judgements in computer-mediated communication by means of even small cues like nicknames

The Name Is the Game: Nicknames as Predictors of Personality and Mating Strategy in Online Dating. Benjamin P. Lange et al. Front. Commun., February 12 2019. https://doi.org/10.3389/fcomm.2019.00003

Abstract
Objective: We investigated the communicative function of online dating nicknames. Our aim was to assess if it is possible to correctly guess personality traits of a user simply by reading his/her nickname.

Method: We had 69 nickname users (average age: 33.59 years, 36 female) complete questionnaires assessing their personality (Big 5 + narcissism) and mating strategy (short- vs. long-term). We then checked (using a total of 638 participants, average age: 26.83 years, 355 female), whether personality and mating strategy of the nickname users could be assessed correctly based only on the nickname. We also captured the motivation to contact the user behind a nickname and looked at linguistic features of the nicknames.

Results: We found that personality and mating strategy could be inferred from a nickname. Furthermore, going by trends, women were better at intersexual personality judgments, whereas men were better in intrasexual judgements. We also found several correlates of the motivation to contact the person behind the nickname. Among other factors, long nicknames seemed to deter people from contacting the nickname user.

Conclusions: Findings display that humans are capable of making accurate personality judgements in computer-mediated communication by means of even small cues like nicknames.

Introduction

Language-based face-to-face (ftf) interaction can be considered the most natural way of communication (Kock, 2004). New social media have transformed communication, though, as sender and receiver are not necessarily copresent in such a mediated context. However, communication in the digital world is still language-based, even when only in the form of written language (Koch et al., 2005).

Research on such computer-mediated communication (cmc) can be divided into different approaches. Two of them are: (1) the reduced-social-cues approach (rsc) (Sproull and Kiesler, 1986), and (2) the hyperpersonal communication approach (hp) (Walther, 1996). The first assumes that cmc filters out social context cues. The second emphasizes that cmc might surpass ftf communication, as the sender has the opportunity to optimize their self-representation while the receiver idealizes the sender on the basis of the available cues. Here lies the question whether people are able to, and actually do hide their “true selves,” that is their identity (e.g., personality), or whether they, despite being relatively anonymous, inevitably communicate aspects of their respective identity and personality that are in turn perceived by the receiver (Walther and Parks, 2002).

Sex or gender, respectively, are central features of one's identity and personality (e.g., Mealey, 2000; Ellis et al., 2008). As a matter of fact, sex has been central in cmc research. For instance, Guiller and Durndell (2007) found that in cmc men are more dominant than women, whereas women are more supportive than men—findings reminiscent of sex differences in ftf communication (Eckert and McConnell-Ginet, 2003).

A large body of research (e.g., Savicki et al., 1999; Thomson and Murachver, 2001; Koch et al., 2005) shows that only by reading text, people are able to guess the sex of the writers above chance. The same seems to be true for personality judgments (Park et al., 2015). Entire texts are not necessary, though. Lange et al. (2016b) used pseudonyms chosen by students in written exams, and had participants rate them on assumed sex of the user and other attributes. They found that sex could be guessed correctly above chance with a large effect size. Also, participants ascribed typical female and male attributes to the pseudonyms and even tried to retrieve information on the users' personality. It was also found that women, more than men, used diminutive suffixes in their pseudonyms (like -i in “cuti”). In line with these findings, Heisler and Crabill (2006) demonstrated that the majority of their participants considered themselves capable of correctly guessing the sex and age of the users of e-mail usernames. Moreover, their participants attempted to rate the supposed owners of the e-mail addresses also with respect to, among other aspects, their relationship status.

Not only is sex a matter of interest with respect to the digital world, the phenomenon of online dating is, too (Valkenburg and Peter, 2007). Considering that mate choice is one of the most important areas in social life (Buss, 2003) and that people are increasingly shifting their activities from the offline to the online world, it does not surprise that online dating has become a billion-dollar business (Sautter et al., 2010).

Human mating in general and sex differences in human mating have attracted numerous researchers and have produced a veritable deluge of related literature (e.g., Buss and Barnes, 1986; Buss, 1989; Buss and Schmitt, 1993; for an overview, see Buss, 2003, 2016; Schwarz and Hassebrauck, 2012). This research has, on the one hand, identified several characteristics that both sexes prefer in a mate (e.g., healthy), as well as those that are more preferred by women (e.g., good earning capacity, college graduate) and those more preferred by men (e.g., physically attractive) (Buss et al., 1990). The role of language in human mate choice has also been examined recently (e.g., Lange et al., 2014, 2016a). On the other hand, empirical mate choice research has documented that women are more exacting in mate choice decisions, while men face stronger same-sex competition (for an overview, see Buss, 2003). The first process, called intersexual selection, is the actual mate choice, which in most species occurs as female mate choice. That is, women because of having higher obligatory costs (Trivers, 1972), are more selective, while men, whose obligatory costs are lower, compete more strongly with other men in order to be chosen. This is called intrasexual selection (for an overview, see Buss, 2003).

Another area of interest in mate choice research is the distinction between short-term mating (the search for an affair, a one-night stand, etc.) and long-term mating (the search for a committed, steady relationship) (Buss and Schmitt, 1993), which can be referred to as a person's mating strategy (Schmitt, 2005). This distinction is somewhat linked to females being choosier than males. As the costs for males are lower than for females, men show a tendency to be relatively indiscriminate in short-term mating. A bad mate choice imposes higher costs on women than on men—and this applies more to short-term than to long-term mating. Generally, women show a preference for a long-term mate (Buss and Schmitt, 1993). As a result, men for whom short-term mating is a particularly useful strategy might want to pretend to be interested in long-term mating, while in fact they are not. Thus, women should be particularly interested in detecting a man's mating strategy (Buss, 2003).

Not only dating in general but online dating as well has excited some research interest—among others, also with respect to rsc and hp (for an overview, see Finkel et al., 2012). It has been assumed, taking the hp perspective, that the cmc limitations in online dating can be compensated by language style and choice of words (Walther et al., 2005). While physical cues are missing in cmc, the importance of verbal cues might be rising. The question then might very well be, this time with respect to online dating: what about single words instead of entire texts?

As emphasized above, communication only by means of single words is even more limited than communicating through written texts. Still, those single words might communicate crucial information (Lange et al., 2016a). In accordance with findings on mate choice in “real life,” Whitty and Buchanan (2010) found that women were more attracted to online user names (hereinafter called nicknames) (e.g., in terms of the motivation to contact the person behind the name) that signaled intelligence, while men were more attracted to nicknames indicative of physical attractiveness. So the choice of a nickname in online dating can be used for impression management—just like hp would predict. Online dating is indeed an area in the digital world in which making a good first impression is essential (Whitty and Buchanan, 2010).

Apart from classical mate choice criteria, the personality of a potential mate is crucial, too (e.g., Buss et al., 1990; Botwin et al., 1997; Escorial and Martín-Buro, 2012). In this context, research by Back et al. (2008) is particularly relevant for the research presented in the article at hand. They retrieved personality scores of 599 participants (Big Five, e.g., extraversion; narcissism) and additionally asked them for their e-mail addresses. Back et al. (2008) then presented the e-mail names to 100 participants who judged the personality dimensions of the e-mail name users on the same personality items used before. Personality dimensions were detected correctly, with results being statistically significant for all dimensions except for extraversion. Back et al. (2008) also showed that personality ratings were linked to certain attributes of the e-mail address. For instance, the perception of conscientiousness was positively correlated with both the number of characters and dots the names consisted of, while number of digits was negatively correlated with it.

The current study had the objective of replicating the findings by Back et al. (2008) with respect to online dating as well as to extend them. Back et al. (2008) used e-mail names and had a general cmc context. We, on our part, wanted to focus more on nicknames. This was inspired by research on the psychology of pseudonyms (e.g., Lange et al., 2016b) as well as based on the following assumption: While e-mail addresses are often created based on the rule “first name.last name” (e.g., john.smith@…), nicknames are assumed to be more creative (cf. Whitty and Buchanan, 2010). Also unlike Back et al. (2008), we were interested in the context of online dating and mate choice. Whitty and Buchanan (2010) have already shown that such an approach is worthwhile. Still, the scarcity of such research calls for more studies of this kind.

The question might also be asked, as to whether people are able to detect the mating strategy of a potential mate. It was also of interest whether the motivations for contacting a person behind a nickname, based only on the nickname, might differ (Whitty and Buchanan, 2010). Furthermore, we wanted additionally to investigate whether one of the two sexes are better at judging women's and men's personality based on their nicknames. Mating is an area of social life, where making a proper choice seems particularly important (Buss, 2003). So, it seemed of practical relevance to elucidate what mate choice-relevant information can be retrieved form an online dating nickname.

Finally, we were interested in the linguistic features of the nicknames, and the subsequent question whether we would find correlations between these features and other variables of interest (Back et al., 2008; Lange et al., 2016b).

We proposed the following hypothesis (cf. Back et al., 2008):

H1: People are able to correctly guess online daters' personality by means only of their nicknames. Under personality, we understood the Big Five dimensions which are: openness to experience, conscientiousness, extraversion, agreeableness, and neuroticism (McCrae and John, 1992). The Big Five have been used quite often in research focusing on personality perceptions by means of certain cues (e.g., Küfner et al., 2010; Qui et al., 2015). As another personality dimensions, we added narcissism following the mentioned study by Back et al. (2008). Other researchers have also included this trait, which is one of three traits of the so-called Dark Triad, into their research in order to elucidate, whether it can be detected (e.g., Buffardi and Campbell, 2008; Vander Molen et al., 2018).

Furthermore, we had four research questions that were derived from mate choice research (see above) and other studies on the psychology of nicknames or usernames (Back et al., 2008; Whitty and Buchanan, 2010; Lange et al., 2016b):

RQ1: Are people able to correctly guess online daters' mating strategy by means only of their nicknames?

RQ2: What are the correlates of the motivation to contact a person behind a nickname?

RQ3: Does one sex show greater accuracy in personality judgments than the other?

RQ4: What are the linguistic correlates of the personality of the nickname users and how are they perceived? In other words, are linguistic features significant mediators of judgments?