
Last week the Heritage Foundation released my new study on the effect of state policies easing access to puberty blockers and cross-sex hormones on youth suicide rates. The study generated a large amount of attention from policymakers, the traditional media, and on Twitter. Given that the topic is a highly emotional and politicized one and given the low quality of discourse on Twitter, much of the social media response was inaccurate and ad hominem. Fortunately, Twitter is not the real world and the serious response of a number of policymakers with steps they are taking to address the risks posed by these drugs makes responding to Twitter critics pointless.
The reactions of two prominent commentators, Jack Turban and Jesse Singal, however, warrant a response, not for the merit of their criticisms but because they have enough influence outside of Twitter that their mistaken criticisms might undercut the positive policy responses my study has facilitated. Turban is a professor at Stanford Medical School and is the author of two of the three studies claiming that puberty blockers and hormones are protective against suicide and therefore must be made widely and readily availably. Singal is a journalist who has engaged in extensive criticism of Turban’s work and is therefore someone that skeptics of these medical interventions might look to for guidance on what to think about my new study.
Other than dismissive and ad hominem comments, Turban’s main objection to my research is that minors are not supposed to be getting access to these drugs without parental consent so that looking at variation in state policies regarding the ability of minors to access health care without parental consent would be irrelevant. He writes, “One thing to note is that @TheEndoSociety and @wpath guidelines require parental consent to access gender-affirming hormones. This entire report is based on the incorrect assumption that minors can easily access hormones without parental consent.” He adds, “Since trans youth account for about 1.9% of teens, only a fraction of those desire hormones, only a few percent of those are able to access them, and of those even fewer would access without parental consent, the logical jump made by the Heritage people doesn’t make sense.”
I never claim that “minors can easily access hormones without parental consent.” My study is based on the natural policy experiment that results from some states having one fewer barrier to minors getting these drugs by having a provision in law that allows minors to access healthcare without parental consent, at least under some circumstances. It just has to be easier for minors to get these drugs, not that they can do so “easily.”
But Turban seems to suggest on Twitter that this virtually never happens. Where would I get the idea that minors can access these drugs without parental consent? It comes from Turban’s own research. In his 2022 study on the effects of cross-sex hormones on suicidal ideation, he analyzes the results of a survey given to a convenience sample of adults who identify as transgender. In Table 1, he reports that 3.7% of those who say that they received cross-sex hormones between the ages of 14 and 17 are still not “out” to their families as transgender. These respondents must have obtained those drugs without parental consent since their parents do not even know that they identify as transgender. In addition, we see in that same table that only 79.4% of those adults who got hormones between the ages of 14 and 17 say that their families are supportive. We might reasonably assume that unsupportive families would not have given consent to their teenage children getting hormones, especially if they continue to be unsupportive several years later when those children are now adults. Yet somehow, nearly a fifth of those who got the drugs as teenagers did so despite the lack of support from their families.
If we accept Turban’s claim that 1.9% of teenagers identify as transgender, that translates into 1,900 out of every 100,000 teenagers. My finding is that easing access by reducing the parental consent barrier increases youth suicide rates by 1.6 per 100,000 young people. For my result to be plausible, it would only have to be the case that .08% of teenagers who identify as transgender would have to seek these drugs, find that it is easier to access them when states have minor consent provisions, and then kill themselves to result in an additional 1.6 suicides per 100,000 young people. And this assumes that the entirety of the increase in suicides occurs among individuals who identify as transgender and get the drugs, even though we know that there is a contagion effect with suicide so that the 1.6 increase would include young people who did not get the drugs but were influenced by the deaths of others.
The bottom line is that the magnitude of my study’s estimated increase in suicide risk associated with these drugs is entirely plausible given Turban’s own numbers about the frequency with which minors are accessing these drugs without parental consent.
Other than name-calling with words like “misleading,” “crude,” and “shodd[y],” Singal has two seemingly substantive objections to offer. First, he claims “this isn’t even about blocker and hormones — it’s about which states have lower ages of medical consent. Then he says ‘Well, around the time blockers became available, the suicide rates go up.’ This is an exceptionally crude approach.” Second, he embraces a criticism expressed by Elsie Birnbaum that it is ridiculous to describe states like Texas and Utah as having “easier” access to these drugs, adding: “This is a good catch and should immediately cause anyone with any knowledge of this subject to deeply question the study.”
With both of these criticisms, Singal appears to think that the proper way to study the effects of puberty blockers and cross-sex hormones would be to compare the suicide rates in places based on the number of prescriptions being dispensed, the number of clinics offering these treatments, or their state reputation as being more or less permissive on transgender issues. While I understand why it is tempting to think that these direct comparisons would be better, if our goal is obtaining an unbiased, even if imprecise, estimate of causal effects, it is far better to focus on state minor access provisions. Because modern research designs for isolating causal effects are not necessarily intuitive, let me try to offer a brief explanation for readers (and Singal) who are not trained researchers.
The gold-standard research design for isolating causal effects is a randomized controlled trial (RCT), in which a lottery would assign some people to get the drugs and others not to get the drugs. Researchers would then compare the outcomes for the two groups over time. Any significant differences in their outcomes would have to be caused by the drugs and not by some preexisting difference between the treatment and control groups, since the two groups would be identical, on average, at the start of the experiment. Unfortunately, the effects of puberty blockers and cross-sex hormones in treating what is called gender dysphoria has never been studied with an RCT. Turban, the Biden administration, and others claiming with confidence that these drugs save lives should support an RCT to prove what they say, but they do not, and we are left without the kind of rigorous evidence that is normally required for initial approval of drugs by the FDA.
Short of an RCT, there are a number of research designs that have been developed that attempt to imperfectly approximate the causal rigor of a true experiment. They do so by looking for ways in which exposure to treatment is “exogenous.” That is, they look for why some people get the drugs while others do not for reasons that are unrelated to factors that might separately influence the outcomes. A lottery is perfectly exogenous because chance determines who gets the treatment and chance has nothing to do with causing outcomes. If the reasons that some people get the treatment while others are in the control are related to the outcomes, however, then we have an “endogeneity” problem and the results are biased.
It is easy to illustrate this endogeneity problem in Turban’s research on this issue. Turban examines a survey of adults who identify as transgender and compares those who sought and got these drugs to those who sought but did not get the drugs in terms of their more recent thoughts about suicide. One of the reasons some people who sought these drugs would be unable to get them would be if they were not considered psychologically stable, since being psychologically stable at the time is supposed to be a criteria for prescribing the drugs. Rather than being random and unrelated to later outcomes, the reason that some people end up in Turban’s treatment or control groups is caused by their psychological condition when they sought treatment, which would be related to the suicide outcomes being measured — or endogenous. It is obviously biased and unhelpful to compare treatment and control groups that begin with different average psychological health in terms of their later psychological outcomes.
The same endogeneity problem applies to how Singal seems to think this issue should be examined. We know that there is significant comorbidity between gender dysphoria and other challenges that young people have, including depression, anxiety, and autism spectrum disorder. To the extent that demand for puberty blockers and cross-sex hormones is related to people having other psychological challenges, comparing places based on the number of prescriptions or clinics would be endogenous and misleading. It would be biased by the likelihood that places with more of these drugs being dispensed would also be places with a higher prevalence of other psychological challenges, which would be related to suicide rates in those places independent of whether the drugs helped, hurt, or made no difference. Similarly, comparing states based on whether they had reputations for being permissive or “blue” states would be endogenous and misleading. It would be comparing treatment and control groups that differed at the start in ways that are related to suicide outcomes.
To find something closer to the true causal effect, we would need exogenous sources of variation in the treatment other than a lottery. My study takes advantage of plausibly exogenous variation in exposure to treatment with respect to WHERE there is a lower barrier to accessing treatment, WHEN that treatment is available, and WHO is affected by the treatment. States adopted policies about the ability of minors to access healthcare without parental consent for reasons that had nothing to do with, and generally long preceded, the transgender issue. On the margin, parental consent is one additional barrier to minors getting puberty blockers and cross-sex hormones.
Singal believes that it is a defect of my study that this variation in minor consent policies is not “about blocker and hormones,” but he fails to understand this is a virtue of the research design. Because minor consent provisions are a barrier to accessing puberty blockers and cross-sex hormones that is not “about” this issue, variation in the existence of this barrier is exogenous and helps isolate causal effects. It’s true that minor access provisions are not the most important or direct barrier to access, but they help generate unbiased effects. To the extent that these provision are entirely unrelated to the issue, they would be random noise and would bias effects toward zero but would not bias the direction of the results.
In addition to exogenous variation regarding where these drugs could be accessed with or without an extra barrier, we have exogenous variation in when the drugs became available. This is why it is important that we observe that there is no difference in youth suicide rates between states with or without minor access provisions before the drugs are introduced but there is after.
Lastly, we have exogenous variation in who would be affected. If states with a minor access provision began to differ systematically with respect to suicide only after 2010, we should observe this pattern also among a slightly older population that would not be affected by minor consent provisions. The fact that there is no effect for a slightly older population is also important.
Obviously, it would be far better to have an RCT if we wanted to isolate the causal effects of puberty blockers and cross-sex hormones on youth suicide. But absent an RCT, my study uses quasi-experimental research design features that generate credibly causal effects. It’s imperfect, but it is a huge improvement over the obviously endogenous research design that Turban and Green use and much better than the direct but endogenous approaches that Singal criticizes my study for lacking.
Channeling all of his Twitter erudition and a penchant for research nihilism, Singal asserts that my study is no different in its defects from prior ones: “The dude makes perfectly fair comparisons of some of the past work on this subject, most notably Jack Turban’s, and then he reaches into basically the same bag of tricks. It’s SUCH a bad article.” While I like being compared to The Dude, claiming that my study is comparable to those by Turban and Green is just incorrect.
Twitter is a dangerous place for young people with gender dysphoria, but it is also a dangerous place to discuss the merits of different studies. If Turban or Singal were willing, I’d be happy to get together in a public forum to discuss these issues at greater length. An audience would benefit far more from such a discussion than Twitter name-calling and drive-by research critiques.
ADDENDUM
I’ll respond to one more substantive objection that Jesse Singal echoes and that was raised initially by Dave Hewitt, an English substacker. Hewitt claims that my results are sensitive to outlier states, such as Alaska or Wyoming, that have above-average youth suicide rates. He attempts to illustrate this concern by switching whether AK and WY are classified as having a minor access provision or not. He then produces a graph that shows the unadjusted difference in suicide rates between states based on the existence of a minor access provision shrinks to zero if both AK and WY are switched in how they are classified.
Importantly, Hewitt only shows us the unadjusted difference, not the final results adjusting for baseline differences in state suicide rates, as displayed in Chart 3 and Appendix Tables 2-5 in the study. States differ in their average suicide rate across the entire time period studied, including the years before puberty blockers and cross-sex hormones were introduced as a therapy for gender dysphoria around 2010. Hewitt is aware of this fact when he notes, “The suicide rate of this age group in Alaska is far higher than any other state… Wyoming has the highest overall suicide rate in the country across all age groups.”
Because there are time-invariant factors that might make some states have higher or lower youth suicide rates, my analysis controls for each state’s suicide rate at baseline. And to the extent that there are time-variant but age-invariant factors that affect the change in suicide rates over time, I control for the annual suicide rate in each state in each year for a slightly older population that should be unaffected by a minor access provision.
Yes, Alaska and Wyoming have high suicide rates and switching states with high rates from the treatment to the control group would alter the unadjusted difference between those groups of states. But this is irrelevant to the question of whether states experience a change in youth suicide rates when cross-sex treatments become available based on having one fewer exogenous barrier to minors accessing those treatments — especially when we control for time-invariant and age-invariant factors that make the rate higher in some states than in others.
Switching states from one category to the other is also the wrong way to test whether the results are sensitive to one or two states. The proper way to test for sensitivity would be to run the regression with all of the controls and to drop individual states from the analysis to see if the result still holds. If any single state is driving the result, then dropping it from the analysis should substantively change the result.
I’ve done this and Hewitt’s (and Singal’s) concerns about sensitivity to outlier cases are unfounded. If I drop Alaska from the analysis presented in Appendix Table 2, the result remains unchanged. If I drop Wyoming, the result remains unchanged. If I drop Alaska and Wyoming at the same time, the result remains unchanged. In fact, I’ve dropped each of the 50 states and DC one by one and the results remain statistically significant and virtually identical in magnitude across all 51 robustness checks.
I’m working on revising this “working paper” and submitting to a peer-reviewed journal and will comply with the replication data set policies of that journal. In the meantime, Hewitt, Singal, or anyone else interested in replicating my analysis and trying other robustness checks can easily do so by downloading and analyzing the data. The study lists the handful of data sources and provides links. The full model specifications are also provided in the appendix tables.