Blog


Consulting as Private-Sector Graduate School

July 28, 2022

As an undergraduate student in the sciences at MIT, contempt for management consulting was commonplace. Consulting was the path for people who had ambition devoid of any real interests, the “sellout road” where you made endless Powerpoints instead of providing any tangible improvement to the world. In contrast, going to graduate school was a choice that showed commitment and integrity. If you were willing to sacrifice your 20s in service to a scientific discipline, that showed true passion and an honorable commitment to the field.

I’m now midway through my PhD, and I’ve come to the conclusion that my previous impressions were mistaken and that consulting and graduate school are in fact more alike than they seem. This change has been spurred by making new friends from the world of management consulting and realizing not only that they enjoyed and benefited greatly from their experience, but also that their experience seemed broadly similar to mine.

This essay is an attempt to outline some similarities and differences between consulting and graduate school, speculate about why these differences exist, and finally determine if graduate schools can learn anything from the consulting model. My tentative conclusion is that, at the margin, research groups would benefit from acting more like consultants and directly solving industry-relevant problems for pay.

Epistemic status: moderate to low. These thoughts are based mainly on my own experience as a chemistry PhD student, and likely do not translate to the humanities or social sciences. I also have only a secondhand knowledge of management consulting and so probably suffer from myriad misconceptions.

Similarities and Differences

At a superficial level, management consulting and graduate school both fill the same role: a safe and prestigious opportunity for a new graduate to diversify his or her skills and accrue “career capital.” In consulting, much like in graduate school, learning is key: in the words of Pete Buttigieg, McKinsey was “a place where I could learn as much as I could by working on interesting problems and challenges.” Both occupations can segue smoothly into a variety of opportunities afterwards, in part due to the shared emphasis on connections, networking, and presentation skills, and as a result both professions attract a steady stream of bright, highly motivated people. Easy access to human capital seems to be a shared prerequisite: without a supply of new graduates willing to work long hours in a high-stress environment, neither BCG nor Harvard could survive.

Given the plethora of interesting, well-paid opportunities available for high-achieving graduates, the popularity of these more grueling professions might be surprising. In her essay “Harvard Creates Managers Instead of Elites,” Saffron Huang describes the thought process behind why so many of her Harvard classmates took “the decreasing returns of another NGO internship or McKinsey job” over more inventive careers, concluding that “school-validated options” appeal to students who are “naïve and uncertain about [their] own futures.” In other words, the safety of taking a job well-known to be prestigious is what makes consulting and similar options so appealing.

Although Huang doesn’t mention graduate school specifically, I would argue that staying within academia is the most “school-validated” of all choices. Getting a PhD not only gives one a defensible claim to domain expertise and a chance at a higher-status job but also allows students to stay within the familiar academic system for longer. Faced with all the manifold diversity of the private sector, the chance for a graduate to stay within the familiar confines of the university for a few more years is a safe and socially acceptable way to delay one’s arrival into corporate America. And the high status that professors have in the eyes of undergraduates only strengthens the appeal of graduate school: if all one’s academic role models went down this path, surely it can’t be a bad choice.

From the perspective of the student, one obvious difference is the pay: a typical starting consulting salary is $100k, while my Harvard graduate student stipend is currently $43k. Given that a first-year consultant and a first-year graduate student have essentially the same skills (i.e. what you’d expect from an undergraduate education and not much more), this difference is surprising. Based on anecdotal reports from consulting, my intuition is that this difference is not limited to salaries: the consulting world is flush with cash, while the academic world often runs on the verge of bankrupcy.1

(I’m intentionally avoiding questions around the ethics of consulting because I think it’s not particularly relevant to this piece, and because I don’t think I have any unique insights on this topic.2)

Academia’s Unique Niche

Why does consulting have so much more money than academia? One simple model of academia is as follows: discoveries that provide “present value” can easily be funded by companies, because there’s a quick return-on-investment. On the other hand, discoveries that provide “future value” are hard to fund through the private sector, because there’s no guarantee that the real-world value will be captured by the funder. Accordingly, the government sponsors research into interesting problems with uncertain timeframes to do what the free market cannot.3 This comports with what Vannevar Bush wrote in his landmark 1945 work Science, The Endless Frontier:

New impetus must be given to research in our country. Such impetus can come promptly only from the Government…. Further, we cannot expect industry adequately to fill the gap. Industry will fully rise to the challenge of applying new knowledge to new products. The commercial incentive can be relied upon for that. But basic research is essentially noncommercial in nature. It will not receive the attention it requires if left to industry.

Viewed within this model, we might hypothesize that consulting is lucrative because it’s easier to finance providing present value than providing future value (or because the free market is more efficient than the NIH/NSF). But this picture is oversimplified. Much current chemistry research at least ostensibly addresses present problems in the chemical industry, and research groups frequently collaborate with (and receive money from) chemical companies. Why, then, is consulting better at capturing returns on present value than academia?

Structural factors disincentivize academic labs from acting as consultants.4 Harvard’s stated policy on academic–industrial collaborations involving specific deliverables is that they are discouraged, allowed “only if the activity in question advances a core academic mission of the faculty member’s school and either provides a significant institutional benefit or a public benefit that is consistent with the University’s mission and charitable status.” This matches my experience collaborating with Merck, a pharmaceutical company; it was clear that we were not accepting money for rendering Merck a service, but instead simply working together because our intellectual interests aligned. Although we did receive some money, it was a fraction of what our total costs in salary, materials, etc were for the project.

Policies like this prevent companies from hiring research labs on a purely transactional basis, forcing academics to decouple their incentives from those of industry. Even if an academic lab is running out of money, it must find some way to justify its collaborations beyond pure economic necessity: research groups cannot simply remake their interests to suit whichever employer they want to attract. Viewed within the above model, this is good! Academia is supposed to focus on problems that can’t be solved by industry, not act as a contractor in service of corporate profits.

Yet the preponderance of academic–industrial collaborations suggests that academia’s ostensible focus on long-term projects is not as strong as it could be. In a world where funding for basic research seems to be declining on a per-lab basis, it is perhaps unsurprising that professors turn to alternate sources of funding to keep their labs afloat; moving forward, we can expect this trend only to intensify.

Perhaps the biggest omission from the above discussion is another key role of academia: training students. Graduate school, after all, seeks not only to advance the frontiers of human knowledge but also to train students in this pursuit. But from the perspective of the typical graduate student, it strikes me as unlikely that the specific nature of the problems under study (i.e. purely academic versus industrially relevant) has a massive impact on the student’s learning. Indeed, many students might be better prepared for their careers by having more encounters with industrial problems and techniques. The existence of current industrial postdoctoral positions suggests that gaining scientific experience through industry-relevant problems can be a successful strategy.

Conclusions

Although the idealized model of the university—a place dedicated to advancing long-term human flourishing through the pursuit of knowledge “without thought of practical ends”—is indeed utopian, the present problems with academic funding suggest that a more pragmatic outlook may be needed in the short term. In particular, finding new ways to efficiently fund scientific research and education is a pressing challenge for the field (absent major changes to the funding ecosystem) which remains, from my point of view, unsolved.

Accordingly, the consulting model presents an interesting alternative to the current system. Consulting firms sustain themselves solely by providing solutions to current problems in industry, training their “students” without any need for external subsidies. Is it possible for research groups to support part-time basic research by consulting the rest of the time? At the margin, should graduate schools be more like consulting firms? This approach would require reducing the stigma around research groups acting as contractors, and in so doing perhaps run the risk of lessening the prestige of the university. On the other hand, directly applying university knowledge to solving practical problems might raise public appreciation for science.

We may see the results of this experiment sooner rather than later. As acquiring scientific funding continues to grow more difficult, I expect that smaller, more poorly funded departments will begin to pursue money from industry more aggressively to keep themselves afloat, moving more and more towards the consulting model out of necessity. Time will tell whether this proves to be an alternate, or even superior, model for funding research, or a negative development that undermines what makes universities distinctive.

If forced to guess, my tentative prediction would be that these changes will be good. The present funding model seems wasteful and unsustainable, a relic of massive growth in federal science funding over the past 100 years. A correction is coming, and it will be brutal when it does. Finding new ways to fund research beyond just federal grants, then, is important for the future of research in the US; it’s been done before, and it can be done again. In fact, some of the greatest scientific discoveries have originated not from universities but from the corporate sphere! Disrupting our institutions of science will be painful, but I think the potential upside is high—that is, if academic researchers can accept corporate money while still preserving some ability to pursue basic science.

Another conclusion from this area of thinking is that federally funded scientists ought, as much as possible, to focus on their area of comparative advantage—long-term research with uncertain payoffs, “essentially noncommercial” in nature. At least in organic chemistry, most funding applications that I’ve seen are very careful to point out how their discoveries could lead to immediate deliverables with practical impact.5 If these claims are really true, then these discoveries should be funded by the private sector, not by federal money. These assertions may be part of what a competitive grant application today requires, but their existence seem to point to a fundamental disconnect between what academic research is and what it should be.

Footnotes

  1. It's tough to find sources on this, but anecdotally even at top universities most research labs seem strapped for cash. For structural discussions, see this NPR article and this New Science report.
  2. This issue has been discussed a lot: one particularly influential piece in this area is Daniel Markovits's “How McKinsey Destroyed The Middle Class”.
  3. This is equivalent to saying that scientific progress is a public good. There are more ways that things could be public goods than just long timeframes, but without loss of generality we’ll elide these considerations here.
  4. Many professors do serve as consultants for industry, but they generally do this apart from the university, without involving their students, and the money goes to them personally, not to their research groups.
  5. The New Science NIH report discusses this phenomenon at length: over the past 20 years, the NIH has changed its standards, such that now new grants are expected to have “clear research goals with obvious practical applications.”

Timescales

July 19, 2022

For many organic chemists, it’s hard to grasp the vast difference between various “fast” units of time. For instance, if a reactive intermediate has a lifetime of microseconds, does that mean it can escape the solvent shell and react with a substrate? What about a lifetime of nanoseconds, picoseconds, or femtoseconds?

To help answer these questions for myself, I made the following graphic about a year ago, which compares the timescale of various molecular processes on a logarithmic axis. Although someone skilled in Adobe Illustrator could doubtless make a prettier version, I've still found this to be a useful reference, and frequently use it as a slide in talks or group meetings:

illustration of chemical timescales
Illustration of timescales of chemical processes—open in new tab for higher resolution

Based on this graphic, it becomes easier to think about the interplay between competing fast processes. Species that persist for less than ~5 ps, for instance, are effectively “frozen” in the solvent configuration they’re formed in, whereas molecules that persist for longer can sample different solvent configurations. If a species can persist for 10-100 ps, it can begin to sample the nearby conformational landscape through low-barrier processes (e.g. single-bond rotation), although larger conformational changes might still be inaccessible.

As lifetimes stretch into the nanosecond regime, diffusion and solvent-cage escape become more realistic possibilities: based on Mayr’s value for the “diffusion limit” (2–4e9 M-1s-1), we can estimate that bimolecular association with a 1.0 M reactant will take 200-400 ps, while association with a 100 mM reactant will take 2-4 ns. On the far right of the graph, being able to distinguish different species by NMR (e.g. amide methyl groups in DMF) means that these species have a very long lifetime indeed.

Framed in these terms, then, it becomes obvious why the 10-100 ps timescales currently accessible by ab initio molecular dynamics (AIMD) are unable to model many important molecular processes. Indeed, work from Grossman and coworkers has shown that the results of AIMD simulations can be very dependent on the pre-AIMD equilibration method used, since the actual solvent environment is unable to fully relax over the timescale of the simulation. For AIMD to become a truly useful alternative to forcefield molecular dynamics, much faster ab initio methods will be needed!

Sources:

Thanks to Richard Liu and Eugene Kwan for feedback on this figure.

Corrections (updated 7/25/2022):

The Apotheosis of 2D Infrared Spectroscopy

July 8, 2022

While IR spectroscopy is still taught in introductory organic chemistry classes, it has been almost completely replaced by NMR spectroscopy and mass spectrometry for routine structural assignments. Still, IR spectroscopy offers unique advantages to the mechanistic chemist: the short timescale of IR allows for the observation of transient molecular interactions even when slower techniques like NMR only yield time-averaged data, and IR absorbances can easily be perturbed by isotopic substitution while leaving the underlying potential-energy surface unchanged.

These advantages are nicely illustrated in the IR spectrum of a mixture of phenol, benzene, and CCl4, which shows two peaks corresponding to free phenol (2665 cm-1) and phenol complexed to benzene (2631 cm-1). (The phenolic proton was replaced by deuterium, moving the O–D stretch away from C–H stretches and into a vacant region of the spectrum.) From the standpoint of assessing purity, it might be upsetting that a pure compound shows two peaks; from the standpoint of a mechanistic chemist, the ability to distinguish two different solvation environments experimentally is incredible.

IR spectrum of phenol in benzene/CCl4
IR spectrum of phenol in benzene/CCl4 (Fayer Figure 2).

Unfortunately, measuring this spectrum tells us about the thermodynamics of the equilibrium, but not the kinetics; there’s not a good way to determine how fast these two species are exchanging from these data.1 In 2005, Fayer and coworkers developed a pump–probe infrared spectroscopy method called “2D IR” to tackle this problem. In 2D IR, the system is excited, allowed to evolve for a variable length of time Tw, and then triggered to emit a vibrational “echo” (in analogy to spin-echo NMR experiments) which still contains phase information from the original excitation. (There are a lot of non-trivial spectroscopic details here which I don’t really understand.)

2D IR spectrum of phenol in benzene/CCl4
2D IR spectrum of phenol in benzene/CCl4 (Fayer Figure 3).

The net result of this is a two-dimensional plot showing initial and final frequencies, in which cross-peaks represent molecules which have moved between one state and another during Tw. By surveying a range of Tw values, the kinetics of exchange can be quantitatively determined: in this case, the time constant τ was found to be 8 ± 2 ps. This result might not seem thrilling (“fast exchange is fast”), but this experiment can be used to measure rates of phenol dissociation from electronically-varied aromatic rings, or compared to results from molecular dynamics simulations for benchmarking purposes.

While many groups are now using 2D IR, this recent paper from Tokmakoff and coworkers studying superconcentrated electrolytes stood out to me as particularly exceptional. In superconcentrated solutions like those found in batteries (e.g. 15 M LiTFSI in acetonitrile), the extreme salt concentration leads to high viscosity and substantial aggregation, leading to questions about how charge transport in batteries occurs. Some simulations seem to suggest that, rather than “vehicular diffusion” wherein a cation diffuses along with its solvent shell, charge transport occurs through “structural diffusion” involving breaking/reforming of cation–solvent interactions. (This is analogous to the Grotthuss mechanism of proton transport in water.)

different transport mechanisms
Illustration of different transport mechanisms (Tokmakoff Figure 8).

Since distinct C≡N stretches are visible for cation-bound and free acetonitrile, it might seem straightforward to simply measure time evolution of the cross-peaks and thereby determine the rate of solvent exchange. Unfortunately studying exchange in the bulk solvent is complicated by the fact that direct vibrational energy transfer can occur through collisions, meaning that cross-peaks are observed even in the absence of exchange. The authors solve this problem by using a mixture of D3CCN and D3C13CN: while cross-peaks between the heavy and light isotopologues can only occur through energy transfer, cross-peaks between the same isotopologue can also occur through chemical exchange.2

2D IR of electrolyte
2D IR measurement of 1.9 M LiTFSI in acetonitrile (Tokmakoff Figure 5).

They find that the time evolution of all cross-peaks is identical under all conditions, indicating that solvent exchange must be slower than energy transfer (~20 ps) for any cation or concentration studied. This suggests that, contrary to a variety of theoretical studies, structural-diffusion mechanisms for cation transport are quite slow and unlikely to be relevant for these electrolytes.

This study is a beautiful example of designing a cutting-edge spectroscopic experiment to solve a key scientific problem, and reminds me how much we still don’t know about “simple” systems like ionic solutions. I would love to see techniques like this applied to study reactive intermediates in the ground state, e.g. Olah-style superacid solutions! More broadly, it’s exciting to see how 2D IR can advance in less than two decades from being limited to simple model systems to now being used to tackle the biggest open questions in chemistry. What new techniques being developed today will rise to prominence in the coming decades?

Thanks to Joe Gair for reading a draft of this.

Footnotes

  1. For a different way to extract rates from vibrational spectra, see this work on line-broadening analysis.
  2. Deuterated acetonitrile is used owing to complications with Fermi resonances in protio-acetonitrile.

Four Handy Bash One-Liners

July 6, 2022

In my experience, most computational chemists only know a handful of basic Bash commands, which is a shame because Bash is incredibly powerful. Although I'm far from an expert, here are a few commands I frequently find myself using:

1. sed For Find-and-Replace.

$ sed -i “s/b3lyp/m062x/” *.gjf

If you want to resubmit a bunch of transition states at a different level of theory, don't use a complex package like cctk! You can easily find and replace text using sed, which runs almost instantly even for hundreds of files. (Note that the syntax for modifying in-place is slightly different on macOS.)

2. Renaming Lots of Files

$ for f in *.gjf; do mv $f ${f/.gjf/_resubmit.gjf}; done

Unfortunately, you can't rename lots of files with a single command in Bash, but using a for; do; done loop is almost as easy. Here, we simply use parameter expansion to replace the end of the filename, but the possibilities are endless.

3. Counting Occurrences in a File

$ for f in */*.out; do echo $f; grep "SCF Done" $f | wc -l; done

Here we again use a for loop, but in this case we use grep to search for the string "SCF Done". We then pipe the output of this search to the wc -l command, which counts the number of lines. Since grep returns each result on a new line, this prints the number of optimization steps completed.

4. Cancelling Jobs By Matching Name

$ squeue -u cwagen | grep "g16_ts_scan" | awk '{print $1}' | xargs -n 1 scancel

Although the slurm workload manager allows one to cancel jobs by partition or by user, to my knowledge there isn't a way to cancel jobs that match a certain name. This is a problem if, for instance, you're working on two projects at once and want to resubmit only one set of jobs. Here, we use squeue to get a list of job names, search for the names that match, extract the job number using awk, and finally cancel each job by building the scancel commands with xargs. (This should be easily modifiable for other workload managers.)

Plenty of Room at the Bottom

June 28, 2022

This thesis, from Christian Sailer at Ludwig Maximilian University in Munich, is one of the most exciting studies I’ve read this year.

Sailer and coworkers are able to generate benzhydryl carbocations from photolysis of the corresponding phosphonium salts, and can monitor their formation and lifetime via femtosecond transient absorption spectroscopy. (There are some technical challenges which I’ll omit here.) They then use this platform to study the addition of alcohols to these cations and obtain nice kinetic data on some shockingly fast reactions:

rates of reactions with electrophiles
Rates of reaction between methanol and various cations (page 77).

Although these species are extremely reactive, substituent effects are still paramount: adding two methyl groups to stabilize the benzhydryl carbocation extends its lifetime by 6.3x, whereas adding two electron-withdrawing fluorines shortens its lifetime by 5.6x. The most electrophilic species studied, (dfp)2CH+, reacts with methanol in only 2.6 ps! These findings demonstrate how even an extremely reactive carbocation like Ph2CH+ doesn’t react at the rate of diffusion with nucleophiles; although this reaction is almost instant on an absolute scale, there are still tens or hundreds of unproductive alcohol–carbocation interactions before product is finally formed.

In contrast to Sailer’s results on the electrophile, which align nicely with results from more conventional kinetic measurements, different alcohols behave very differently:

rates of reactions with nucleophiles
Rates of reaction between mfp(Ph)CH+ and various alcohols (page 75).

This rate difference is surprising, since Mayr’s measurements demonstrate that there’s no conventional difference in nucleophilicity between these species—and intuitively, one wouldn’t expect adding additional carbons to substantially hinder approach to the oxygen. However, as Sailer notes, larger molecules rotate and reorient much more slowly than smaller molecules. This is evident from a number of different physical properties: larger alcohols form more viscous liquids and have slower dielectric relaxation times (how long it takes them to reorient in response to a new charge).

alcohol viscosity (ref) dielectric relaxation time (ref) reaction time (above)
methanol 0.54 cP 5.0 ps 22 ps
ethanol 1.08 cP 16 ps 41 ps
n-propanol 1.95 cP 26 ps 62 ps
iso-propanol 207 cP 106 ps

Sailer thus concludes that, as the timescale of reaction approaches the timescale of molecular motion, nucleophile reorientation becomes rate-limiting. This is a really cool conclusion, and highlights how the environment that reacting molecules “see” differs from our macroscopic chemical intuition: although we think of reorientation as barrierless, for very fast reactions reorientation is actually slower than bond formation. (Philosophically related, in my mind, is Dan Singleton’s work on non-instantaneous solvent relaxation, and how this influences the stability of reactive intermediates.)

Thus, this work simultaneously highlights both similarities and differences between ultrafast reactions and their benchtop congeners. My hope is that future studies can find ways to move beyond just benzhydryl cations and study elementary questions of selectivity with different nucleophiles (e.g. addition/elimination). The ability to observe what is typically the “invisible” step in SN1 is incredibly powerful, and I think we’ve barely scratched the surface with what we can learn from measurements like these.