Natalie S.
Eyke
a,
Timo N.
Schneider
a,
Brooke
Jin
a,
Travis
Hart
a,
Sebastien
Monfette
b,
Joel M.
Hawkins
b,
Peter D.
Morse
b,
Roger M.
Howard
b,
David M.
Pfisterer
b,
Kakasaheb Y.
Nandiwale
b and
Klavs F.
Jensen
*a
aDepartment of Chemical Engineering, Massachusetts Institute of Technology, Cambridge, MA 02139, USA. E-mail: kfensen@mit.edu
bPfizer Worldwide Research and Development, 445 Eastern Point Rd, Groton, CT 06340, USA
First published on 4th August 2023
We present an automated droplet reactor platform possessing parallel reactor channels and a scheduling algorithm that orchestrates all of the parallel hardware operations and ensures droplet integrity as well as overall efficiency. We design and incorporate all of the necessary hardware and software to enable the platform to be used to study both thermal and photochemical reactions. We incorporate a Bayesian optimization algorithm into the control software to enable reaction optimization over both categorical and continuous variables. We demonstrate the capabilities of both the preliminary single-channel and parallelized versions of the platform using a series of model thermal and photochemical reactions. We conduct a series of reaction optimization campaigns and demonstrate rapid acquisition of the data necessary to determine reaction kinetics. The platform is flexible in terms of use case: it can be used either to investigate reaction kinetics or to perform reaction optimization over a wide range of chemical domains.
Parallelization has been incorporated into several platforms as a way to increase throughput.12,14 When parallelization is coupled with miniaturization, large quantities of reaction information can be generated rapidly using small quantities of material, and the resulting expansion in the design spaces that can be screened is transformative. Platforms that study hundreds to thousands of reactions at a time, each at the nanomole scale, have been reported.11,28,29
Efficiency gains can also be realized through the incorporation of iterative optimal experimental design tools into the platform control software,6,9,10,13,15,30 particularly when those tools are capable of leveraging preexisting reaction information.30 The proliferation of open-source optimization algorithms has made it facile to integrate these into preexisting systems.31–36
When the automation objective is not just knowledge generation but also compound synthesis, the ability to automatically perform multi-step synthesis sequences becomes particularly important. A range of approaches have been described for automating multi-step syntheses in ways that retain flexibility, including timed injections of additional reagents into intermediate reaction droplets,37 the use of selector valves to create independent reactor channels that allow for the decoupling of subsequent synthesis steps,12 and robotic reconfiguration of reactor modules.15,38
Despite the incredible progress that has been made, there remains a need for platforms that can reproduce the flexibility, accuracy, and iterative experimental design capabilities of the bench chemist while achieving elevated throughput and using as little material as possible.25,39 It is often claimed that reaction automation represents a path toward reduced variability and improved control over reaction outcomes,40 but it is important to emphasize that substantial engineering hurdles must be overcome to achieve such fine control, particularly when reactions are being studied at the microscale.
Additionally, it is often cheaper and simpler to design platforms that can be used to study only a subset of chemical reaction space.41 For example, many of the platforms that have been developed for conducting high-throughput reaction experimentation were designed with constraints that facilitate the high-throughput nature of the platform while limiting the kinds of chemistries or ranges of operating conditions that can be studied. For example, microfluidic devices constructed using polycarbonate or polydimethylsiloxane (e.g. ref. 42) are incompatible with many common organic solvents, and can't operate at high pressures. Other platforms achieve high throughput through the use of well plates, in which all of the reactions on a particular plate are confined to the same temperature and reaction time. Some well plate-based platforms limit reactions to high-boiling solvents and ambient temperature in order to preserve the fidelity of the reactions.28 For larger-scale reactions (on the order of a milliliter), well plate-style approaches with glass vials for broad chemical compatibility and a top-plate for sealing can be used to enable heating without solvent loss,43 but the confinement of all reactions in the plate to one set of continuous reaction variables remains, at the expense of sampling for reaction kinetic analysis. An alternative approach involves generating unique reaction droplets sequentially in a tube by varying the amount of reagent injected into each droplet,44 but the lack of total independence of the conditions of reactions running in parallel persists. To increase the “universality” of automated platforms, various types of automated modular platforms45 (some of which autonomously rearrange themselves to suit chemistries and process sequences of interest38) have been developed, but there remains a need to translate these concepts of universality to automation at the micro- and nanomole scale.
The platform that we developed is designed to address the aforementioned challenges, and is intended to be used primarily to facilitate efficient reaction development. The platform enables facile exploration of the influence of a variety of categorical and continuous reaction variables on reaction outcomes. For the platform to be useful, it is crucial that the reaction outcomes reported by the platform accurately reflect the reaction conditions specified by the user. The platform combines a bank of parallelized microfluidic reactors that can each individually be controlled across a broad temperature range and operated in either thermal or photochemical mode with a liquid handler and an on-line HPLC. The platform is governed by customized control software that synchronizes all of the hardware and schedules all operations to ensure efficient execution. To meaningfully build on existing capability, at the outset of our design efforts, we set the following goals for performance and operating characteristics of the automated platform:
(1) Excellent reproducibility: <5% standard deviation in reaction outcomes.
(2) Reaction temperatures from 0 to 200 °C (solvent-dependent).
(3) Operating pressure up to 20 atm.
(4) Online analysis: minimal delay between reaction completion and evaluation allows for real-time feedback and eliminates need for quenching and sample stability.
(5) Ability to perform both thermal and photochemical transformations.
(6) Maximally efficient operation through parallelization and scheduling.
(7) Integrated optimal experimental design algorithm for iterative experimentation and optimization.
With all of these capabilities incorporated, our platform fills an important niche. The throughput of our platform is only moderate compared to what can be achieved with nanowell plates, but the conditions in each reactor channel are totally independent, the range of operating conditions is broad, there is flexibility with regard to thermal versus photochemical operation, and the inclusion of fully-automated feedback and automated experimental design allows for fully-automated iterative experimentation. Further, prior work on a precursor to this platform showed that the reaction information obtained via these droplet reactors is directly scalable.46 Thus, the platform is well-suited for situations when it is appropriate to sacrifice maximum throughput in favor of acquiring high-quality data across a wide range of conditions.
Herein, we describe the work we performed to develop our platform and verify that it met all of our design criteria, along with the experiments we conducted to demonstrate the platform's utility.
To convert the original oscillatory droplet reactor platform to a high-throughput version, we swapped the single reactor channel for a reactor bank consisting of multiple independent parallel reactors. Parallelization allowed us to increase the throughput while retaining flexibility, because each individual reactor can operate at a set of conditions independent of those of its neighbors. We chose to make the parallel channels independent of one another to enable integration with experimental design algorithms, which operate most efficiently when they aren't constrained to propose batches of experiments that have conditions in common. To allow for parallelized droplet oscillation, we initially developed a compact oscillation mechanism, replicates of which could be easily produced and attached to each channel.49 The novel mechanism was designed to deliver rapid oscillation for the purposes of efficient mixing, but rapid mixing made solvent loss issues apparent (see Subsection 2.2), which motivated us to switch to stationary operation.
A schematic of the parallelized platform is shown in Fig. 1. In addition to the components described in ref. 37, the platform features the following additional hardware:
• A reactor bank consisting of a series of parallel reactor channels (we chose to use ten parallel channels by considering the length of bottleneck operations and comparing those to anticipated typical reaction times, see Subsection 3). Selector valves are positioned upstream and downstream of the reactor bank to distribute droplets among the various channels, analogous to the deployment of selector valves in the automated platform developed by Chatterjee et al.12
• A six-port, two-position valve for each reactor channel that allows each reaction droplet to be isolated from the rest of the system while the reaction is running.
• Two ten-position selector valves (VICI Valco C5H-3720EUHAY) upstream and downstream of the reactor bank that allow for distribution of droplets to their assigned reactors, and collection of droplets from the reactor bank for analysis.
• An internal injection valve (VICI Valco C84H-1574-.02EUHA) for sampling droplets for HPLC analysis, whose swappable nanoliter-scale rotors (20 nL, 50 nL, 100 nL) enable minuscule injection volumes that (1) eliminate the need to dilute concentrated reactions prior to analysis and (2) mitigate the effects of strong solvents on the analytical outcome.
One of the criteria we emphasized while designing the platform was reproducibility. Every step in the process that influences reaction outcomes can also impact reproducibility. Some of these factors require simple calibration and standardization, such as ensuring that each thermocouple is calibrated and positioned in the same location on the reactor plate; other factors are intrinsic features of equipment that was incorporated into the platform as-is, such as the consistency of the HPLC injection. To prevent carryover between droplets from impacting reproducibility, we incorporated multiple rinse droplets between each droplet transit operation.
The Gilson liquid handler that we used to prepare reaction droplets allows for extensive customization of factors that can have a significant influence on reproducibility. To ensure fidelity to user-specified concentrations, we optimized the liquid handler design and operation parameters to deliver deviations from target concentrations of <5% (see ESI Section S3.2†). We expect these experiments to be of interest to automated experimentation scientists generally, given the widespread use of liquid handlers in automated systems.
Reaction execution also requires special consideration. During our early development experiments with the novel oscillator, we observed significant evaporation effects as reactions proceeded: reaction droplets would shrink over time, and corresponding droplets of condensed liquid would grow outside the reactor. Unlike a reaction in a sealed, jacketed vessel, the droplet does not merely evaporate to saturate the headspace, and then cease evaporating once equilibrium is reached. Rather, several distinct phenomena contribute to a more dynamic behavior, in which volatile components in the droplet evaporate to saturate the headspace, the vapor which is saturated at the elevated reactor temperature exits the heated reactor (due to both convection from the induced oscillation and diffusion, see ESI Section S4†), the vapor then becomes supersaturated in the cooler tubing outside the reactor (which is not temperature-controlled, see Fig. 2), and condenses, thus forcing further evaporation of the droplet to maintain saturation of the heated headspace inside the reactor. To provide a more quantitative picture of evaporation effects, we measured the apparent charge error for a series of preliminary reactions that we performed at a range of temperatures and residence times using the oscillatory reactor (see ESI Section S4†). The apparent charge error values skew positive (corresponding to “overcharge”), with an average value of 70%, which is consistent with significant evaporation effects.
Based on our assessment of the driving forces for evaporation and our experimental observations of its significant impact on data quality and reproducibility, we determined that rapid oscillation of the reaction droplets is fundamentally at odds with another goal of the platform, which is to study reaction droplets at high temperatures (including temperatures that exceed the atmospheric boiling point of the solvent) without limits on the residence time. To remain faithful to our high-fidelity criterion and retain compatibility with high temperatures and essentially unlimited residence times, we eliminated the oscillatory component from the original platform design, which helps mitigate evaporation, but requires us to constrain our reaction space to those reactions which are homogeneous initially. We considered this preferable to using exclusively high-boiling solvents which would have further limited reaction scope.28 Accordingly, we redeveloped the reactor for “stationary” operation with acceptable restrictions on reaction scope; the resulting reaction operation style is similar to the “stopped-flow” approach described by both Chatterjee et al.12 and Avila et al.16
We designed a new single-channel reactor that uses tubing with a small (0.02′′) inner diameter to minimize the interfacial area between the gas and the liquid and therefore slow evaporative behavior. We also minimized the total reactor headspace to minimize solvent loss due to vapor-liquid equilibrium; we determined an appropriate total volume of tubing by accounting for thermal expansion and gas evolution from a model decarboxylation reaction (see ESI Section S1†).
The updated stationary reactor that accounts for these conditions with safety factors has a heated volume of 60 μL and connects to the six-port valve via eight total inches of temperature-uncontrolled 0.02′′ ID FEP tubing, which translates to a volume of about 40 μL. A SolidWorks sketch of the reactor is included in Fig. 3b. Evaporation studies verified that zero measurable evaporative losses from a droplet of pure THF are observed over 30 min at 125 °C (see ESI Section S4†). The droplets do expand upon heating, but we did not observe any translation of the droplets that could be due, for example, to thermocapillary effects, since there is little to no temperature gradient within the reactor itself (see ESI Section S4†).
Before constructing the full parallelized reactor, we performed preliminary demonstration experiments using a model homogeneous nucleophilic aromatic substitution (SNAr) reaction (Scheme 1) that motivated additional modifications to our droplet handling procedure (see ESI Section S3.2 for additional details†). After the additional modifications were incorporated, we performed a final set of validation experiments using the same model reaction. The final validation experiments indicated excellent reproducibility (see ESI Section S3.2†).
Scheme 1 Nucleophilic aromatic substitution reaction used to validate the single-channel version of the platform. |
A common photoreactor design challenge is temperature control. To keep the reactor cool while the LEDs are operational, a Peltier module was incorporated into the design: a Peltier device (Multicomp Pro part number MCPE-241-14-15) was sandwiched between the stationary reactor and a backboard using thermal paste, with the “cold side” of the Peltier touching the reactor. On the opposite end of the backboard, we mounted a heat sink (Alpha Novatech part number FH6030MU) and fan to remove the heat generated on the “hot side” of the Peltier and maintain its efficiency. A 3D sketch of the module is provided in Fig. 3c.
To supply power and control to the Peltier module, LED board, and heat sink fans, we designed and constructed a custom circuit board (see ESI Section S1.2†). The LabVIEW-based platform control software communicates with the circuit boards via a Python node.
To validate the platform's photochemical capabilities, we sought to:
(1) Define an absolute minimum separation distance between the reactor and the LED board that was large enough for the Peltier cooler to maintain the reactor temperature below 25 °C when the LEDs are operated at full power;
(2) Further define an appropriate separation distance (equal to or greater than the temperature-based distance) that delivered sufficiently uniform radiation to the reactor; and
(3) Affirm that the performance was consistent across three identical copies of the LED board, to validate readiness for parallelization.
For the purposes of these initial tests, three identical boards were sent for production at Bittele‡ using 455 nm LEDs (Cree LED part number XTEARY-00-0000-000000N02). We started by determining the minimum distance between the LED board and the reactor (see ESI Section S2†).
Next, we further optimized the distance between the reactor and LED board on the basis of irradiance uniformity by performing five different 9,10-diphenylanthracene (DPA) oxidation reactions (Scheme 2) at each of a series of separation distances, while allowing the natural variation in droplet placement in the reactor to provide a representative assessment of the performance at different locations within the reactor (Table 1). Since the absolute irradiance reaching the reaction mixture decreases as distance from the LED board increases, we increased the residence time of the droplets in parallel with the increasing separation distance in order to maintain comparable conversion values (approximately 50% in each case) across the various separation distances. Based on the results in Table 1, the reaction is fairly reproducible across the range of separation distances we tested. At the minimum separation distance of 2 cm, the standard deviation in final DPA concentration is slightly elevated compared to the rest of the values, suggesting greater nonuniformity than the cases where the reactor and LED board are further apart. On the basis of these results, we chose to fix the reactor-board separation distance at 4 cm.
Scheme 2 Aerobic oxidation of 9,10-diphenylanthracene to the corresponding peroxide in the presence of Ru(bpy)3Cl2 used to validate irradiance uniformity. |
Finally, we used a model photoredox reaction (Scheme 3) to verify that the three LED boards we ordered behave similarly. For each board, we performed one photoredox reaction at each of two different power levels. Then, for each power level, we computed the standard deviation in conversion values across the three boards to assess cross-board variation (Table 2). As a baseline for the variation assessment, we also performed three photoredox reactions at the higher of the two power levels (39%) using LED board #1. The standard deviation in conversion across these three experiments was 3.1%, which is similar to the degree of variation observed across the boards, confirming that the boards behave very similarly.
LED power [%] | Board #1 [% conversion] | Board #2 [% conversion] | Board #3 [% conversion] | Standard deviation [%] |
---|---|---|---|---|
39 | 55.6 | 58.5 | 52.4 | 3.1 |
12 | 27.1 | 30.9 | 25.8 | 2.7 |
The reason that “collisions” as defined broadly above must be avoided is twofold: first, under collision conditions, droplets won't necessarily merge, but they might; second, it would be difficult to design a tracking system that would be capable of confidently distinguishing between the droplets. Therefore, the surest way to avoid errors is to design the scheduling algorithm such that no more than one droplet is expected to occupy a given section of hardware at any one time. The scheduler must satisfy the following requirements:
• Once prepared, a droplet must be immediately transported to its assigned reactor to minimize the impact of any room-temperature reactivity on the outcome.
• For the same reason listed in the bullet above, once a droplet has been confined to its reactor for its specified residence time, it must be immediately transported to analysis.
• All droplet transit operations must be immediately followed by a rinse/vent operation to prevent carryover to subsequent droplets. Two types of rinse and vent operations are necessary: one that does not include a reactor in the flowpath, which follows droplet transit from the liquid handler injection valve to its reactor; and another that follows droplet transit from the reactor to the analytical sampling valve, and which therefore does rinse the reactor in question as well.
Initially, we considered ways to design the scheduling algorithm such that it produced a schedule a priori based on user-provided operation times for the liquid handler, both transit operations, both rinse/vent operations, the analytical method, and of course the reactions themselves, but there are too many possible cases to construct the scheduler in that manner. Instead, we wrote the scheduler so that it starts by assuming that droplet preparation can be back-to-back. Then, the scheduler scans through the reactions in order of execution, looking for cases where droplets would collide (as defined above). If collisions are detected, the algorithm begins iteratively shifting the preparation of the first colliding droplet back by one unit of time until all collisions are eliminated. The algorithm is written in MATLAB, and it is very fast: it can schedule 500 reactions in less than a tenth of a second. See ESI Section S3† for additional details.
Scheduling is one important focus of the rich field of operations research, which emphasizes operational efficiency. Two quantitative measures of the efficiency of a process are total operation time (which is, ideally, minimized) and utilization (ideally maximized). The parallelized droplet reactor is a version of a classic scheduling problem in which m unique jobs are to be scheduled on n identical parallel machines and the objective is to minimize the “makespan”, or total time necessary to complete all m jobs. Minimizing the makespan in this case is complex; an early approximation to a universal solution is known as the Longest Processing Time (LPT) rule,50 in which the m jobs are sorted from longest to shortest (in our case, from the longest target residence time to the shortest) and run in that order. There are alternatives to LPT that improve on its performance in edge cases (see, e.g. ref. 51), but LPT remains a popular heuristic for its simplicity and short run time. We implemented LPT in the scheduling algorithm in order to approximately minimize the total run time of a given set of reactions.
The scheduling algorithm does not directly account for reactor heating or cooling rates; instead, the control software pauses droplet production until the temperature of the droplet's assigned reactor has reached its target value. A more direct accounting of the heating and cooling rates by the scheduling algorithm would increase overall efficiency. Heating operations – which rely on cartridge heaters – are quite rapid compared to the time necessary to prepare a reaction droplet, but cooling can take several minutes, depending on the magnitude of the gradient and the initial temperature of the heat sink.
To showcase the benefits of collecting kinetic data on our platform, we sought to reproduce the results of a preexisting kinetics study52 examining a nucleophilic aromatic substitution reaction between 4-fluoronitrobenzene and 1-methylpiperazine in acetonitrile as solvent (Scheme 4). First, we verified the rate law determined by Ashworth et al. (eqn (1)) by performing the reaction at 70 °C in the presence of two different concentrations of the aryl halide (the limiting reagent) and tracking the consumption of the aryl halide. We took advantage of the ten parallel reactors to rapidly evaluate both conditions at residence times of 5, 10, 15, 25 and 40 minutes. Next, we measured the temperature dependence of the rate constant by studying the reaction at 60, 80, and 90 °C. Note that 90 °C exceeds the atmospheric boiling point of the solvent, which is accessible thanks to the platform's unique pressurized design.
r = k[ArF][R2NH]2 | (1) |
Scheme 4 Reaction used to demonstrate the accurate and rapid determination of kinetic parameters on the platform. |
We used the MATLAB functions nlinfit and ode45 to determine the rate constants at each temperature, and then fitted the data to the Eyring equation (eqn (2)) to determine the enthalpy and entropy of activation of the reaction for comparison to the literature.52
(2) |
To robustly determine kinetic parameters for the reaction, we performed a total of thirty reactions, which required just thirteen hours of platform time and a combined 600 mg of the two starting materials. We observed excellent agreement between the rate law and rate constants determined by Ashworth et al. and those that we measured (Fig. 5a and b and Table 3).
Fig. 5 Results of the kinetics investigation. (a) Temporal reaction profiles at 70 °C for two different initial concentrations of the limiting reagent, 1-fluoro-4-nitrobenzene: 0.8 M and 1.04 M. Scatter points represent experimental data; line plots represent the kinetic model. (b) Temporal reaction profiles at temperatures from 60 °C to 90 °C in ten-degree increments, with the initial concentration of the limiting reagent, 1-fluoro-4-nitrobenzene, fixed at 1.04 M. Scatter points represent experimental data; line plots represent the kinetic model. (c) Eyring plot for measured rate constants. Linear regression of the experimental data yielded eqn (3), which is plotted with a dashed line in the figure. |
Temperature | k [M−2 s−1] × 104 | Literature values52 |
---|---|---|
60 °C | 3.32 ± 0.07 | 2.87, 3.03 |
70 °C | 5.00 ± 0.14 | 5.10, 5.15, 5.08 |
80 °C | 7.90 ± 0.30 | 7.77, 7.75 |
90 °C | 10.6 ± 0.40 | — |
We constructed an Eyring plot (Fig. 5c) to determine the enthalpy and entropy of activation of the reaction. Linear regression of the experimental data yielded the following equation:
(3) |
The linear fit was used to determine the following enthalpy and entropy of activation: ΔH‡ = 36.5 ± 1.9 kJ mol−1 and ΔS‡ = −203 ± 5 J mol−1 K−1. The thermodynamic values we determined differ somewhat from the literature values of ΔH‡ = 44.5 kJ mol−1 and ΔS‡ = −180 J mol−1 K−1,52 due in part to the fact that we explored a wider temperature range than Ashworth et al. When the thermodynamic constants are recomputed without the 90 °C data, the values are closer to those previously reported in the literature: ΔH‡ = 39.5 kJ mol−1 and ΔS‡ = −194 J mol−1 K−1.
We decided to deploy the Dragonfly implementation of Bayesian optimization for several reasons. One of the advantages of the Dragonfly package is that it avoids performance artefacts that are specific to a single acquisition function by adaptively sampling from a set of popular acquisition functions (upper confidence bound, expected improvement, and Thompson sampling) in a manner that progressively prefers acquisition functions that perform better on the problem in question.53 Unlike a popular, reaction-specific option (EDBO35 and EDBO+36), Dragonfly explicitly optimizes the acquisition function to select the next point(s) (using DiRect, PDOO, or evolutionary algorithms, automatically selected by Dragonfly depending on the nature and dimensionality of the domain). Further, Dragonfly allows for multiobjective optimization and for a wide variety of variable types (continuous, discrete numeric, and discrete categorical).
We applied Dragonfly to a series of reaction optimization case studies, each consisting of both continuous and categorical optimization variables. We operated Dragonfly in ask-tell mode with primarily default settings (see ESI Section S5.0.1 for details†). We used one-hot encoding to distinguish between the various settings of each categorical variable. The optimization objective in each case study was product peak area determined by HPLC, normalized by the peak area of an internal standard. Algorithm termination was based on manual assessment of convergence.
We used Dragonfly to perform closed-loop optimization of the Buchwald–Hartwig amination between 9H-carbazol-2-yl trifluoromethanesulfonate and 3-aminopyridine (Scheme 5) over catalyst (tBuBrettPhos Pd G3 or tBuXPhos Pd G3), base (DBU or BTMG), temperature (50 to 100 °C) and residence time (5 to 60 minutes). We performed the optimization campaign twice, once in dimethylformamide (DMF) and once in dimethyl sulfoxide (DMSO), to examine the influence of solvent on reaction performance. In both solvents, an optimum was identified within a few dozen experiments (ESI Fig. S21a and b†); for the DMF campaign, 28 experiments were performed in 12.5 hours, consuming 132 mg of the aryl triflate, and for the DMSO campaign, 30 experiments were performed over 14 hours, requiring the consumption of 142 mg of aryl triflate.
When the reaction is performed in DMF, many conditions are found that deliver nearly full conversion: a quarter of the reactions that were performed during the campaign had outcomes that came within ten percent of the optimum. In contrast, when the reaction is performed in DMSO, high-performing conditions are more difficult for the algorithm to find, and the optimal objective function value that the algorithm identified is 30% lower than the optimum identified in DMF. The relatively poor performance of this reaction in DMSO compared to its performance in DMF highlights the importance of platforms that enable screening of solvents other than DMSO (Fig. 6).
Across both campaigns, the best-performing catalyst–base combination is tBuBrettPhos Pd G3 and DBU, and the best performance is observed at the maximum temperature in the range examined. Residence time was not found to have a large effect on reaction outcomes in either solvent, despite the fact that we examined a wide residence time range; high-conversion conditions could be identified even at the low end of the residence time range.
We separately optimized another Buchwald–Hartwig amination with 9H-carbazol-2-yl trifluoromethanesulfonate, this time using tetrahydrofuran (THF) as the solvent and (3-aminophenyl)(phenyl)methanone as the amine (Scheme 6). We optimized this reaction over catalyst (tBuBrettPhos Pd G3 or XantPhos Pd G3), base (DBU or BTMG), temperature (70 to 120 °C) and residence time (5 to 30 minutes). For this reaction, the XantPhos precatalyst yielded little to no product, whereas tBuBrettPhos resulted in full conversion at the high end of the temperature range, regardless of the residence time. This campaign was completed in 28 experiments, which required ten hours of execution time and the consumption of 132 mg of aryl triflate.
The successful use of THF in the second optimization campaign highlights the platform's compatibility with relatively volatile solvents. DMSO is commonly used as a solvent in high-throughput reaction screening platforms because its low vapor pressure and high boiling point help ensure the fidelity of reaction concentrations and because it easily dissolves a wide variety of compounds, but its high solvation tendency poses challenges when processes are later scaled and optimized for isolated yield and safety. Further, the performance of the first optimization campaign in DMSO highlights the significance of solvent effects and motivates the design and implementation of platforms like ours, which can screen reactions in more process-realistic solvents.
There are important challenges that remain to be addressed. The throughput of our platform is only moderate, and could be improved through the incorporation of tandem bottleneck instruments as well as shorter analytical methods and shorter liquid handler processing times. Fast, accurate, affordable, and easy-to-use solids-handling equipment could have eliminated the time-consuming manual effort required to prepare stock solutions for the liquid handler. Efficiency gains could also be realized through further optimization of the scheduling algorithm to account for reactor heating and cooling rates.
Additionally, fully-automated analysis is a persistent challenge. Ideally, it would be straightforward to determine the yield of each product of every reaction, but yield determination requires calibration, which in turn requires authentic product standards. Internal standards can be useful to allow for simple normalized response as output rather than yield (the approach we used here). The responses of the various reaction components must be distinguishable from one another in the chosen analytical approach. The recently-reported MOCCA software (and other previous efforts)57 is helping on this front in the U/HPLC space: MOCCA uses the full UV-Vis spectra recorded by diode array detectors at each timepoint, which enables deconvolution of overlapping peaks. Automated analysis could also be facilitated through the incorporation of automatic response factor estimation.58
To improve the efficiency of the reaction optimization protocol, robust approaches to defining batches of experiments that can be performed in parallel are needed. In general, performing batches of experiments from an iterative experimental design routine on a parallelized platform involves a tradeoff between rapid collection of additional data (which is gained through parallel operation) and the possibility that the experiments in the batch are redundant. Thompson sampling represents one approach to batched Bayesian optimization that has been found to perform well in a variety of reaction optimization scenarios.32,59,60
Footnotes |
† Electronic supplementary information (ESI) available: Complete description of all of the hardware components included in the platform, the preliminary investigations that were performed to optimize hardware performance, details about development of the software and settings for the optimization algorithm, and tabulated results of all of the experiments used to demonstrate the capabilities of the platform. See DOI: https://doi.org/10.1039/d3sc02082g |
‡ https://www.7pcb.com |
This journal is © The Royal Society of Chemistry 2023 |