Sunday, 8 May 2016

A real world perspective on molecular design

I'll be taking a look at a Real-World Perspective on Molecular Design which has already been reviewed by Ash. I don't agree that this study can accurately be described as 'prospective' although, in fairness, it is actually very difficult to publish molecular design work in a genuinely prospective manner. Another point to keep in mind is that molecular modelers (like everybody else in drug discovery) are under pressure to demonstrate that they are making vital contributions. Let's take a look at what the authors have to say:

"The term “molecular design” is intimately linked to the widely accepted concept of the design cycle, which implies that drug discovery is a process of directed evolution (Figure 1). The cycle may be subdivided into the two experimental sections of synthesis and testing, and one conceptual phase. This conceptual phase begins with data analysis and ends with decisions on the next round of compounds to be synthesized. What happens between analysis and decision making is rather ill-defined. We will call this the design phase. In any actual project, the design phase is a multifaceted process, combining information on status and goals of the project, prior knowledge, personal experience, elements of creativity and critical filtering, and practical planning. The task of molecular design, as we understand it, is to turn this complex process into an explicit, rational and traceable one, to the extent possible. The two key criteria of utility for any molecular design approach are that they should lead to experimentally testable predictions and that whether or not these predictions turn out to be correct in the end, the experimental result adds to the understanding of the optimization space available, thus improving chances of correct prediction in an iterative manner. The primary deliverable of molecular design is an idea [4] and success is a meaningful contribution to improved compounds that interrogate a biological system."

This is a certainly a useful study although I will make some criticisms in the hope that doing so stimulates discussion. I found the quoted section to lack coherence and would argue that  the design cycle is actually more of a logistic construct than a conceptual one. That said, I have to admit that it's not easy to clearly articulate what is meant by the term 'molecular design'. One definition of molecular design is control of behavior of compounds and materials by manipulation of molecular properties. Using the term 'behavior' captures the idea that we design compounds to 'do' rather than merely to 'be'. I also find it useful to draw a distinction between hypothesis-driven molecular design (ask good questions) and prediction-driven molecular design (synthesize what the models, metrics or tea leaves tell you to). Asking good questions is not as easy as it sounds because it it is not generally possibly to perform controlled experiments in the context of molecular design as discussed in another post from Ash. Hypothesis-driven molecular design can also be thought of as a framework in which to efficiently obtain the information required to make decisions and, in this sense, there are analogies with statistical molecular designI believe that the molecular design that the authors describe in the quoted section is of the hypothesis-driven variety but hand-wringing about how "ill-defined" it is doesn't really help move things forward. The principal challenges for hypothesis-driven molecular design are to make it more objective, systematic and efficient. I'll refer you to a trio of blog posts ( 1 | 2 | 3) in which some of this is discussed in more detail.

I'll not say anything specific about the case studies presented in this study except to note that sharing specific examples of application of  molecular design as case studies does help to move the field forward even when the studies are incomplete. The examples do illustrate how the computational tools and structural databases can be used to provide a richer understanding of molecular properties such as conformational preferences and interaction potential. The CSD (Cambridge Structural Database) is a particularly powerful tool and, even in my Zeneca days, I used to push hard to get medicinal chemists using it. Something that we in the medicinal chemistry community might think about is how incomplete studies can be published so that specific learning points can be shared widely in a timely manner.  

But now I'd like to move on to the conclusions, starting with 1 (value of quantitative statements), The authors note:

"Frequently, a single new idea or a pointer in a new direction is sufficient guidance for a project team. Most project impact comes from qualitative work, from sharing an insight or a hypothesis rather than a calculated number or a priority order. The importance of this observation cannot be overrated in a field that has invested enormously in quantitative prediction methods. We believe that quantitative prediction alone is a misleading mission statement for molecular design. Computational tools, by their very nature, do of course produce numerical results, but these should never be used as such. Instead, any ranked list should be seen as raw input for further assessment within the context of the project. This principle can be applied very broadly and beyond the question of binding affinity prediction, for example, when choosing classification rather than regression models in property prediction."
  
This may be uncomfortable reading for QSAR advocates, metric touts and those who would have you believe that they are going to disrupt drug discovery by putting cheminformatics apps on your phone. It also is close to my view of the role of computational chemistry in molecular design (the observant reader will have noticed that I didn't equate the two activities) although, in the interests of balance, I'll refer you to a review article on predictive modelling. We also need to acknowledge that predictive capability will continue to improve (although pure prediction-driven pharmaceutical design is likely to be at least a couple of decades away) and readers might find this blog post to be relevant. 

Let's take a look at conclusion 5 (Staying close to experiment) and the authors note:

"One way of keeping things as simple as possible is to preferentially utilize experimental data that may support a project, wherever this is meaningful. This may be done in many different ways: by referring to measured parameters instead of calculated ones or by utilizing existing chemical building blocks instead of designing new ones or by making full use of known ligands and SAR or related protein structures. Rational drug design has a lot to do with clever recycling."

This makes a lot of sense although I don't recommend use of the tautological term 'rational drug design' (has anybody ever done irrational drug design?). What they're effectively saying here is that it is easier to predict the effect of structural changes on properties of compounds than it is to predict those properties directly from molecular structure. The implications of this for cheminformaticians (and others seeking to predict behaviour of compounds) is that they need to look at activity and chemical properties in terms of relationships between the molecular structures of compounds. I've explored this theme, both in an article and a blog post, although I should point out that there is a very long history of associating changes in the values of properties of compounds with modifications to molecular structures.

However, there is another side to "staying close to experiment" and that is recognizing what is and what isn't an experimental observable. The authors are clearly aware of this point when they state: 

"MD trajectories cannot be validated experimentally, so extra effort is required to link such simulation results back to truly testable hypotheses, for example, in the qualitative prediction of mechanisms or protein movements that may be exploited for the design of binders."

When interpreting structures of protein-ligand complexes, it is important to remember that the contribution of an intermolecular contact to affinity is not, in general, an experimental observable. As such, it would have been helpful if the authors had been a bit more explicit about exactly which experimental observable(s) form the basis of the "Scorpion network analysis of favorable interactions". The authors make a couple of references to ligand efficiency and I do need to point out that scaling free energy of binding has no thermodynamic basis because, in general, our perception of efficiency changes with the concentration used to define the standard state. On a lighter note there is a connection between ligand efficiency and homeopathy that anybody writing about molecular design might care to ponder and that's where I'll leave things.

Friday, 1 April 2016

LELP metric validated

<< previous || next >>

So I was wrong all along about LELP.  

I now have to concede that the LELP metric actually represents a seminal contribution to the science of drug design. Readers of this blog will recall our uncouth criticism of LELP which, to my eternal shame, I must now admit is actually the elusive, universal metric that achieves simultaneous normalization (and renormalization) of generalized interaction potential with respect to the size-scaled octanol/water partition function.

What changed my view so radically? Previously we observed that LELP treats the ADME risk associated with logP of 1 and 75 heavy atoms as equivalent to that associated with logP of 3 and 25 heavy atoms. Well it turns out that I was using the rest mass of the hydrogen atom to make this comparison which unequivocally invalidates the criticism of what turns out to be the most fundamental (and beautiful) of all the ligand efficiency metrics.

It is not intuitively obvious why relativistic correction is necessary for the correct normalization of affinity with respect to both molecular size and lipophilicity. However, I was fortunate to receive to receive a rare copy of the seminal article in the Carpathian Journal of Thermodynamics by T. T. Macoute, O. B. Ya and B. Samedi. The math is quite formidable and is based on convergence characteristics of the non-linear response to salt concentration of the Soucouyant Tensor, following application of the Douen p-Transform. B. Samedi is actually better known for his even more seminal study (with A. Bouchard) of the implications for cognitive function of the slow off-rate of tetrodotoxin in its dissociation from Duvalier's Tyrosine Kinase (DTK).

So there you have it. Ignore all false metrics and use LELP with confidence in your Sacred Quest for the Grail.  

Thursday, 10 March 2016

Ligand efficiency beyond the rule of 5


One recurring theme in this blog is that the link between physicochemical properties and undesirable behavior of compounds in vivo may not be as strong as property-based design 'experts' would have us believe. To be credible, guidelines for drug discovery need to reflect trends observed in relevant, measured data and the strengths of these trends tells you how much weight you should give to the guidelines. Drug discovery guidelines are often specified in terms of metrics, such as Ligand Efficiency (LE) or property forecast index (PFI), and it is important to be aware that every metric encodes assumptions (although these are rarely articulated).

The most famous set of guidelines for drug discovery is known as the rule of 5 (Ro5) which is essentially a statement of physicochemical property distributions for compounds that had progressed at least as far as as Phase II at some point before the Ro5 article was published in 1997. It is important to remember (some 'experts' have short memories) that Ro5 was originally presented as a set of guidelines for oral absorption. Personally, I have never regarded Ro5 as particularly helpful in practical lead optimization since it provides no guidance as to how suboptimal ADMET characteristics of compliant compounds can be improved. Furthermore, Ro5 is not particularly enlightening with respect to the consequences of straying out the allowed region and into 'die Verbotenezone'.

Nobody reading this blog needs to be reminded that drug discovery is an activity that has been under the cosh for some time and a number of publications ( 1 | 2 | 3 | 4 ) examine potential opportunities outside the chemical space 'enclosed' by Ro5. Given that drug-likeness is not the secure concept that those who claim to be leading our thoughts would have us believe, I do think that we really need to be a bit more open minded in our views as to the regions of chemical space in which we are prepared to work. That said, you cannot afford to perform shaky analysis when proposing that people might consider doing things differently because that will only hand a heavy cudgel to the roundheads for them to beat you with.

The article that I'll be discussing has already been Pipelined and this post has a much narrower focus than Derek's post. The featured study defines three regions of chemical space:  Ro5 (rule of 5), eRo5 (extended rule of 5) and bRo5 (beyond rule of 5). The authors note that "eRo5 space may be thought of as a buffer zone between Ro5 and bRo5 space".  I would challenge this point because there is a region (MW less than 500 Da and ClogP between 5 and 7.5) between Ro5 and bRo5 spaces that is not covered by the eRo5 specifications. As such, it is not meaningful to compare properties of eRo5 compounds with properties of Ro5 or bRo5 compounds. The authors of featured article really do need to fix this problem if they're planning to carve out niche in this area of study because failing to do so will make it easier for conservative drug-likeness 'experts' to challenge their findings. Problems like this are particularly insidious because the activation barriers for fixing them just keep getting higher the longer you ignore them.   

But enough of Bermuda Triangles in the space between Ro5 and bRo5 because the focus of this post is ligand efficiency and specifically its relevance (or otherwise) to bRo5 cmpounds. I'll write a formula for generalized LE is a way that makes it clear that DG° is a function of temperature, pressure and the standard concentration: 


LEgen = -DG°(T,p,C°)/HA

When LE is calculated it is usually assumed that C° is 1 M although there is nothing in the original definition of LE that says this has to be so and few, if any, users of the metric are even aware that they are making the assumption. When analyzing data it is important to be aware of all assumptions that you're making and the effects that making these assumptions may have on the inferences drawn from the analysis.

Sometimes LE is used to specify design guidelines.  For example we might assert that acceptable fragment hits must have LE above a particular cutoff. It's important to remember that setting a cutoff for LE is equivalent to imposing an affinity cutoff that depends on molecular size. I don't see any problem with allowing the affinity cutoff to increase with molecular size (or indeed lipophilicity) although the response of the cutoff to molecular size should reflect analysis of measured data (rather than noisy sermons of self-appointed thought-leaders). When you set a cutoff for LE, you're assuming (whether or not you are aware of it) that the affinity cutoff is a line that intersects the affinity axis at a point corresponding to Kof 1 M. Before heading back to bRo5, I'd like you to consider a question. If you're not comfortable setting an affinity cutoff as a function of molecular size would you be comfortable setting a cutoff for LE?

So let's take a look at what the featured article has to say about affinity: 


"Affinity data were consistent with those previously reported [44] for a large dataset of drugs and drugs in Ro5, eRo5 and bRo5 space had similar means and distributions of affinities (Figure 6a)"

So the article is saying that, on average, bRo5 compounds don't need to be of higher affinity than Ro5 compounds and that's actually useful information. One might hypothesize that unbound concentrations of bRo5 compounds tend to be lower than for Ro5 compounds because the former are less drug-like and precisely the abominations that MAMO (Mothers Against Molecular Obesity) have been trying to warn honest, god-fearing folk about for years. If you look at Figure 6a in the featured article, you'll see that the mean affinity does not differ significantly between the three categories of compound. Regular readers of this blog will be well aware that that categorizing continuous data in this manner tends to exaggerate trends in data. Given that the authors are saying that there isn't a trend, correlation inflation is not an issue here. 

Now look at Figure 6b. The authors note:


"As the drugs in eRo5 and bRo5 space are significantly bigger than Ro5 drugs, i.e., they have higher molecular weights and more heavy atoms, their LE is significantly lower"

If you're thinking about using these results in your own work, you really need to be asking whether or not the results provide any real insight (i.e. something beyond the the trivial result that 1/HA gets smaller when HA gets larger? This would also be a good time to think very carefully about all the assumptions you're going to make in your analysis. The featured article states:

"Ligand efficiency metrics have found widespread use;[45however, they also have some limitations associated with their application, particularly outside traditional Ro5 drug space. [46We nonetheless believe it is useful to characterize the ligand efficiency (LE) and lipophilic ligand efficiency (LLE) distributions observed in eRo5 and bRo5 space to provide guides for those who wish to use them in drug development"

Given that I have asserted that LE is not even wrong and have equated it with homeopathy, I'm not sure that I agree with sweeping LE's probems under the carpet by making a vague reference to "some limitations".  Let's not worry too much about trivial details because declaring a ligand efficiency metric to be useful is a recognized validation tool (even for LELP which appears have jumped straight from the pages of a Mary Shelley novel).  There is a rough analogy with New Math where "the important thing is to understand what you're doing rather to get the right answer" although that analogy shouldn't be taken too far because it's far from clear whether or not LE advocates actually understand what they are doing. As an aside, New Math is what inspired "the rule of 3 is just like the rule of 5 if you're missing two fingers" that I have occasionally used when delivering harangues on fragment screening library design.

So let's see what happens when one tries to set an LE threshold for for bRo5 compounds. The featured article states:

"Instead, the size and flexibility of the ligand and the shape of the target binding site should be taken into account, allowing progression of compounds that may give candidate drugs with ligand efficiencies of ≥0.12 kcal/(mol·HAC), a guideline that captures 90% of current oral drugs and clinical candidates in bRo5 space"

So let's see how this recommended LE threshold of 0.12 kcal/(mol.HA) translates to affinity thresholds for compounds with molecular weights of 700 Da and 3000 Da. I'll assume a temperature of 298 K and C° of 1 M when calculating DG°and will use 14 Da/HA to convert  molecular weight to heavy atoms.  I'll conclude the post by asking you to consider the following two questions? 


  • The recommended LE threshold transforms to pKD threshold of 4.4 at 700 Da. When considering progression of compounds that may give candidate drugs, would you consider a recommendation that KD should be less than 40 mM to be useful?



  • The recommended LE threshold transforms to a pKD threshold of 19 at 3000 Da. How easy do you think it would be to measure a pKD value of 19? When considering progression of compounds that may give candidate drugs, would you consider a recommendation that  pKD be greater than 19 to be useful?

      

Monday, 7 March 2016

On Sci-Hub

Many readers will have heard of Sci-Hub which makes almost 50 million copyrighted journal articles freely available. Derek has blogged about Sci-Hub and has also suggested that it might not matter as much as some think that it does. Readers might also want to take a look at some other posts ( 1 | 2 | 3 ) on the topic. I'll focus more on some of the fallout that might result from Sci-Hub's activities but won't be expressing an opinion as to who is right and who is wrong. Briefly, one side says that knowledge should be free, the other side says that laws have been broken. I'll leave it to readers to decide for themselves which side they wish to take because nothing I write is likely to change people's views on this subject. 

Sci-Hub and its creatrix are based in Russia and, given the current frosty relations between Russia and the countries which host the aggrieved journal publishers, it is safe to assume that Sci-Hub will be able to thumb its nose at the those publishers for the foreseeable future.  Sci-Hub relies relies on helpers to provide it with access to to the copyrighted material and these helpers presumably do this by making their institutional subscription credentials available to Sci-Hub. It's worth noting that one usually accesses copyrighted material through a connection that is recognized by the publisher and only a very small number of people at an institution actually know the access keys/passwords. One important question is whether or not publishers can trace the PDFs supplied by Sci-Hub. I certainly recall seeing PDFs from certain sources being marked with the name of the institution and date of download so I don't think that one can safely assume that no PDF is traceable. If a publisher can link a PDF supplied by Sci-Hub to a particular institution then presumably the publisher could sue the institution because providing third parties with access is specifically verboten by most (all?) subscription contracts. An institution facing a legal challenge from a publisher would be under some pressure to identify the leaks and publishers would be keen for some scalps pour encourager les autres.

While it would be understatement to say that the publishers are pissed off that Sci-Hub has managed to 'liberate' almost 50 million copyrighted journal articles, it is not clear how much lasting damage has been done.  The fee for downloading an article to which one does not have subscription access is typically in the range $20 to $50 but my guess is that only a tiny proportion of publishers' revenues comes from these downloads. I actually think the publishers set the download fees to provide institutions with the incentive to purchase subscriptions rather than to generate revenue from pay-per-view. If this is the case, Sci-Hub will only do real damage to the publishers if, by continuing to operate, it causes institutions to stop subscribing or helps them to negotiate cheaper subscriptions.

There is not a lot that the publishers can do about the material that Sci-Hub already has in its possession but there are a number of tactics that they might employ in order to prevent further 'liberation' of copyrighted material. I don't know if it is possible to engineer a finite lifetime into PDF files but they can be protected with passwords and publishers may try to only allow a small number of individuals at each institution to direct access the copyrighted material as PDF files. Alternatively the publishers might require that individual users create accounts and change passwords regularly in order to make it more difficult (and dangerous) for Sci-Hub's helpers to share their access. Countermeasures put in place by publishers to protect content are likely to add complexity to the process of accessing that content. This in turn would make it more difficult to mine content and the existence (and scale) of Sci-Hub could even be invoked as a counter to arguments that the right to read is the right to mine.

Given that almost 50 million articles are freely available on Sci-Hub, one might consider potential implications for Open Access (OA). There is a lot of heated debate about OA although the issues are perhaps not as clear cut as OA advocates would have you believe and this theme was explored in a post from this blog last year.  Although there is currently a lot of pressure to reduce the costs of subscriptions, it is difficult to predict how far Sci-Hub will push subscription-based journal publishers towards a purely OA business model. For example, we may see scientific publication moving towards a 'third way' in the form of pre-publication servers with post-publication peer review. I wouldn't be surprised to learn that 'direct to internet' has usurped both subscription-based and OA scholarly publishing models twenty years from now. That, however, is going off on a tangent and, to get things back on track, I'd like you to think of Sci-Hub from the perspective of an author who has paid a subscription-based journal $2000 to make an article OA. Would it be reasonable for this author to ask for a refund?       

   

Monday, 29 February 2016

The boys who cried wolf

So it's back to blogging and it's taken a bit longer to get into it this year since I had to finish a few things before leaving Brazil. This is a long post so make sure to have some strong coffee to hand.

This post features an article, 'Molecular Property Design: Does Everyone Get It?' by two unwitting 'collaborators' in our correlation inflation Perspective. There are, however, a number of things that the authors of this piece just don't 'get' which makes their choice of title particularly unfortunate.  The first thing that they don't 'get' is that doing questionable data analysis in the past means that people in the present are less likely to heed your warnings about the decline in quality of compounds in today's pipelines. As has been pointed out more than once by this blog, rules/guidelines in drug discovery are typically based on trends observed in measured data and the strength of the trend tells you how rigidly you should adhere to the rule/guideline. Correlation inflation (see also voodoo correlations) is a serious problem in drug discovery because it causes drug discovery scientists to to give more weight to rules/guidelines (and 'expert' opinion) than is justified by the data. In drug discovery, we need to make a distinction between what we believe and what we know. If we can't (or won't) make this distinction then those who fund our activities may conclude that the difficulties that we face are actually of our own making and that's something else that the authors of the featured article just don't seem to 'get'. "Views obtained from senior medicinal chemistry leaders..." does come across as arm-waving and I'm surprised that the editor and reviewers (if there were any) let them get away with it.  

If you're familiar with the correlation inflation problem, you'll know that one of the authors of the featured article did some averaging of groups of data points prior to analysis which was presented in support of an assertion that, "Lipophilicity plays a dominant role in promoting binding to unwanted drug targets". This may indeed be the case but it is not correct to suggest that the analysis supports this opinion because the reported correlations are between promiscuity and median lipophilicity rather than lipophilicity itself. The author concedes that the analysis has been criticized but does not make any attempt to rebut the criticism. Readers can draw their own conclusions from the lack of rebuttal.

The other author of the featured article also 'contributed' to our correlation inflation study although it would be stretching it to term that contribution as 'data analysis'. The approach used there was to first bin the data and then to plot bar charts which were compared visually. You might wonder how a bar chart of binned data can be used to quantify the strength of a  trend and, if attempting to do this, keep your arms loose because you'll be waving them a lot. Here are a couple of examples of how the approach is applied:    

The clearer stepped differentiation within the bands is apparent when log DpH7.4 rather than log P is used, which reflects the considerable contribution of ionization to solubility. 

This graded bar graph (Figure 9) can be compared with that shown in Figure 6b to show an increase in resolution when considering binned SFI versus binned c log DpH7.4 alone. 

This second approach to data 'analysis' is actually more relevant than the first to this blog post because it is used as 'support' ('a crutch' might be a more appropriate term) for SFI (Solubility Forecast Index), which is the old name for PFI (Property Forecast Index) which the featured article touts as a metric. If you're thinking that it's rather strange to 'convert' one form of continuous data (e.g. measured logD) into another form of continuous data (values of metrics) by first making it categorical and turning it into pictures, you might not be alone. What 'senior medicinal chemistry leaders' would make of such data 'analysis' is open to speculation.  

But enough of voodoo correlations and 'pictorial' data analysis because I should make some general comments on property-based design. Here's a figure that provides an admittedly abstract view of property-based design.
 
One challenge for drug-likeness advocates analyzing large, structurally heterogenous data sets is to make the results of analysis relevant to the medicinal chemists working on one or two series in a specific lead optimization project. Affinity (for association with both therapeutic target and antitargets) and free concentration at site of action are the key determinants of drug action. In general, the response of activity to lipophilicity depends on chemotype and, in the case of affinity, also on the relevant protein target (or antitarget). If you're going to tell medicinal chemists how to do their jobs then you can't really afford to have any data-analytic skeletons rattling around in the closet and that's something else that the authors of the featured article just don't 'get'.

The featured article asserts:

The principle of minimal hydrophobicity, proposed by Hansch and colleagues in 1987 states that “without convincing evidence to the contrary, drugs should be made as hydrophilic as possible without loss of efficacy.” This hypothesis is surviving the test of time and has been quantified as lipophilic ligand efficiency (LLE or LipE).

A couple of points need to be made here. Firstly, when Hansch et al refer to 'hydrophobicity', they mean octanol/water logP (as opposed to logD). Secondly, the observation that excessive lipophilicity is a bad thing doesn't actually justify using LLE/LipE in lead optimization. The principle proposed by Hansch et al suggests that a metric of the following functional form may be useful for normalization of activity with respect to liophilicity:


pIC50  -  (l  ´ logP)

However, the principle does not tell us what value of l is most appropriate (or indeed whether a single value of l is appropriate for all situations).  The 'sound and fury' article reviewed in an earlier post makes a similar error with ligand efficiency. 

So it's now time to take a look at PFI and the featured article asserts:


The likelihood of meeting multiple criteria, a typical requirement for a candidate drug, increases substantially with  ‘low fat, low flat’ molecules where PFI is <7, versus >7. In considering a portfolio of drug candidates, the probabilistic argument hypothesizes that successful outcomes will increase as the portfolio’s balance of biological and physicochemical properties becomes more similar to that of marketed drugs.

The first thing that a potential user of PFI should be asking him/herself is where this magic value of 7 comes from since the featured article does imply that the likelihood of good things will increase substantially when PFI is reduced from 7.1 to 6.9. Potential users also need to ask whether this step jump in likelihood is backed by statistical analysis of experimental data or by 'clearer stepped variation' in pictures created using an arbitrary binning scheme. It's also worth remembering that thresholds used to apply guidelines often reflect the binning schemes used to convert continuous data to categorical data and the correlation inflation Perspective discusses the 4/400 rule in this context. Something that molecular property design 'experts' really do need to 'get' is that simple yes/no guidelines are of limited use in practical lead optimization even when these are backed by competent analysis of relevant experimental data. Molecular property 'experts' also need to 'get' that measured lipophilicity is not actually a molecular property. 

PFI is defined as the sum of chromatographic logD (at pH 7.4) and the number of aromatic rings: 


PFI = Chrom logDpH7.4 +  Ar rings

Now suppose that you're a medicinal chemist in a department where the head of medicinal chemistry has decreed that that 80% of compounds synthesized by departmental personnel must have PFI less than 7.  When senior medicinal chemistry leaders set targets like these, the primary objective (i.e topic of your annual review) is to meet them. Delivering clinical candidates is secondary objective since these will surely materialize in the pipeline as if by magic provided that the compound quality targets are met. 

There is a difference between logD (whether measured  by shake-flask or chromatographically) and logP and one which it is important for compound quality advocates to 'get'. When we measure lipophilicity, we determine logD rather than logP and so it is not generally valid to invoke Hansch's principle of minimal hydrophobicity (which is based on logP) when using logD. If the compound in question is not significantly ionized under experimental conditions (pH) then logP and logD will be identical. However, this is not the case when ionization is significant as is usually the case for amines and carboxylic acids at a physiological pH like 7.4. If ionization is significant then logD will typically be lower than logP and we sometimes assume that only the neutral form of the compound partitions into the organic phase for the purposes of prediction or interpretation of log D values. If this is indeed the case we can write logD as a function of logP and the fraction of compound existing in neutral form(s):

  log D(pH) = log P + log Fneut(pH)
Ionized forms can sometimes partition into the organic phase although measuring the extent to which this happens is not easy and the effective partition coefficient for a charged entity depends on whatever counter ion is present (and its concentration). 

So let's get back to the problem of reducing logD so our medicinal chemist can achieve those targets and get an A+ rating in the annual review.  Two easy ways to lower logD are to add ionizable groups (if compound is neutral) and to increase extent of ionization (if compound already has ionizable groups). Increasing the extent of ionization will generally be expected to increase aqueous solubility but I hope readers can see why we wouldn't expect this to help when a compound binds in an ionized form to an antitarget such as hERG (see here for a more complete discussion of this point).  Now I'd like you to take a close look at Figure 2(a) in the featured article. You'll notice that the profiles for the last two entries (hERG and promiscuity) have actually been generated using intrinsic PFI (iPFI) rather than PFI itself and you may be wondering what iPFI is and why it was used instead of PFI. In answer to the first question, iPFI is calculated using logP rather than logD:


 iPFI = logP +  Ar rings

This definition of iPFI is not quite complete because the authors of the featured article don't actually say what they mean by logP.  Is it actually obtained directly from experimental measurements (e..g. logD/pH profile) or is it calculated (in which case it should be stated which method was used for the calculation).

Some medicinal chemists reading this will be asking what iPFI was even doing in the article in the first place and my response would be, as I say frequently in Brazil, 'boa pergunta'.  My guess is that using PFI rather than iPFI for the hERG row of Figure 2(a) would have the effect of shifting the cells in this row one or two cells to the left (based on the assumption that logP will be 1 to 2 units greater than logD at pH 7.4).  Such a shift would make compounds with PFI less than 7 look 'dirtier' than the PFI advocates would like you to think.

There is another term in PFI and that's the number of aromatic rings (# Ar rings) which is meant to measure how 'flat' a molecular structure is.  That it might do but then again it might not because two 'flat' aromatic rings will look a lot less flat when linked by a sulfonyl group and their rigidity could prove to be a liability when trying to pack them into a crystal lattice. However, number of aromatic rings will also quantify molecular size (especially in typical Pharma compound collections) and this is something my friends at Practical Fragments have also noted. Molecular size had been recognized as a pharmaceutical risk factor for at least a decade before people started to tout PFI (or SFI) as a compound quality metric and we can legitimately ask whether or not using a more conventional measure of molecular size (e.g. molecular weight, number of non-hydrogen atoms or molecular volume) would have resulted in a more predictive (or useful) metric.

So let's assume for a moment that you're a medicinal chemist in a place where the 'senior medicinal chemistry leaders' actually believe that optimizing PFI is useful. In case you don't know, jobs for medicinal chemists don't exactly grow on trees these days and so it makes a lot of sense to adopt an appropriately genuflectory attitude to the prevailing 'wisdom' of your 'leaders'. The problem is that your 1 nM enzyme inhibitor with the encouraging pharmacokinetic profile has a PFI of 8 and your lily-livered manager is taking some flak from the Compound Respository Advisory Panel for having permitted you to make it in the first place. Fear not because, you have two benzene rings at the periphery of the molecular structure which will make the synthesis relatively easy. Basically you need to think of a metric like PFI as a Gordian knot that needs to be cut efficiently and you can do this either by eliminating rings or by eliminating aromaticity. Substitution of benzoquinone (either isomer) or cyclopentadiene for the offending benzene rings will have the desired effect.

It's been a long post and I really do need to start wrapping things up. One common reaction when you criticize of drug discovery metrics is the straw man defense in which your criticism is interpreted as an assertion that one doesn't need to worry about physicochemical properties.  In other words, this is precisely the sort of deviant behavior that MAMO (Mothers Against Molecular Obesity) have been trying to warn about. To the straw men, I will say that we described lipophilicity and molecular size as pharmaceutical risk factors in our critique of ligand efficiency metrics. In that critique, we also explain what it means to normalize activity to with respect to risk factor and that's something that not even the NRDD ligand efficiency metric review does. There's a bit more to defining a compound quality metric than dreaming up arbitrary functions of molecular size and lipophilicity and that's something else that the authors of the featured article just don't seem to 'get'. When you use PFI you're assuming that a one unit decrease in chromatographic logD is equivalent to eliminating an aromatic ring (or the aromaticity of a ring) from the molecular structure. 

The essence of my criticism of metrics is that the assumptions encoded by the metrics are rarely (if ever) justified by analysis of relevant measured data. A plot of pIC50 against the relevant property for your project compounds is a good starting point for property-based design and it allows you to use the actual trend observed in your data for normalization of activity values (see the conclusion to our ligand efficiency metric critique for a more detailed discussion of this). If you want to base your decisions on 'clearer stepped differentiation' in pictures or on the blessing of 'senior medicinal chemistry leaders', as a consenting adult, you are free to do so.

Wednesday, 6 January 2016

Looking back at 2015


I'll start the year by taking a look back at some of the 2015 blog posts. The dynamic range of the bullshitometer was severely tested last year and there was an element of 'pour encourager les autres' to more than one of the posts. I thought that it'd be fun to share some travel pics and the first is of the Danube in Belgrade (I'd dropped by to catch up with friends and deliver a harangue at the university).  The early evening light was quite perfect although I hope that I won't spoil your experience of the photo by telling you that there was a pig carcass floating about 100 m from it was taken.

 

I changed the title of the blog this year. I've not been involved with FBDD for some years now and molecular design was always my main interest. One of the ideas that I try to communicate is that there's more to design than just making predictions. After Belgrade, I dropped in at Fidelta in Zagreb where I delivered another harangue before heading south to Sarajevo.  I'm a keen student of history so it was inevitable that this would be the first photo I'd take in Sarajevo.


It seems so bizarre today. There had already been one assassination attempt for the day when the driver of the car took the fateful wrong turn that gave Gavrilo Princip the opportunity to fire two shots at the royal couple. Back in Vienna, Sophie was not always allowed out in public with Franz Ferdinand so the trip to Sarajevo may have been a special treat for her. What if SatNav had already been invented but, then again, what if Queen Victoria's eldest child had succeeded her to the throne?

Part of the problem was that, as a lowly Czech countess, Sophie was not considered an appropriate match for the Habsburg heir by Franz Josef (the reigning emperor and a puritanical old killjoy) and there were rules (although metrics and Lean Six Sigma 'belts' had, thankfully, not yet been invented). One of the rules was that the children of Sophie and Franz Ferdinand were barred from succession. It is somewhat ironic that poor Franz Ferdinand was never even supposed to be crown prince in the first place and only got the job because his cousin Rudolf had abruptly removed himself from the Habsburg line of succession a quarter of a century previously. 

All this talk of puritanical rules serves as a reminder that, before moving on, I need to point you towards a friend's blog post on roundheads (who were bigger killjoys than Franz Josef or even Lean Six Sigma 'belts') and cavaliers in drug discovery.  I really like the term 'roundhead' and I think you do have to agree that it's a lot politer than 'compound quality jackboot'. Terms like 'roundhead' and 'jackboot' are invariably associated with pain and that brings me to the next topic which is PAINS. My interest in this topic was piqued by a PAINS-shaming post at Practical Fragments and I have to thank my friends there for launching me on what has proven to be a most stimulating, although at times disturbing, line of inquiry.

My first post on PAINS examined some of the basic science and cheminformatics behind the substructural filters used. One observation that I'll make is that cheminformaticians would have done themselves rather more credit if, instead of implementing PAINS filters quite so enthusiastically, they'd first taken a more forensic look at how the filters had been derived. Singlet oxygen is an integral component of the AlphaScreen technology used in all six assays that formed the basis of the original PAINS study and the second post explored some of the consequences of this reliance on singlet oxygen. The third post was written as a 'get out of jail' card for those who need to get their use of PAINS past manuscript reviewers but, on a more serious note, it does pose some questions about how much we actually know about the behavior of PAINS compounds. The final PAINS post emphasized the need to make a clear distinction in science between what we know and what we believe. If we are unable (or unwilling) to demonstrate that we can do this in drug discovery then those who fund our work may conclude that the difficulties we face are of our own making.

There's actually a lot more to Sarajevo than dead Habsburgs and the city hosted the  1984 Winter Olympics. I took a taxi to the top of the bobsled run and walked back down to the city. Here are some photos. 




  


  
  

So I guess you're wondering where the 1984 bobsled run fits into drug discovery.  Ligand efficiency is, in essence, about slopes and intercepts and, like bobsledders, ligand efficiency advocates prefer not to think about intercepts.  I did two posts on ligand efficiency in 2015. The first post was a response to an article in which our criticism of ligand efficiency metrics was denounced as noise although, in the manner of Pravda, the article didn't actually say what the criticism was and I was left with the impression of a panicky batsman desperately trying to fend off a throat ball that had lifted sharply off just short of a length. The second post explored the link between ligand efficiency and homeopathy. 

I have described ligand efficiency as not even wrong and it also fits snugly into the voodoo thermodynamics category. Sometimes I think that if a coiled dog turd could be converted to molar energy units and scaled by coil radius then it would get adopted as a metric (which we might call 'scatological efficiency'). Voodoo thermodynamics is likely to feature more frequently in 2016 although I did manage one post on this topic in 2015. 

I took the train from Sarajevo to Mostar and the next four photos show a guy jumping, as is the local custom, off the reconstructed Stari Most into the Neretva River.


 





Now I guess you're wondering what a guy jumping off a bridge in Herzegovina has to do with molecular design and the quick answer is nothing at all. During the course of the year I jumped off a bridge of sorts (more accurately out of my applicability domain) with a post on Open Access and there'll hopefully be more of this sort of thing this year. This is probably a good point to wrap up the review of 2015 and I look forward to seeing you towards the end of the month when you'll meet the boys who cried wolf.