Hi Greg, reopening this old question. I can see that there are potential differences between rdkit version and especially Linux and Windows but let's lieave that aside for now.
After further "playing around" however I really have the impression there is a real issue with running rdkit (or python?) in a virtualized operating sytem. Since most production software and/or when using the cloud will mostly run in a virtualized operating system, I think this should be a fairly relevant topic worth investigation. As you showed yourself, the AWS System also was fairly slow. For following observations I'm keeping the same datasets as before which is from your blog post ( /Regress/Scripts/fingerprint_screenout.py). basically it's that code slightly adapted: mols = [] with gzip.open(data_dir + 'chembl21_25K.pairs.txt.gz', 'rb') as inf: for line in inf: line = line.decode().strip().split() smi1 = line[1] smi2 = line[3] m1 = Chem.MolFromSmiles(smi1) m2 = Chem.MolFromSmiles(smi2) mols.append(m1) mols.append(m2) frags = [Chem.MolFromSmiles(x.split()[0]) for x in open(data_dir + 'zinc.frags.500.q.smi', 'r')] mfps = [Chem.PatternFingerprint(m, 512) for m in mols] fragsfps = [Chem.PatternFingerprint(m, 512) for m in frags] %%timeit -n1 -r1 for i, fragfp in enumerate(fragsfps): hits = 0 for j, mfp in enumerate(mfps): if DataStructs.AllProbeBitsMatch(fragfp, mfp): if mols[j].HasSubstructMatch(frags[i]): hits = hits + 1 I want to focus on the last cell and namley the "AllProbeBitsMatch" method: %%timeit DataStructs.AllProbeBitsMatch(fragsfps[10], mfps[10]) Results: Windows 10 native i7-8850H: 567 ns ± 5.48 ns per loop (mean ± std. dev. of 7 runs, 1000000 loops each) Lubuntu 16.04 virtualized i7-8850H: 1.81 µs ± 56.7 ns per loop (mean ± std. dev. of 7 runs, 1000000 loops each) // the high variation is consistent Windows Server 2012 R2 virtualized Xeon E5-2620 v4: 1.18 µs ± 4.09 ns per loop (mean ± std. dev. of 7 runs, 1000000 loops each) So it seems virtualization seems to reduce the performance of this specific method by half which is also what I see by running the full substructure search code which takes double the time on the virtualized machines. (The windows server actually runs on ESX (eg type 1 hypervisor) while the Lubuntu VM is a type 2 (Vmware workstation) but both seem to suffer the same.). we can try same thing with %%timeit mols[10].HasSubstructMatch(frags[10]) The difference here is smaller but VMs also take >50% more time. So there seems to be a consistent large performance impact in VMs. Of course the VM will be a bit slower but not by that much? What am I missing? Other experiences? Best Regards, Thomas ________________________________ Von: Greg Landrum <greg.land...@gmail.com> Gesendet: Montag, 16. Dezember 2019 17:10 An: Thomas Strunz <beginn...@hotmail.de> Cc: rdkit-discuss@lists.sourceforge.net <rdkit-discuss@lists.sourceforge.net> Betreff: Re: [Rdkit-discuss] Observations about RDKit performance: PatternFingerprinter, Windows, Linux and Virtual machines Hi Thomas, First it is important to compare equivalent major versions to each other. Particularly in this case. On my linux box generating the pattern fingerprints takes 24.2 seconds with v2019.03.x and 15.9 seconds with v2019.09.x (that's due to the improvements in the substructure matcher that the blog post you link to discusses). Comparing the same versions to each other: Performance on windows vs linux Windows performance with the RDKit has always lagged behind linux performance. There's something in the code (or in the way we use the compiler) that leads to big differences on some benchmarks. The most straightforward way I can demonstrate this is with results from my windows 10 laptop. Here's the output when running the fingerprint_screenout.py benchmark using the windows build: | 2019.09.1 | 13.6 | 0.3 | 38.1 | 0.8 | 25.5 | 25.9 | 84.1 | and here's the output from a linux build running on the Windows Linux Subsystem: | 2019.09.2 | 10.7 | 0.2 | 19.3 | 0.4 | 19.4 | 19.2 | 53.2 | You can see the differences are not small. I haven't invested massive time into it, but I haven't been able to figure out what causes this. Performance on (linux) VMs I can't think of any particular reason why there should be huge differences and it's really difficult to compare apples to apples here. Since I have the numbers, here's one comparison Here's a run on my linux workstation: | 2019.09.2 | 7.6 | 0.3 | 15.9 | 0.4 | 21.4 | 20.4 | 55.7 | and here's the same thing on an AWS t3.xlarge instance: | 2019.09.2 | 9.6 | 0.2 | 20.3 | 0.4 | 38.4 | 38.2 | 94.7 | The VM is significantly slower, but t3.xlarge an instance type that's intended to be used for compute intensive jobs (I don't have on of those active and configured at the moment). Does that help at all? -greg On Mon, Dec 16, 2019 at 8:27 AM Thomas Strunz <beginn...@hotmail.de<mailto:beginn...@hotmail.de>> wrote: Hi All, I was looking at a blog post from greg: https://rdkit.blogspot.com/2019/07/a-couple-of-substructure-search-topics.html about fingerprint screenout. The part that got me confused was the timings in his blog post because run times in my case where a lot slower. Gregs numbers: [07:21:19] INFO: mols from smiles [07:21:27] INFO: Results1: 7.77 seconds, 50000 mols [07:21:27] INFO: queries from smiles [07:21:27] INFO: Results2: 0.16 seconds [07:21:27] INFO: generating pattern fingerprints for mols [07:21:43] INFO: Results3: 16.11 seconds [07:21:43] INFO: generating pattern fingerprints for queries [07:21:43] INFO: Results4: 0.34 seconds [07:21:43] INFO: testing frags queries [07:22:03] INFO: Results5: 19.90 seconds. 6753 tested (0.0003 of total), 3989 found, 0.59 accuracy. 0 errors. [07:22:03] INFO: testing leads queries [07:22:23] INFO: Results6: 19.77 seconds. 1586 tested (0.0001 of total), 1067 found, 0.67 accuracy. 0 errors. [07:22:23] INFO: testing pieces queries [07:23:19] INFO: Results7: 55.37 seconds. 3333202 tested (0.0810 of total), 1925628 found, 0.58 accuracy. 0 errors. | 2019.09.1dev1 | 7.8 | 0.2 | 16.1 | 0.3 | 19.9 | 19.8 | 55.4 | Machine 1: Virtual machine, Windows Server 2012 R2 with an intel xeon (4 virtual cores) Since the test is single-threaded it makes a bit of sense that it isn't fast here but it's not just a bit slower, but a lot slower, depending on test almost 3xtimes slower [09:03:19] INFO: mols from smiles [09:03:38] INFO: Results1: 19.44 seconds, 50000 mols [09:03:38] INFO: queries from smiles [09:03:38] INFO: Results2: 0.36 seconds [09:03:38] INFO: generating pattern fingerprints for mols [09:04:54] INFO: Results3: 75.99 seconds [09:04:54] INFO: generating pattern fingerprints for queries [09:04:56] INFO: Results4: 1.55 seconds [09:04:56] INFO: testing frags queries [09:05:34] INFO: Results5: 37.59 seconds. 6753 tested (0.0003 of total), 3989 f ound, 0.59 accuracy. 0 errors. [09:05:34] INFO: testing leads queries [09:06:11] INFO: Results6: 37.34 seconds. 1586 tested (0.0001 of total), 1067 f ound, 0.67 accuracy. 0 errors. [09:06:11] INFO: testing pieces queries [09:08:39] INFO: Results7: 147.79 seconds. 3333202 tested (0.0810 of total), 19 25628 found, 0.58 accuracy. 0 errors. | 2019.03.3 | 19.4 | 0.4 | 76.0 | 1.5 | 37.6 | 37.3 | 147.8 | I thought maybe another issue with windows being slow so I tested on a linux VM on my laptop Machine 2: Virtual machine, Lubuntu 16.04 on a laptop i7-8850H 6-core [09:23:31] INFO: mols from smiles [09:23:54] INFO: Results1: 23.71 seconds, 50000 mols [09:23:54] INFO: queries from smiles [09:23:55] INFO: Results2: 0.48 seconds [09:23:55] INFO: generating pattern fingerprints for mols [09:24:53] INFO: Results3: 58.31 seconds [09:24:53] INFO: generating pattern fingerprints for queries [09:24:54] INFO: Results4: 1.19 seconds [09:24:54] INFO: testing frags queries [09:25:41] INFO: Results5: 46.22 seconds. 6753 tested (0.0003 of total), 3989 found, 0.59 accuracy. 0 errors. [09:25:41] INFO: testing leads queries [09:26:26] INFO: Results6: 45.84 seconds. 1586 tested (0.0001 of total), 1067 found, 0.67 accuracy. 0 errors. [09:26:26] INFO: testing pieces queries [09:28:33] INFO: Results7: 126.78 seconds. 3333202 tested (0.0810 of total), 1925628 found, 0.58 accuracy. 0 errors. | 2019.03.3 | 23.7 | 0.5 | 58.3 | 1.2 | 46.2 | 45.8 | 126.8 | Pretty weird sometimes even slower sometimes faster than the windows VM but still a lot slower than Gregs numbers (I repeated with rdkit 2019.09.2 and got comparable results) So I also tested on above laptop directly: Machine 3: physical install, windows 10 on a laptop i7-8850H 6-core (same machine as 2) [09:51:43] INFO: mols from smiles [09:51:54] INFO: Results1: 10.59 seconds, 50000 mols [09:51:54] INFO: queries from smiles [09:51:54] INFO: Results2: 0.20 seconds [09:51:54] INFO: generating pattern fingerprints for mols [09:52:24] INFO: Results3: 29.50 seconds [09:52:24] INFO: generating pattern fingerprints for queries [09:52:24] INFO: Results4: 0.61 seconds [09:52:24] INFO: testing frags queries [09:52:44] INFO: Results5: 19.71 seconds. 6753 tested (0.0003 of total), 3989 found, 0.59 accuracy. 0 errors. [09:52:44] INFO: testing leads queries [09:53:04] INFO: Results6: 19.48 seconds. 1586 tested (0.0001 of total), 1067 found, 0.67 accuracy. 0 errors. [09:53:04] INFO: testing pieces queries [09:54:05] INFO: Results7: 61.94 seconds. 3333202 tested (0.0810 of total), 1925628 found, 0.58 accuracy. 0 errors. | 2019.09.1 | 10.6 | 0.2 | 29.5 | 0.6 | 19.7 | 19.5 | 61.9 | This is much closer to Gregs results, except for the fingerprinting which takes almost double the time. Also notice how the fingerprinting on the linux VM is much faster also compared to other results than on the windows VM? Conclusions: 1. Form what I see, it seems that the pattern fingerprinter runs a lot slower on windows. Is this known issue? 2. In virtual machines the rdkits performance simply tanks, is much worse. A certain penalty is to be expected but not this much. Or what am I missing? Machine 1 runs on central infrastructure so I would assume virtualization is configured correctly. For the local VM, vt-x is enabled. Yet it is much slower compared to the physical machine (plus that AFAIK rdkit runs faster in linux vs windows) Especially the virtual machine aspect is kind of troubling because I would assume many real-world applications are deployed as VM and hence might suffer from this too? I don't have a well defined question but more interested in other users experience especially regarding the virtualization. Best Regards, Thomas _______________________________________________ Rdkit-discuss mailing list Rdkit-discuss@lists.sourceforge.net<mailto:Rdkit-discuss@lists.sourceforge.net> https://lists.sourceforge.net/lists/listinfo/rdkit-discuss
_______________________________________________ Rdkit-discuss mailing list Rdkit-discuss@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/rdkit-discuss