F@H deals with a specific set to work with.
DF approaches from a broad idea to isolate stuff that doesn't work and stuff that has a higher potential to work roughly 5 terra of memory or the size of a persons brain. Trials were going on using small and now larger proteins to target against using now 10 billion structures.
phase 1a:
1ENH: 100%, best RMSD: 3.64 (March 15, 2002)
1PMC: 100%, best RMSD: 3.96 (March 23, 2002)
5PTI: 100%, best RMSD: 5.40 (April 1, 2002)
1VII: 100%, best RMSD: 2.19 (April 8, 2002)
1SHG: 100%, best RMSD: 6.52 (April 15, 2002)
phase 1b:
4ICB: 34.0%, best RMSD: 4.61
CASP5 targets:
T0129: 100%, best pseudo-energy: 42.32 (June 11, 2002)
T0131: 100%, best pseudo-energy: 1.3 (June 18, 2002)
T0135: 100%, best pseudo-energy: 0.70 (June 25, 2002)
T0138: 100%, best pseudo-energy: 0.39 (July 3, 2002)
T0139: 100%, best pseudo-energy: 6.79 (July 9, 2002)
T0148: 100%, best pseudo-energy: 47.79 (July 16, 2002)
T0147N: 100%, best pseudo-energy: 22.35 (July 23, 2002)
T0157: 100%, best pseudo-energy: 54.42 (July 30, 2002)
T0156: 981.8 million, best pseudo-energy: 44.16
Proteins have a vast number of folds, larger than we could hope to compute even with distributed computing. Usually only one fold is found in nature. The Distributed Folding Project aims to test our new protein folding algorithm. We want see if it can reproduce natural protein folds after making extremely large samples of many different folds.
With your help, we will create the largest samples of protein folds ever computed. First we will make 1 Billion (1,000,000,000) folds for 5 small proteins, then we will try 10 Billion (10,000,000,000) for another 10 large proteins. By the end of our first phase, we hope to make over 100 Billion protein folds spanning 15 different proteins.
muttley