You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: README.md
+45-1
Original file line number
Diff line number
Diff line change
@@ -53,9 +53,10 @@ pip install faesm
53
53
54
54
## ESM2
55
55
56
-
FAESM is a drop-in replacement for the official ESM implementation. You can use the same code as you would use the official ESM implementation. For example:import torch
56
+
FAESM is a drop-in replacement for the official ESM implementation. You can use the same code as you would use the official ESM implementation. For example:
assertabs(ce_eval -2.4) <0.1# 2.4 is the reference ce for the official progen2-small
116
+
```
117
+
76
118
## ESM-C
77
119
78
120
Right after EvolutionaryScale release [ESM-C](https://www.evolutionaryscale.ai/blog/esm-cambrian), we follow up with the flash attention version of ESM-C in FAESM. You can run ESM-C easily with the following code:
@@ -94,6 +137,7 @@ It's recommended to use the flash attention for training. Because in the forward
94
137
95
138
# Benchmarking
96
139
140
+
97
141
### FAESM vs. Official ESM2
98
142
99
143
Below is the comparison of peak memory usage and inference time of FAESM with the official ESM2. We show that FAESM can save memory usage by up to 60% and inference time by up to 70% (length 1000). The benchmarking is done on ESM-650M with batch size 8, and a single A100 with 80GB of memory.
0 commit comments