You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Most fuzzy search libraries rely on optimizing the hell out of the same couple of fuzzy search algorithms (Hamming distance, Levenshtein distance). Sometimes unfortunately due to the complexity of these algorithms, no amount of optimization will get you the speed, that you want.
15
29
@@ -93,7 +107,7 @@ from sklearn.feature_extraction.text import TfidfVectorizer
93
107
process = Process(vectorizer, metric="cosine")
94
108
```
95
109
96
-
### Dimentionality Reduction
110
+
### Dimensionality Reduction
97
111
98
112
You might find that the speed of your fuzzy search process is not sufficient. In this case it might be desirable to reduce the dimentionality of the produced vectors with some matrix decomposition method or topic model.
99
113
@@ -107,7 +121,7 @@ from sklearn.pipeline import make_pipeline
0 commit comments