Skip to content

Commit 694739d

Browse files
committed
Simplify grep complexity
1 parent 9974028 commit 694739d

File tree

1 file changed

+5
-23
lines changed

1 file changed

+5
-23
lines changed

section_grep.md

+5-23
Original file line numberDiff line numberDiff line change
@@ -136,41 +136,23 @@ Audience question
136136

137137
<!-- .slide: class="audience-question" -->
138138

139-
# Complexity
140-
141-
## Grep
139+
# Grep Complexity
142140

143141
Search every query term as a string in every document: <!-- .element: class="fragment" -->
144142

145143
$$O(\text{num query terms} \times \text{total length of all documents})$$<!-- .element: class="fragment" -->
146144

147-
## Union
148-
149-
Merge result lists (without duplicates): <!-- .element: class="fragment" -->
150-
151-
$$O(\text{number of results})$$<!-- .element: class="fragment" -->
152-
153-
## Intersect
154-
155-
Compare the first result list with every other: <!-- .element: class="fragment" -->
156-
157-
$$O(\text{num query terms} \times \text{num results per query term})$$<!-- .element: class="fragment" -->
145+
Can take reaaally long<!-- .element: class="fragment" -->
158146

159147
Notes:
160148
Audience question
161149
---
162150

163-
# Grep complexity
164-
165-
* $O(\text{num query terms} \times \text{total length of all documents})$
166-
* Can take reaaally long
167-
168-
#### <!-- .element: class="fragment" data-fragment-index="1" --> Example
151+
# Grep complexity example
169152

170-
* &shy;<!-- .element: class="fragment" data-fragment-index="1" -->*English Wikipedia*: 6M articles, 12B characters, 1.2M
153+
* *English Wikipedia*: 6M articles, 12B characters, 1.2M
171154
distinct terms
172-
* &shy;<!-- .element: class="fragment" data-fragment-index="1" -->grep: 2 query terms &times; 12GB = **24 billion string
173-
comparisons**
155+
* grep: 2 query terms &times; 12GB = **24 billion string comparisons**
174156

175157
Notes:
176158
How can this be improved?

0 commit comments

Comments
 (0)