| 1 | /*! |
| 2 | This module contains types and routines for implementing determinization. |
| 3 | |
| 4 | In this crate, there are at least two places where we implement |
| 5 | determinization: fully ahead-of-time compiled DFAs in the `dfa` module and |
| 6 | lazily compiled DFAs in the `hybrid` module. The stuff in this module |
| 7 | corresponds to the things that are in common between these implementations. |
| 8 | |
| 9 | There are three broad things that our implementations of determinization have |
| 10 | in common, as defined by this module: |
| 11 | |
| 12 | * The classification of start states. That is, whether we're dealing with |
| 13 | word boundaries, line boundaries, etc., is all the same. This also includes |
| 14 | the look-behind assertions that are satisfied by each starting state |
| 15 | classification. |
| 16 | * The representation of DFA states as sets of NFA states, including |
| 17 | convenience types for building these DFA states that are amenable to reusing |
| 18 | allocations. |
| 19 | * Routines for the "classical" parts of determinization: computing the |
| 20 | epsilon closure, tracking match states (with corresponding pattern IDs, since |
| 21 | we support multi-pattern finite automata) and, of course, computing the |
| 22 | transition function between states for units of input. |
| 23 | |
| 24 | I did consider a couple of alternatives to this particular form of code reuse: |
| 25 | |
| 26 | 1. Don't do any code reuse. The problem here is that we *really* want both |
| 27 | forms of determinization to do exactly identical things when it comes to |
| 28 | their handling of NFA states. While our tests generally ensure this, the code |
| 29 | is tricky and large enough where not reusing code is a pretty big bummer. |
| 30 | |
| 31 | 2. Implement all of determinization once and make it generic over fully |
| 32 | compiled DFAs and lazily compiled DFAs. While I didn't actually try this |
| 33 | approach, my instinct is that it would be more complex than is needed here. |
| 34 | And the interface required would be pretty hairy. Instead, I think splitting |
| 35 | it into logical sub-components works better. |
| 36 | */ |
| 37 | |
| 38 | use alloc::vec::Vec; |
| 39 | |
| 40 | pub(crate) use self::state::{ |
| 41 | State, StateBuilderEmpty, StateBuilderMatches, StateBuilderNFA, |
| 42 | }; |
| 43 | |
| 44 | use crate::{ |
| 45 | nfa::thompson, |
| 46 | util::{ |
| 47 | alphabet, |
| 48 | look::{Look, LookSet}, |
| 49 | primitives::StateID, |
| 50 | search::MatchKind, |
| 51 | sparse_set::{SparseSet, SparseSets}, |
| 52 | start::Start, |
| 53 | utf8, |
| 54 | }, |
| 55 | }; |
| 56 | |
| 57 | mod state; |
| 58 | |
| 59 | /// Compute the set of all reachable NFA states, including the full epsilon |
| 60 | /// closure, from a DFA state for a single unit of input. The set of reachable |
| 61 | /// states is returned as a `StateBuilderNFA`. The `StateBuilderNFA` returned |
| 62 | /// also includes any look-behind assertions satisfied by `unit`, in addition |
| 63 | /// to whether it is a match state. For multi-pattern DFAs, the builder will |
| 64 | /// also include the pattern IDs that match (in the order seen). |
| 65 | /// |
| 66 | /// `nfa` must be able to resolve any NFA state in `state` and any NFA state |
| 67 | /// reachable via the epsilon closure of any NFA state in `state`. `sparses` |
| 68 | /// must have capacity equivalent to `nfa.len()`. |
| 69 | /// |
| 70 | /// `match_kind` should correspond to the match semantics implemented by the |
| 71 | /// DFA being built. Generally speaking, for leftmost-first match semantics, |
| 72 | /// states that appear after the first NFA match state will not be included in |
| 73 | /// the `StateBuilderNFA` returned since they are impossible to visit. |
| 74 | /// |
| 75 | /// `sparses` is used as scratch space for NFA traversal. Other than their |
| 76 | /// capacity requirements (detailed above), there are no requirements on what's |
| 77 | /// contained within them (if anything). Similarly, what's inside of them once |
| 78 | /// this routine returns is unspecified. |
| 79 | /// |
| 80 | /// `stack` must have length 0. It is used as scratch space for depth first |
| 81 | /// traversal. After returning, it is guaranteed that `stack` will have length |
| 82 | /// 0. |
| 83 | /// |
| 84 | /// `state` corresponds to the current DFA state on which one wants to compute |
| 85 | /// the transition for the input `unit`. |
| 86 | /// |
| 87 | /// `empty_builder` corresponds to the builder allocation to use to produce a |
| 88 | /// complete `StateBuilderNFA` state. If the state is not needed (or is already |
| 89 | /// cached), then it can be cleared and reused without needing to create a new |
| 90 | /// `State`. The `StateBuilderNFA` state returned is final and ready to be |
| 91 | /// turned into a `State` if necessary. |
| 92 | pub(crate) fn next( |
| 93 | nfa: &thompson::NFA, |
| 94 | match_kind: MatchKind, |
| 95 | sparses: &mut SparseSets, |
| 96 | stack: &mut Vec<StateID>, |
| 97 | state: &State, |
| 98 | unit: alphabet::Unit, |
| 99 | empty_builder: StateBuilderEmpty, |
| 100 | ) -> StateBuilderNFA { |
| 101 | sparses.clear(); |
| 102 | |
| 103 | // Whether the NFA is matched in reverse or not. We use this in some |
| 104 | // conditional logic for dealing with the exceptionally annoying CRLF-aware |
| 105 | // line anchors. |
| 106 | let rev = nfa.is_reverse(); |
| 107 | // The look-around matcher that our NFA is configured with. We don't |
| 108 | // actually use it to match look-around assertions, but we do need its |
| 109 | // configuration for constructing states consistent with how it matches. |
| 110 | let lookm = nfa.look_matcher(); |
| 111 | |
| 112 | // Put the NFA state IDs into a sparse set in case we need to |
| 113 | // re-compute their epsilon closure. |
| 114 | // |
| 115 | // Doing this state shuffling is technically not necessary unless some |
| 116 | // kind of look-around is used in the DFA. Some ad hoc experiments |
| 117 | // suggested that avoiding this didn't lead to much of an improvement, |
| 118 | // but perhaps more rigorous experimentation should be done. And in |
| 119 | // particular, avoiding this check requires some light refactoring of |
| 120 | // the code below. |
| 121 | state.iter_nfa_state_ids(|nfa_id| { |
| 122 | sparses.set1.insert(nfa_id); |
| 123 | }); |
| 124 | |
| 125 | // Compute look-ahead assertions originating from the current state. Based |
| 126 | // on the input unit we're transitioning over, some additional set of |
| 127 | // assertions may be true. Thus, we re-compute this state's epsilon closure |
| 128 | // (but only if necessary). Notably, when we build a DFA state initially, |
| 129 | // we don't enable any look-ahead assertions because we don't know whether |
| 130 | // they're true or not at that point. |
| 131 | if !state.look_need().is_empty() { |
| 132 | // Add look-ahead assertions that are now true based on the current |
| 133 | // input unit. |
| 134 | let mut look_have = state.look_have().clone(); |
| 135 | match unit.as_u8() { |
| 136 | Some(b' \r' ) => { |
| 137 | if !rev || !state.is_half_crlf() { |
| 138 | look_have = look_have.insert(Look::EndCRLF); |
| 139 | } |
| 140 | } |
| 141 | Some(b' \n' ) => { |
| 142 | if rev || !state.is_half_crlf() { |
| 143 | look_have = look_have.insert(Look::EndCRLF); |
| 144 | } |
| 145 | } |
| 146 | Some(_) => {} |
| 147 | None => { |
| 148 | look_have = look_have |
| 149 | .insert(Look::End) |
| 150 | .insert(Look::EndLF) |
| 151 | .insert(Look::EndCRLF); |
| 152 | } |
| 153 | } |
| 154 | if unit.is_byte(lookm.get_line_terminator()) { |
| 155 | look_have = look_have.insert(Look::EndLF); |
| 156 | } |
| 157 | if state.is_half_crlf() |
| 158 | && ((rev && !unit.is_byte(b' \r' )) |
| 159 | || (!rev && !unit.is_byte(b' \n' ))) |
| 160 | { |
| 161 | look_have = look_have.insert(Look::StartCRLF); |
| 162 | } |
| 163 | if state.is_from_word() == unit.is_word_byte() { |
| 164 | look_have = look_have |
| 165 | .insert(Look::WordAsciiNegate) |
| 166 | .insert(Look::WordUnicodeNegate); |
| 167 | } else { |
| 168 | look_have = |
| 169 | look_have.insert(Look::WordAscii).insert(Look::WordUnicode); |
| 170 | } |
| 171 | if !unit.is_word_byte() { |
| 172 | look_have = look_have |
| 173 | .insert(Look::WordEndHalfAscii) |
| 174 | .insert(Look::WordEndHalfUnicode); |
| 175 | } |
| 176 | if state.is_from_word() && !unit.is_word_byte() { |
| 177 | look_have = look_have |
| 178 | .insert(Look::WordEndAscii) |
| 179 | .insert(Look::WordEndUnicode); |
| 180 | } else if !state.is_from_word() && unit.is_word_byte() { |
| 181 | look_have = look_have |
| 182 | .insert(Look::WordStartAscii) |
| 183 | .insert(Look::WordStartUnicode); |
| 184 | } |
| 185 | // If we have new assertions satisfied that are among the set of |
| 186 | // assertions that exist in this state (that is, just because we added |
| 187 | // an EndLF assertion above doesn't mean there is an EndLF conditional |
| 188 | // epsilon transition in this state), then we re-compute this state's |
| 189 | // epsilon closure using the updated set of assertions. |
| 190 | // |
| 191 | // Note that since our DFA states omit unconditional epsilon |
| 192 | // transitions, this check is necessary for correctness. If we re-did |
| 193 | // the epsilon closure below needlessly, it could change based on the |
| 194 | // fact that we omitted epsilon states originally. |
| 195 | if !look_have |
| 196 | .subtract(state.look_have()) |
| 197 | .intersect(state.look_need()) |
| 198 | .is_empty() |
| 199 | { |
| 200 | for nfa_id in sparses.set1.iter() { |
| 201 | epsilon_closure( |
| 202 | nfa, |
| 203 | nfa_id, |
| 204 | look_have, |
| 205 | stack, |
| 206 | &mut sparses.set2, |
| 207 | ); |
| 208 | } |
| 209 | sparses.swap(); |
| 210 | sparses.set2.clear(); |
| 211 | } |
| 212 | } |
| 213 | |
| 214 | // Convert our empty builder into one that can record assertions and match |
| 215 | // pattern IDs. |
| 216 | let mut builder = empty_builder.into_matches(); |
| 217 | // Set whether the StartLF look-behind assertion is true for this |
| 218 | // transition or not. The look-behind assertion for ASCII word boundaries |
| 219 | // is handled below. |
| 220 | if nfa.look_set_any().contains_anchor_line() |
| 221 | && unit.is_byte(lookm.get_line_terminator()) |
| 222 | { |
| 223 | // Why only handle StartLF here and not Start? That's because Start |
| 224 | // can only impact the starting state, which is special cased in |
| 225 | // start state handling. |
| 226 | builder.set_look_have(|have| have.insert(Look::StartLF)); |
| 227 | } |
| 228 | // We also need to add StartCRLF to our assertions too, if we can. This |
| 229 | // is unfortunately a bit more complicated, because it depends on the |
| 230 | // direction of the search. In the forward direction, ^ matches after a |
| 231 | // \n, but in the reverse direction, ^ only matches after a \r. (This is |
| 232 | // further complicated by the fact that reverse a regex means changing a ^ |
| 233 | // to a $ and vice versa.) |
| 234 | if nfa.look_set_any().contains_anchor_crlf() |
| 235 | && ((rev && unit.is_byte(b' \r' )) || (!rev && unit.is_byte(b' \n' ))) |
| 236 | { |
| 237 | builder.set_look_have(|have| have.insert(Look::StartCRLF)); |
| 238 | } |
| 239 | // And also for the start-half word boundary assertions. As long as the |
| 240 | // look-behind byte is not a word char, then the assertions are satisfied. |
| 241 | if nfa.look_set_any().contains_word() && !unit.is_word_byte() { |
| 242 | builder.set_look_have(|have| { |
| 243 | have.insert(Look::WordStartHalfAscii) |
| 244 | .insert(Look::WordStartHalfUnicode) |
| 245 | }); |
| 246 | } |
| 247 | for nfa_id in sparses.set1.iter() { |
| 248 | match *nfa.state(nfa_id) { |
| 249 | thompson::State::Union { .. } |
| 250 | | thompson::State::BinaryUnion { .. } |
| 251 | | thompson::State::Fail |
| 252 | | thompson::State::Look { .. } |
| 253 | | thompson::State::Capture { .. } => {} |
| 254 | thompson::State::Match { pattern_id } => { |
| 255 | // Notice here that we are calling the NEW state a match |
| 256 | // state if the OLD state we are transitioning from |
| 257 | // contains an NFA match state. This is precisely how we |
| 258 | // delay all matches by one byte and also what therefore |
| 259 | // guarantees that starting states cannot be match states. |
| 260 | // |
| 261 | // If we didn't delay matches by one byte, then whether |
| 262 | // a DFA is a matching state or not would be determined |
| 263 | // by whether one of its own constituent NFA states |
| 264 | // was a match state. (And that would be done in |
| 265 | // 'add_nfa_states'.) |
| 266 | // |
| 267 | // Also, 'add_match_pattern_id' requires that callers never |
| 268 | // pass duplicative pattern IDs. We do in fact uphold that |
| 269 | // guarantee here, but it's subtle. In particular, a Thompson |
| 270 | // NFA guarantees that each pattern has exactly one match |
| 271 | // state. Moreover, since we're iterating over the NFA state |
| 272 | // IDs in a set, we are guarateed not to have any duplicative |
| 273 | // match states. Thus, it is impossible to add the same pattern |
| 274 | // ID more than once. |
| 275 | // |
| 276 | // N.B. We delay matches by 1 byte as a way to hack 1-byte |
| 277 | // look-around into DFA searches. This lets us support ^, $ |
| 278 | // and ASCII-only \b. The delay is also why we need a special |
| 279 | // "end-of-input" (EOI) sentinel and why we need to follow the |
| 280 | // EOI sentinel at the end of every search. This final EOI |
| 281 | // transition is necessary to report matches found at the end |
| 282 | // of a haystack. |
| 283 | builder.add_match_pattern_id(pattern_id); |
| 284 | if !match_kind.continue_past_first_match() { |
| 285 | break; |
| 286 | } |
| 287 | } |
| 288 | thompson::State::ByteRange { ref trans } => { |
| 289 | if trans.matches_unit(unit) { |
| 290 | epsilon_closure( |
| 291 | nfa, |
| 292 | trans.next, |
| 293 | builder.look_have(), |
| 294 | stack, |
| 295 | &mut sparses.set2, |
| 296 | ); |
| 297 | } |
| 298 | } |
| 299 | thompson::State::Sparse(ref sparse) => { |
| 300 | if let Some(next) = sparse.matches_unit(unit) { |
| 301 | epsilon_closure( |
| 302 | nfa, |
| 303 | next, |
| 304 | builder.look_have(), |
| 305 | stack, |
| 306 | &mut sparses.set2, |
| 307 | ); |
| 308 | } |
| 309 | } |
| 310 | thompson::State::Dense(ref dense) => { |
| 311 | if let Some(next) = dense.matches_unit(unit) { |
| 312 | epsilon_closure( |
| 313 | nfa, |
| 314 | next, |
| 315 | builder.look_have(), |
| 316 | stack, |
| 317 | &mut sparses.set2, |
| 318 | ); |
| 319 | } |
| 320 | } |
| 321 | } |
| 322 | } |
| 323 | // We only set the word byte if there's a word boundary look-around |
| 324 | // anywhere in this regex. Otherwise, there's no point in bloating the |
| 325 | // number of states if we don't have one. |
| 326 | // |
| 327 | // We also only set it when the state has a non-zero number of NFA states. |
| 328 | // Otherwise, we could wind up with states that *should* be DEAD states |
| 329 | // but are otherwise distinct from DEAD states because of this look-behind |
| 330 | // assertion being set. While this can't technically impact correctness *in |
| 331 | // theory*, it can create pathological DFAs that consume input until EOI or |
| 332 | // a quit byte is seen. Consuming until EOI isn't a correctness problem, |
| 333 | // but a (serious) perf problem. Hitting a quit byte, however, could be a |
| 334 | // correctness problem since it could cause search routines to report an |
| 335 | // error instead of a detected match once the quit state is entered. (The |
| 336 | // search routine could be made to be a bit smarter by reporting a match |
| 337 | // if one was detected once it enters a quit state (and indeed, the search |
| 338 | // routines in this crate do just that), but it seems better to prevent |
| 339 | // these things by construction if possible.) |
| 340 | if !sparses.set2.is_empty() { |
| 341 | if nfa.look_set_any().contains_word() && unit.is_word_byte() { |
| 342 | builder.set_is_from_word(); |
| 343 | } |
| 344 | if nfa.look_set_any().contains_anchor_crlf() |
| 345 | && ((rev && unit.is_byte(b' \n' )) || (!rev && unit.is_byte(b' \r' ))) |
| 346 | { |
| 347 | builder.set_is_half_crlf(); |
| 348 | } |
| 349 | } |
| 350 | let mut builder_nfa = builder.into_nfa(); |
| 351 | add_nfa_states(nfa, &sparses.set2, &mut builder_nfa); |
| 352 | builder_nfa |
| 353 | } |
| 354 | |
| 355 | /// Compute the epsilon closure for the given NFA state. The epsilon closure |
| 356 | /// consists of all NFA state IDs, including `start_nfa_id`, that can be |
| 357 | /// reached from `start_nfa_id` without consuming any input. These state IDs |
| 358 | /// are written to `set` in the order they are visited, but only if they are |
| 359 | /// not already in `set`. `start_nfa_id` must be a valid state ID for the NFA |
| 360 | /// given. |
| 361 | /// |
| 362 | /// `look_have` consists of the satisfied assertions at the current |
| 363 | /// position. For conditional look-around epsilon transitions, these are |
| 364 | /// only followed if they are satisfied by `look_have`. |
| 365 | /// |
| 366 | /// `stack` must have length 0. It is used as scratch space for depth first |
| 367 | /// traversal. After returning, it is guaranteed that `stack` will have length |
| 368 | /// 0. |
| 369 | pub(crate) fn epsilon_closure( |
| 370 | nfa: &thompson::NFA, |
| 371 | start_nfa_id: StateID, |
| 372 | look_have: LookSet, |
| 373 | stack: &mut Vec<StateID>, |
| 374 | set: &mut SparseSet, |
| 375 | ) { |
| 376 | assert!(stack.is_empty()); |
| 377 | // If this isn't an epsilon state, then the epsilon closure is always just |
| 378 | // itself, so there's no need to spin up the machinery below to handle it. |
| 379 | if !nfa.state(start_nfa_id).is_epsilon() { |
| 380 | set.insert(start_nfa_id); |
| 381 | return; |
| 382 | } |
| 383 | |
| 384 | stack.push(start_nfa_id); |
| 385 | while let Some(mut id) = stack.pop() { |
| 386 | // In many cases, we can avoid stack operations when an NFA state only |
| 387 | // adds one new state to visit. In that case, we just set our ID to |
| 388 | // that state and mush on. We only use the stack when an NFA state |
| 389 | // introduces multiple new states to visit. |
| 390 | loop { |
| 391 | // Insert this NFA state, and if it's already in the set and thus |
| 392 | // already visited, then we can move on to the next one. |
| 393 | if !set.insert(id) { |
| 394 | break; |
| 395 | } |
| 396 | match *nfa.state(id) { |
| 397 | thompson::State::ByteRange { .. } |
| 398 | | thompson::State::Sparse { .. } |
| 399 | | thompson::State::Dense { .. } |
| 400 | | thompson::State::Fail |
| 401 | | thompson::State::Match { .. } => break, |
| 402 | thompson::State::Look { look, next } => { |
| 403 | if !look_have.contains(look) { |
| 404 | break; |
| 405 | } |
| 406 | id = next; |
| 407 | } |
| 408 | thompson::State::Union { ref alternates } => { |
| 409 | id = match alternates.get(0) { |
| 410 | None => break, |
| 411 | Some(&id) => id, |
| 412 | }; |
| 413 | // We need to process our alternates in order to preserve |
| 414 | // match preferences, so put the earliest alternates closer |
| 415 | // to the top of the stack. |
| 416 | stack.extend(alternates[1..].iter().rev()); |
| 417 | } |
| 418 | thompson::State::BinaryUnion { alt1, alt2 } => { |
| 419 | id = alt1; |
| 420 | stack.push(alt2); |
| 421 | } |
| 422 | thompson::State::Capture { next, .. } => { |
| 423 | id = next; |
| 424 | } |
| 425 | } |
| 426 | } |
| 427 | } |
| 428 | } |
| 429 | |
| 430 | /// Add the NFA state IDs in the given `set` to the given DFA builder state. |
| 431 | /// The order in which states are added corresponds to the order in which they |
| 432 | /// were added to `set`. |
| 433 | /// |
| 434 | /// The DFA builder state given should already have its complete set of match |
| 435 | /// pattern IDs added (if any) and any look-behind assertions (StartLF, Start |
| 436 | /// and whether this state is being generated for a transition over a word byte |
| 437 | /// when applicable) that are true immediately prior to transitioning into this |
| 438 | /// state (via `builder.look_have()`). The match pattern IDs should correspond |
| 439 | /// to matches that occurred on the previous transition, since all matches are |
| 440 | /// delayed by one byte. The things that should _not_ be set are look-ahead |
| 441 | /// assertions (EndLF, End and whether the next byte is a word byte or not). |
| 442 | /// The builder state should also not have anything in `look_need` set, as this |
| 443 | /// routine will compute that for you. |
| 444 | /// |
| 445 | /// The given NFA should be able to resolve all identifiers in `set` to a |
| 446 | /// particular NFA state. Additionally, `set` must have capacity equivalent |
| 447 | /// to `nfa.len()`. |
| 448 | pub(crate) fn add_nfa_states( |
| 449 | nfa: &thompson::NFA, |
| 450 | set: &SparseSet, |
| 451 | builder: &mut StateBuilderNFA, |
| 452 | ) { |
| 453 | for nfa_id in set.iter() { |
| 454 | match *nfa.state(nfa_id) { |
| 455 | thompson::State::ByteRange { .. } => { |
| 456 | builder.add_nfa_state_id(nfa_id); |
| 457 | } |
| 458 | thompson::State::Sparse { .. } => { |
| 459 | builder.add_nfa_state_id(nfa_id); |
| 460 | } |
| 461 | thompson::State::Dense { .. } => { |
| 462 | builder.add_nfa_state_id(nfa_id); |
| 463 | } |
| 464 | thompson::State::Look { look, .. } => { |
| 465 | builder.add_nfa_state_id(nfa_id); |
| 466 | builder.set_look_need(|need| need.insert(look)); |
| 467 | } |
| 468 | thompson::State::Union { .. } |
| 469 | | thompson::State::BinaryUnion { .. } => { |
| 470 | // Pure epsilon transitions don't need to be tracked as part |
| 471 | // of the DFA state. Tracking them is actually superfluous; |
| 472 | // they won't cause any harm other than making determinization |
| 473 | // slower. |
| 474 | // |
| 475 | // Why aren't these needed? Well, in an NFA, epsilon |
| 476 | // transitions are really just jumping points to other states. |
| 477 | // So once you hit an epsilon transition, the same set of |
| 478 | // resulting states always appears. Therefore, putting them in |
| 479 | // a DFA's set of ordered NFA states is strictly redundant. |
| 480 | // |
| 481 | // Look-around states are also epsilon transitions, but |
| 482 | // they are *conditional*. So their presence could be |
| 483 | // discriminatory, and thus, they are tracked above. |
| 484 | // |
| 485 | // But wait... why are epsilon states in our `set` in the first |
| 486 | // place? Why not just leave them out? They're in our `set` |
| 487 | // because it was generated by computing an epsilon closure, |
| 488 | // and we want to keep track of all states we visited to avoid |
| 489 | // re-visiting them. In exchange, we have to do this second |
| 490 | // iteration over our collected states to finalize our DFA |
| 491 | // state. In theory, we could avoid this second iteration if |
| 492 | // we maintained two sets during epsilon closure: the set of |
| 493 | // visited states (to avoid cycles) and the set of states that |
| 494 | // will actually be used to construct the next DFA state. |
| 495 | // |
| 496 | // Note that this optimization requires that we re-compute the |
| 497 | // epsilon closure to account for look-ahead in 'next' *only |
| 498 | // when necessary*. Namely, only when the set of look-around |
| 499 | // assertions changes and only when those changes are within |
| 500 | // the set of assertions that are needed in order to step |
| 501 | // through the closure correctly. Otherwise, if we re-do the |
| 502 | // epsilon closure needlessly, it could change based on the |
| 503 | // fact that we are omitting epsilon states here. |
| 504 | // |
| 505 | // ----- |
| 506 | // |
| 507 | // Welp, scratch the above. It turns out that recording these |
| 508 | // is in fact necessary to seemingly handle one particularly |
| 509 | // annoying case: when a conditional epsilon transition is |
| 510 | // put inside of a repetition operator. One specific case I |
| 511 | // ran into was the regex `(?:\b|%)+` on the haystack `z%`. |
| 512 | // The correct leftmost first matches are: [0, 0] and [1, 1]. |
| 513 | // But the DFA was reporting [0, 0] and [1, 2]. To understand |
| 514 | // why this happens, consider the NFA for the aforementioned |
| 515 | // regex: |
| 516 | // |
| 517 | // >000000: binary-union(4, 1) |
| 518 | // 000001: \x00-\xFF => 0 |
| 519 | // 000002: WordAscii => 5 |
| 520 | // 000003: % => 5 |
| 521 | // ^000004: binary-union(2, 3) |
| 522 | // 000005: binary-union(4, 6) |
| 523 | // 000006: MATCH(0) |
| 524 | // |
| 525 | // The problem here is that one of the DFA start states is |
| 526 | // going to consist of the NFA states [2, 3] by computing the |
| 527 | // epsilon closure of state 4. State 4 isn't included because |
| 528 | // we previously were not keeping track of union states. But |
| 529 | // only a subset of transitions out of this state will be able |
| 530 | // to follow WordAscii, and in those cases, the epsilon closure |
| 531 | // is redone. The only problem is that computing the epsilon |
| 532 | // closure from [2, 3] is different than computing the epsilon |
| 533 | // closure from [4]. In the former case, assuming the WordAscii |
| 534 | // assertion is satisfied, you get: [2, 3, 6]. In the latter |
| 535 | // case, you get: [2, 6, 3]. Notice that '6' is the match state |
| 536 | // and appears AFTER '3' in the former case. This leads to a |
| 537 | // preferential but incorrect match of '%' before returning |
| 538 | // a match. In the latter case, the match is preferred over |
| 539 | // continuing to accept the '%'. |
| 540 | // |
| 541 | // It almost feels like we might be able to fix the NFA states |
| 542 | // to avoid this, or to at least only keep track of union |
| 543 | // states where this actually matters, since in the vast |
| 544 | // majority of cases, this doesn't matter. |
| 545 | // |
| 546 | // Another alternative would be to define a new HIR property |
| 547 | // called "assertion is repeated anywhere" and compute it |
| 548 | // inductively over the entire pattern. If it happens anywhere, |
| 549 | // which is probably pretty rare, then we record union states. |
| 550 | // Otherwise we don't. |
| 551 | builder.add_nfa_state_id(nfa_id); |
| 552 | } |
| 553 | // Capture states we definitely do not need to record, since they |
| 554 | // are unconditional epsilon transitions with no branching. |
| 555 | thompson::State::Capture { .. } => {} |
| 556 | // It's not totally clear whether we need to record fail states or |
| 557 | // not, but we do so out of an abundance of caution. Since they are |
| 558 | // quite rare in practice, there isn't much cost to recording them. |
| 559 | thompson::State::Fail => { |
| 560 | builder.add_nfa_state_id(nfa_id); |
| 561 | } |
| 562 | thompson::State::Match { .. } => { |
| 563 | // Normally, the NFA match state doesn't actually need to |
| 564 | // be inside the DFA state. But since we delay matches by |
| 565 | // one byte, the matching DFA state corresponds to states |
| 566 | // that transition from the one we're building here. And |
| 567 | // the way we detect those cases is by looking for an NFA |
| 568 | // match state. See 'next' for how this is handled. |
| 569 | builder.add_nfa_state_id(nfa_id); |
| 570 | } |
| 571 | } |
| 572 | } |
| 573 | // If we know this state contains no look-around assertions, then |
| 574 | // there's no reason to track which look-around assertions were |
| 575 | // satisfied when this state was created. |
| 576 | if builder.look_need().is_empty() { |
| 577 | builder.set_look_have(|_| LookSet::empty()); |
| 578 | } |
| 579 | } |
| 580 | |
| 581 | /// Sets the appropriate look-behind assertions on the given state based on |
| 582 | /// this starting configuration. |
| 583 | pub(crate) fn set_lookbehind_from_start( |
| 584 | nfa: &thompson::NFA, |
| 585 | start: &Start, |
| 586 | builder: &mut StateBuilderMatches, |
| 587 | ) { |
| 588 | let rev = nfa.is_reverse(); |
| 589 | let lineterm = nfa.look_matcher().get_line_terminator(); |
| 590 | let lookset = nfa.look_set_any(); |
| 591 | match *start { |
| 592 | Start::NonWordByte => { |
| 593 | if lookset.contains_word() { |
| 594 | builder.set_look_have(|have| { |
| 595 | have.insert(Look::WordStartHalfAscii) |
| 596 | .insert(Look::WordStartHalfUnicode) |
| 597 | }); |
| 598 | } |
| 599 | } |
| 600 | Start::WordByte => { |
| 601 | if lookset.contains_word() { |
| 602 | builder.set_is_from_word(); |
| 603 | } |
| 604 | } |
| 605 | Start::Text => { |
| 606 | if lookset.contains_anchor_haystack() { |
| 607 | builder.set_look_have(|have| have.insert(Look::Start)); |
| 608 | } |
| 609 | if lookset.contains_anchor_line() { |
| 610 | builder.set_look_have(|have| { |
| 611 | have.insert(Look::StartLF).insert(Look::StartCRLF) |
| 612 | }); |
| 613 | } |
| 614 | if lookset.contains_word() { |
| 615 | builder.set_look_have(|have| { |
| 616 | have.insert(Look::WordStartHalfAscii) |
| 617 | .insert(Look::WordStartHalfUnicode) |
| 618 | }); |
| 619 | } |
| 620 | } |
| 621 | Start::LineLF => { |
| 622 | if rev { |
| 623 | if lookset.contains_anchor_crlf() { |
| 624 | builder.set_is_half_crlf(); |
| 625 | } |
| 626 | if lookset.contains_anchor_line() { |
| 627 | builder.set_look_have(|have| have.insert(Look::StartLF)); |
| 628 | } |
| 629 | } else { |
| 630 | if lookset.contains_anchor_line() { |
| 631 | builder.set_look_have(|have| have.insert(Look::StartCRLF)); |
| 632 | } |
| 633 | } |
| 634 | if lookset.contains_anchor_line() && lineterm == b' \n' { |
| 635 | builder.set_look_have(|have| have.insert(Look::StartLF)); |
| 636 | } |
| 637 | if lookset.contains_word() { |
| 638 | builder.set_look_have(|have| { |
| 639 | have.insert(Look::WordStartHalfAscii) |
| 640 | .insert(Look::WordStartHalfUnicode) |
| 641 | }); |
| 642 | } |
| 643 | } |
| 644 | Start::LineCR => { |
| 645 | if lookset.contains_anchor_crlf() { |
| 646 | if rev { |
| 647 | builder.set_look_have(|have| have.insert(Look::StartCRLF)); |
| 648 | } else { |
| 649 | builder.set_is_half_crlf(); |
| 650 | } |
| 651 | } |
| 652 | if lookset.contains_anchor_line() && lineterm == b' \r' { |
| 653 | builder.set_look_have(|have| have.insert(Look::StartLF)); |
| 654 | } |
| 655 | if lookset.contains_word() { |
| 656 | builder.set_look_have(|have| { |
| 657 | have.insert(Look::WordStartHalfAscii) |
| 658 | .insert(Look::WordStartHalfUnicode) |
| 659 | }); |
| 660 | } |
| 661 | } |
| 662 | Start::CustomLineTerminator => { |
| 663 | if lookset.contains_anchor_line() { |
| 664 | builder.set_look_have(|have| have.insert(Look::StartLF)); |
| 665 | } |
| 666 | // This is a bit of a tricky case, but if the line terminator was |
| 667 | // set to a word byte, then we also need to behave as if the start |
| 668 | // configuration is Start::WordByte. That is, we need to mark our |
| 669 | // state as having come from a word byte. |
| 670 | if lookset.contains_word() { |
| 671 | if utf8::is_word_byte(lineterm) { |
| 672 | builder.set_is_from_word(); |
| 673 | } else { |
| 674 | builder.set_look_have(|have| { |
| 675 | have.insert(Look::WordStartHalfAscii) |
| 676 | .insert(Look::WordStartHalfUnicode) |
| 677 | }); |
| 678 | } |
| 679 | } |
| 680 | } |
| 681 | } |
| 682 | } |
| 683 | |