summaryrefslogtreecommitdiffstats
path: root/lib/libc/regex/regcomp.c
Commit message (Collapse)AuthorAgeFilesLines
* Replace various spelling with FALLTHROUGH which is lint()ablecharnier2002-08-251-1/+1
|
* Fix the style of the SCM ID's.obrien2002-03-221-2/+2
| | | | I believe have made all of libc .c's as consistent as possible.
* Remove __P() usage.obrien2002-03-211-43/+43
|
* Remove 'register' keyword.obrien2002-03-211-203/+203
|
* The algorithm that computes the tables used in the BM search algorithm sometimesdcs2001-11-091-1/+2
| | | | | | | | | | access an array beyond it's length. This only happens in the last iteration of a loop, and the value fetched is not used then, so the bug is a relatively innocent one. Fix this by not fetching any value on the last iteration of said loop. Submitted by: MKI <mki@mozone.net> MFC after: 1 week
* altoffset() always returned whenever it recursed, because at the enddcs2000-07-091-0/+4
| | | | | | | | | | | | | | of the processing of the recursion, "scan" would be pointing to O_CH (or O_QUEST), which would then be interpreted as being the end character for altoffset(). We avoid this by properly increasing scan before leaving the switch. Without this, something like (a?b?)?cc would result in a g->moffset of 1 instead of 2. I added a case to the soon-to-be-imported regex(3) test code to catch this error.
* Add some casts here and there.dcs2000-07-091-3/+3
|
* Do not free NULL pointers.dcs2000-07-071-1/+1
|
* Deal with the signed/unsigned chars issue in a more proper manner. Wedcs2000-07-071-3/+5
| | | | | | | | | use a CHAR_MIN-based array, like elsewhere in the code. Remove a number of unused variables (some due to the above change, one that was left after a number of optimizing steps through the source). Brucified by: bde
* I hate signed chars.^W^W^W^W^WCast to unsigned char before using signeddcs2000-07-061-1/+1
| | | | chars as array indices.
* Correct comment to work with test code.dcs2000-07-061-2/+2
| | | | Prevent out of bounds array access in some specific cases.
* Use UCHAR_MAX consistently.dcs2000-07-061-1/+1
|
* Enhance the optimization provided by pre-matching. Fix style bugs withdcs2000-07-021-5/+164
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | previous commits. At the time we search the pattern for the "must" string, we now compute the longest offset from the beginning of the pattern at which the must string might be found. If that offset is found to be infinite (through use of "+" or "*"), we set it to -1 to disable the heuristics applied later. After we are done with pre-matching, we use that offset and the point in the text at which the must string was found to compute the earliest point at which the pattern might be found. Special care should be taken here. The variable "start" is passed to the automata-processing functions fast() and slow() to indicate the point in the text at which they should start working from. The real beginning of the text is passed in a struct match variable m, which is used to check for anchors. That variable, though, is initialized with "start", so we must not adjust "start" before "m" is properly initialized. Simple tests showed a speed increase from 100% to 400%, but they were biased in that regexec() was called for the whole file instead of line by line, and parenthized subexpressions were not searched for. This change adds a single integer to the size of the "guts" structure, and does not change the ABI. Further improvements possible: Since the speed increase observed here is so huge, one intuitive optimization would be to introduce a bias in the function that computes the "must" string so as to prefer a smaller string with a finite offset over a larger one with an infinite offset. Tests have shown this to be a bad idea, though, as the cost of false pre-matches far outweights the benefits of a must offset, even in biased situations. A number of other improvements suggest themselves, though: * identify the cases where the pattern is identical to the must string, and avoid entering fast() and slow() in these cases. * compute the maximum offset from the must string to the end of the pattern, and use that to set the point at which fast() and slow() should give up trying to find a match, and return then return to pre-matching. * return all the way to pre-matching if a "match" was found and later invalidated by back reference processing. Since back references are evil and should be avoided anyway, this is of little use.
* Initialize variables used by the Boyer-Moore algorithm.dcs2000-06-291-0/+2
| | | | | | | This should fix core dumps when the must pattern is of length three or less. Bug found by: knu
* Add Boyler-Moore algorithm to pre-matching test.dcs2000-06-291-0/+144
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | The BM algorithm works by scanning the pattern from right to left, and jumping as many characters as viable based on the text's mismatched character and the pattern's already matched suffix. This typically enable us to test only a fraction of the text's characters, but has a worse performance than the straight-forward method for small patterns. Because of this, the BM algorithm will only be used if the pattern size is at least 4 characters. Notice that this pre-matching is done on the largest substring of the regular expression that _must_ be present on the text for a succesful match to be possible at all. For instance, "(xyzzy|grues)" will yield a null "must" substring, and, therefore, not benefit from the BM algorithm at all. Because of the lack of intelligence of the algorithm that finds the "must" string, things like "charjump|matchjump" will also yield a null string. To optimize that, "(char|match)jump" should be used. The setup time (at regcomp()) for the BM algorithm will most likely outweight any benefits for one-time matches. Given the slow regex(3) we have, this is unlikely to be even perceptible, though. The size of a regex_t structure is increased by 2*sizeof(char*) + 256*sizeof(int) + strlen(must)*sizeof(int). This is all inside the regex_t's "guts", which is allocated dynamically by regcomp(). If allocation of either of the two tables fail, the other one is freed. In this case, the straight-forward algorithm is used for pre-matching. Tests exercising the code path affected have shown a speed increase of 50% for "must" strings of length four or five. API and ABI remain unchanged by this commit. The patch submitted on the PR was not used, as it was non-functional. PR: 14342
* unsigned char cleanupache1999-07-261-17/+17
| | | | | | | fix wrong index from p_simp_re() PR: 8790 Submitted by: Alexander Viro <viro@math.psu.edu> (partially)
* Replace memory leaking instances of realloc with non-leaking reallocf.imp1998-09-161-4/+4
| | | | | | | | | | | In some cases replace if (a == null) a = malloc(x); else a = realloc(a, x); with simple reallocf(a, x). Per ANSI-C, this is guaranteed to be the same thing. I've been running these on my system here w/o ill effects for some time. However, the CTM-express is at part 6 of 34 for the CAM changes, so I've not been able to do a build world with the CAM in the tree with these changes. Shouldn't impact anything, but...
* Speedup in case locale not usedache1997-04-041-5/+11
|
* collate_range_cmp -> __collate_range_cmpache1996-10-311-4/+5
|
* Convert to newly aded collate compare functionache1996-08-121-4/+4
|
* Remove static collcmp, use new internal function nowache1996-08-121-37/+4
|
* Use collate data for national alpha character ranges like [a-z]ache1996-08-111-4/+45
|
* Short value is better for hash due to easy overflow in 8bit charactersache1996-08-111-1/+1
|
* Use locale for character classes instead of hardcoded valuesache1996-08-111-6/+66
| | | | Misc 8bit cleanup
* General -Wall warning cleanup, part I.jkh1996-07-121-28/+32
| | | | Submitted-By: Kent Vander Velden <graphix@iastate.edu>
* 8bit clean fixesache1996-03-251-3/+5
|
* More cleanup.phk1995-10-221-2/+0
| | | | Uhm, I also forgot: I took "EXTRA_SANITY" out of malloc.c
* BSD 4.4 Lite Lib Sourcesrgrimes1994-05-271-0/+1698
OpenPOWER on IntegriCloud