]>
git.ipfire.org Git - thirdparty/squid.git/blob - src/parser/Tokenizer.cc
2 #include "parser/Tokenizer.h"
12 /* Native 64 bit system without strtoll() */
13 #if defined(LONG_MIN) && (SIZEOF_LONG == 8)
14 #define INT64_MIN LONG_MIN
17 #define INT64_MIN (-9223372036854775807LL-1LL)
22 /* Native 64 bit system without strtoll() */
23 #if defined(LONG_MAX) && (SIZEOF_LONG == 8)
24 #define INT64_MAX LONG_MAX
27 #define INT64_MAX 9223372036854775807LL
31 /// convenience method: consumes up to n bytes, counts, and returns them
33 Parser::Tokenizer::consume(const SBuf::size_type n
)
35 // careful: n may be npos!
36 const SBuf result
= buf_
.consume(n
);
37 parsed_
+= result
.length();
41 /// convenience method: consume()s up to n bytes and returns their count
43 Parser::Tokenizer::success(const SBuf::size_type n
)
45 return consume(n
).length();
49 Parser::Tokenizer::token(SBuf
&returnedToken
, const CharacterSet
&delimiters
)
51 const Tokenizer
saved(*this);
53 const SBuf::size_type tokenLen
= buf_
.findFirstOf(delimiters
); // not found = npos => consume to end
54 if (tokenLen
== SBuf::npos
) {
58 returnedToken
= consume(tokenLen
); // cannot be empty
64 Parser::Tokenizer::prefix(SBuf
&returnedToken
, const CharacterSet
&tokenChars
, const SBuf::size_type limit
)
66 const SBuf::size_type prefixLen
= buf_
.substr(0,limit
).findFirstNotOf(tokenChars
);
69 if (prefixLen
== SBuf::npos
&& (atEnd() || limit
== 0))
71 returnedToken
= consume(prefixLen
); // cannot be empty after the npos check
76 Parser::Tokenizer::skipAll(const CharacterSet
&tokenChars
)
78 const SBuf::size_type prefixLen
= buf_
.findFirstNotOf(tokenChars
);
81 return success(prefixLen
);
85 Parser::Tokenizer::skipOne(const CharacterSet
&chars
)
87 if (!buf_
.isEmpty() && chars
[buf_
[0]])
93 Parser::Tokenizer::skip(const SBuf
&tokenToSkip
)
95 if (buf_
.startsWith(tokenToSkip
))
96 return success(tokenToSkip
.length());
101 Parser::Tokenizer::skip(const char tokenChar
)
103 if (!buf_
.isEmpty() && buf_
[0] == tokenChar
)
108 /* reworked from compat/strtoll.c */
110 Parser::Tokenizer::int64(int64_t & result
, int base
)
115 //fixme: account for buf_.size()
117 const char *s
= buf_
.rawContent();
118 const char *end
= buf_
.rawContent() + buf_
.length();
123 } else if (*s
== '+') {
126 if (s
>= end
) return false;
127 if (( base
== 0 || base
== 16) && *s
== '0' && (s
+1 <= end
) &&
128 tolower(*(s
+1)) == 'x') {
140 if (s
>= end
) return false;
144 cutoff
= neg
? -static_cast<uint64_t>(INT64_MIN
) : INT64_MAX
;
145 const int cutlim
= cutoff
% static_cast<int64_t>(base
);
146 cutoff
/= static_cast<uint64_t>(base
);
150 for (c
= *s
++; s
<= end
; c
= *s
++) {
153 } else if (xisalpha(c
)) {
154 c
-= xisupper(c
) ? 'A' - 10 : 'a' - 10;
160 if (any
< 0 || static_cast<uint64_t>(acc
) > cutoff
|| (static_cast<uint64_t>(acc
) == cutoff
&& c
> cutlim
))
169 if (any
== 0) // nothing was parsed
172 acc
= neg
? INT64_MIN
: INT64_MAX
;
179 return success(s
- buf_
.rawContent() - 1);