Lines Matching refs:i
1216 i = 0 # The current index in the string being tokenized
1240 i = initial_token_match.end()
1248 while i < strlen:
1251 id_keyword_match = _id_keyword_re_match(s, i)
1257 i = id_keyword_match.end()
1288 while i < strlen and s[i].isspace():
1289 i += 1
1290 if i == strlen:
1292 c = s[i]
1293 i += 1
1303 if i >= len(s):
1305 c = s[i]
1309 if i + 1 >= len(s):
1311 val += s[i + 1]
1312 i += 2
1315 i += 1
1316 i += 1
1322 end = s.find(c, i)
1325 append(s[i:end])
1326 i = end + 1
1330 if i >= len(s) or s[i] != "&": continue
1332 i += 1
1336 if i >= len(s) or s[i] != "|": continue
1338 i += 1
1341 if i < len(s) and s[i] == "=":
1343 i += 1
3090 self.i = 0
3093 if self.i >= self.length:
3095 item = self.items[self.i]
3096 self.i += 1
3100 return None if self.i >= self.length else self.items[self.i]
3105 if self.i < self.length and self.items[self.i] == token:
3106 self.i += 1
3111 self.i = 0