Lines Matching refs:tokens

652                 tokens = self.end_line_tokens
653 tokens.unget_all()
665 tokens = self._tokenize(line, False, line_feeder.filename,
668 t0 = tokens.get_next()
679 sym = tokens.get_next()
694 kconfig_file = tokens.get_next()
718 dep_expr = self._parse_expr(tokens, None, line,
732 comment.text = tokens.get_next()
746 menu.title = tokens.get_next()
763 name = tokens.get_next()
807 text = tokens.get_next()
823 def parse_val_and_cond(tokens, line, filename, linenr): argument
827 return (self._parse_expr(tokens, stmt, line, filename, linenr,
829 self._parse_expr(tokens, stmt, line, filename, linenr)
830 if tokens.check(T_IF) else None)
852 tokens = self._tokenize(line, False, filename, linenr)
854 t0 = tokens.get_next()
862 if not tokens.check(T_ON):
866 parsed_deps = self._parse_expr(tokens, stmt, line, filename,
906 target = tokens.get_next()
913 self._parse_expr(tokens, stmt, line, filename, linenr)
914 if tokens.check(T_IF) else None))
917 target = tokens.get_next()
924 self._parse_expr(tokens, stmt, line, filename, linenr)
925 if tokens.check(T_IF) else None))
929 if tokens.peek_next() is not None:
930 new_prompt = parse_val_and_cond(tokens, line, filename,
934 new_def_exprs.append(parse_val_and_cond(tokens, line, filename,
939 if tokens.peek_next() is not None:
940 new_def_exprs.append(parse_val_and_cond(tokens, line,
948 new_prompt = parse_val_and_cond(tokens, line, filename, linenr)
951 low = tokens.get_next()
952 high = tokens.get_next()
958 self._parse_expr(tokens, stmt, line, filename, linenr)
959 if tokens.check(T_IF) else None))
963 if tokens.peek_next() is not None:
964 new_def_exprs.append(parse_val_and_cond(tokens, line,
968 if tokens.check(T_ENV) and tokens.check(T_EQUAL):
969 env_var = tokens.get_next()
991 elif tokens.check(T_DEFCONFIG_LIST):
994 elif tokens.check(T_MODULES):
1011 elif tokens.check(T_ALLNOCONFIG_Y):
1023 if not tokens.check(T_IF):
1031 parsed_deps = self._parse_expr(tokens, stmt, line, filename,
1047 self.end_line_tokens = tokens
1233 tokens = []
1256 tokens = [keyword]
1263 append = tokens.append
1372 previous = tokens[-1]
1374 return _Feed(tokens)