Server IP : 66.29.132.122 / Your IP : 18.118.162.128 Web Server : LiteSpeed System : Linux business142.web-hosting.com 4.18.0-553.lve.el8.x86_64 #1 SMP Mon May 27 15:27:34 UTC 2024 x86_64 User : admazpex ( 531) PHP Version : 7.2.34 Disable Function : NONE MySQL : OFF | cURL : ON | WGET : ON | Perl : ON | Python : ON | Sudo : OFF | Pkexec : OFF Directory : /proc/self/root/proc/thread-self/root/proc/thread-self/root/opt/alt/ruby33/share/rubygems/rubygems/request_set/lockfile/ |
Upload File : |
# frozen_string_literal: true # ) frozen_string_literal: true require_relative "parser" class Gem::RequestSet::Lockfile::Tokenizer Token = Struct.new :type, :value, :column, :line EOF = Token.new :EOF def self.from_file(file) new File.read(file), file end def initialize(input, filename = nil, line = 0, pos = 0) @line = line @line_pos = pos @tokens = [] @filename = filename tokenize input end def make_parser(set, platforms) Gem::RequestSet::Lockfile::Parser.new self, set, platforms, @filename end def to_a @tokens.map {|token| [token.type, token.value, token.column, token.line] } end def skip(type) @tokens.shift while !@tokens.empty? && peek.type == type end ## # Calculates the column (by byte) and the line of the current token based on # +byte_offset+. def token_pos(byte_offset) # :nodoc: [byte_offset - @line_pos, @line] end def empty? @tokens.empty? end def unshift(token) @tokens.unshift token end def next_token @tokens.shift end alias_method :shift, :next_token def peek @tokens.first || EOF end private def tokenize(input) require "strscan" s = StringScanner.new input until s.eos? do pos = s.pos pos = s.pos if leading_whitespace = s.scan(/ +/) if s.scan(/[<|=>]{7}/) message = "your #{@filename} contains merge conflict markers" column, line = token_pos pos raise Gem::RequestSet::Lockfile::ParseError.new message, column, line, @filename end @tokens << if s.scan(/\r?\n/) token = Token.new(:newline, nil, *token_pos(pos)) @line_pos = s.pos @line += 1 token elsif s.scan(/[A-Z]+/) if leading_whitespace text = s.matched text += s.scan(/[^\s)]*/).to_s # in case of no match Token.new(:text, text, *token_pos(pos)) else Token.new(:section, s.matched, *token_pos(pos)) end elsif s.scan(/([a-z]+):\s/) s.pos -= 1 # rewind for possible newline Token.new(:entry, s[1], *token_pos(pos)) elsif s.scan(/\(/) Token.new(:l_paren, nil, *token_pos(pos)) elsif s.scan(/\)/) Token.new(:r_paren, nil, *token_pos(pos)) elsif s.scan(/<=|>=|=|~>|<|>|!=/) Token.new(:requirement, s.matched, *token_pos(pos)) elsif s.scan(/,/) Token.new(:comma, nil, *token_pos(pos)) elsif s.scan(/!/) Token.new(:bang, nil, *token_pos(pos)) elsif s.scan(/[^\s),!]*/) Token.new(:text, s.matched, *token_pos(pos)) else raise "BUG: can't create token for: #{s.string[s.pos..-1].inspect}" end end @tokens end end