hc99's picture
Add files using upload-large-folder tool
b40c49c verified
raw
history blame
10.6 kB
"""
Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
SPDX-License-Identifier: MIT-0
# Code is taken from jsonschema package and adapted CloudFormation use
# https://github.com/yaml/pyyaml/blob/a2d19c0234866dc9d4d55abf3009699c258bb72f/lib/yaml/scanner.py#L46
"""
import fileinput
import logging
import sys
from yaml import MappingNode, ScalarNode, SequenceNode
from yaml.composer import Composer
from yaml.constructor import ConstructorError, SafeConstructor
from yaml.reader import Reader
from yaml.resolver import Resolver
from yaml.scanner import Scanner
from cfnlint.decode.mark import Mark
from cfnlint.decode.node import dict_node, list_node, str_node
from cfnlint.rules import Match
from cfnlint.rules.errors import ParseError
try:
from yaml._yaml import CParser as Parser # pylint: disable=ungrouped-imports,
cyaml = True
except ImportError:
from yaml.parser import Parser # type: ignore # pylint: disable=ungrouped-imports
cyaml = False
UNCONVERTED_SUFFIXES = ["Ref", "Condition"]
FN_PREFIX = "Fn::"
LOGGER = logging.getLogger(__name__)
class CfnParseError(ConstructorError):
"""
Error thrown when the template contains Cfn Error
"""
def __init__(self, filename, errors):
if isinstance(errors, Match):
errors = [errors]
# Call the base class constructor with the parameters it needs
super().__init__(errors[0].message)
# Now for your custom code...
self.filename = filename
self.matches = errors
def build_match(filename, message, line_number, column_number, key):
return Match.create(
message=message,
filename=filename,
rule=ParseError(),
linenumber=line_number + 1,
columnnumber=column_number + 1,
linenumberend=line_number + 1,
columnnumberend=column_number + 1 + len(key),
)
class NodeConstructor(SafeConstructor):
"""
Node Constructors for loading different types in Yaml
"""
def __init__(self, filename):
# Call the base class constructor
super().__init__()
self.filename = filename
# To support lazy loading, the original constructors first yield
# an empty object, then fill them in when iterated. Due to
# laziness we omit this behaviour (and will only do "deep
# construction") by first exhausting iterators, then yielding
# copies.
def construct_yaml_map(self, node):
# Check for duplicate keys on the current level, this is not desirable
# because a dict does not support this. It overwrites it with the last
# occurance, which can give unexpected results
mapping = {}
self.flatten_mapping(node)
matches = []
for key_node, value_node in node.value:
key = self.construct_object(key_node, False)
value = self.construct_object(value_node, False)
if key is None:
raise CfnParseError(
self.filename,
[
build_match(
filename=self.filename,
message=(
f"Null key {key_node.value!r} not supported "
f"(line {key_node.start_mark.line + 1})"
),
line_number=key_node.start_mark.line,
column_number=key_node.start_mark.column,
key=key_node.value,
),
],
)
for key_dup in mapping:
if key_dup == key:
if not matches:
matches.extend(
[
build_match(
filename=self.filename,
message=(
f'Duplicate found "{key}" (line'
f" {key_dup.start_mark.line + 1})"
),
line_number=key_dup.start_mark.line,
column_number=key_dup.start_mark.column,
key=key,
),
build_match(
filename=self.filename,
message=(
f'Duplicate found "{key}" (line'
f" {key_node.start_mark.line + 1})"
),
line_number=key_node.start_mark.line,
column_number=key_node.start_mark.column,
key=key,
),
],
)
else:
matches.append(
build_match(
filename=self.filename,
message=(
f'Duplicate found "{key}" (line'
f" {key_node.start_mark.line + 1})"
),
line_number=key_node.start_mark.line,
column_number=key_node.start_mark.column,
key=key,
),
)
try:
mapping[key] = value
except Exception as exc:
raise CfnParseError(
self.filename,
[
build_match(
filename=self.filename,
message=(
f'Unhashable type "{key}" (line'
f" {key.start_mark.line + 1})"
),
line_number=key.start_mark.line,
column_number=key.start_mark.column,
key=key,
),
],
) from exc
if matches:
raise CfnParseError(
self.filename,
matches,
)
(obj,) = SafeConstructor.construct_yaml_map(self, node)
return dict_node(obj, node.start_mark, node.end_mark)
def construct_yaml_str(self, node):
obj = SafeConstructor.construct_yaml_str(self, node)
assert isinstance(obj, (str))
return str_node(obj, node.start_mark, node.end_mark)
def construct_yaml_seq(self, node):
(obj,) = SafeConstructor.construct_yaml_seq(self, node)
assert isinstance(obj, list)
return list_node(obj, node.start_mark, node.end_mark)
NodeConstructor.add_constructor( # type: ignore
"tag:yaml.org,2002:map", NodeConstructor.construct_yaml_map
)
NodeConstructor.add_constructor( # type: ignore
"tag:yaml.org,2002:str", NodeConstructor.construct_yaml_str
)
NodeConstructor.add_constructor( # type: ignore
"tag:yaml.org,2002:seq", NodeConstructor.construct_yaml_seq
)
class _Scanner(Scanner):
def __init__(self) -> None:
super().__init__()
self.ESCAPE_REPLACEMENTS = {
"0": "\0",
"a": "\x07",
"b": "\x08",
"t": "\x09",
"\t": "\x09",
"n": "\x0A",
"v": "\x0B",
"f": "\x0C",
"r": "\x0D",
"e": "\x1B",
" ": "\x20",
'"': '"',
"\\": "\\",
"N": "\x85",
"_": "\xA0",
"L": "\u2028",
"P": "\u2029",
}
# pylint: disable=too-many-ancestors
class MarkedLoader(Reader, _Scanner, Parser, Composer, NodeConstructor, Resolver):
"""
Class for marked loading YAML
"""
# pylint: disable=non-parent-init-called,super-init-not-called
def __init__(self, stream, filename):
Reader.__init__(self, stream)
_Scanner.__init__(self)
if cyaml:
Parser.__init__(self, stream)
else:
Parser.__init__(self)
Composer.__init__(self)
SafeConstructor.__init__(self)
Resolver.__init__(self)
NodeConstructor.__init__(self, filename)
def construct_getatt(self, node):
"""
Reconstruct !GetAtt into a list
"""
if isinstance(node.value, (str)):
return list_node(node.value.split(".", 1), node.start_mark, node.end_mark)
if isinstance(node.value, list):
return [self.construct_object(child, deep=False) for child in node.value]
raise ValueError(f"Unexpected node type: {type(node.value)}")
def multi_constructor(loader, tag_suffix, node):
"""
Deal with !Ref style function format
"""
if tag_suffix not in UNCONVERTED_SUFFIXES:
tag_suffix = f"{FN_PREFIX}{tag_suffix}"
constructor = None
if tag_suffix == "Fn::GetAtt":
constructor = loader.construct_getatt
elif isinstance(node, ScalarNode):
constructor = loader.construct_scalar
elif isinstance(node, SequenceNode):
constructor = loader.construct_sequence
elif isinstance(node, MappingNode):
constructor = loader.construct_mapping
else:
raise f"Bad tag: !{tag_suffix}"
return dict_node({tag_suffix: constructor(node)}, node.start_mark, node.end_mark)
def loads(yaml_string, fname=None):
"""
Load the given YAML string
"""
loader = MarkedLoader(yaml_string, fname)
loader.add_multi_constructor("!", multi_constructor)
template = loader.get_single_data()
# Convert an empty file to an empty dict
if template is None:
template = dict_node({}, Mark(0, 0), Mark(0, 0))
return template
def load(filename):
"""
Load the given YAML file
"""
content = ""
if (filename is None) and (not sys.stdin.isatty()):
filename = "-" # no filename provided, it's stdin
fileinput_args = {"files": filename}
if sys.version_info.major <= 3 and sys.version_info.minor >= 10:
fileinput_args["encoding"] = "utf-8"
with fileinput.input(**fileinput_args) as f:
for line in f:
content = content + line
else:
with open(filename, encoding="utf-8") as fp:
content = fp.read()
return loads(content, filename)