1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384
|
import binascii
import json
import logging
from base64 import b64decode
from contextlib import suppress as noop
from enum import Enum
from typing import Any, List, Tuple, Union
from cloudinit import sources, url_helper, util
from cloudinit.net import find_fallback_nic, get_interfaces_by_mac
from cloudinit.net.ephemeral import EphemeralIPNetwork
from cloudinit.sources.helpers.akamai import (
get_dmi_config,
get_local_instance_id,
is_on_akamai,
)
LOG = logging.getLogger(__name__)
BUILTIN_DS_CONFIG = {
"base_urls": {
"ipv4": "http://169.254.169.254",
"ipv6": "http://[fd00:a9fe:a9fe::1]",
},
"paths": {
"token": "/v1/token",
"metadata": "/v1/instance",
"userdata": "/v1/user-data",
},
# configures the behavior of the datasource
"allow_local_stage": True,
"allow_init_stage": True,
"allow_dhcp": True,
"allow_ipv4": True,
"allow_ipv6": True,
# mac address prefixes for interfaces that we would prefer to use for
# local-stage initialization
"preferred_mac_prefixes": [
"f2:3",
],
}
class MetadataAvailabilityResult(Enum):
"""
Used to indicate how this instance should behave based on the availability
of metadata to it
"""
NOT_AVAILABLE = 0
AVAILABLE = 1
DEFER = 2
class DataSourceAkamai(sources.DataSource):
dsname = "Akamai"
local_stage = False
def __init__(self, sys_cfg, distro, paths):
LOG.debug("Setting up Akamai DataSource")
sources.DataSource.__init__(self, sys_cfg, distro, paths)
self.metadata = dict()
# build our config
self.ds_cfg = util.mergemanydict(
[
get_dmi_config(),
util.get_cfg_by_path(
sys_cfg,
["datasource", "Akamai"],
{},
),
BUILTIN_DS_CONFIG,
],
)
def _build_url(self, path_name: str, use_v6: bool = False) -> str:
"""
Looks up the path for a given name and returns a full url for it. If
use_v6 is passed in, the IPv6 base url is used; otherwise the IPv4 url
is used unless IPv4 is not allowed in ds_cfg
"""
if path_name not in self.ds_cfg["paths"]:
raise ValueError("Unknown path name {}".format(path_name))
version_key = "ipv4"
if use_v6 or not self.ds_cfg["allow_ipv4"]:
version_key = "ipv6"
base_url = self.ds_cfg["base_urls"][version_key]
path = self.ds_cfg["paths"][path_name]
return "{}{}".format(base_url, path)
def _should_fetch_data(self) -> MetadataAvailabilityResult:
"""
Returns whether metadata should be retrieved at this stage, at the next
stage, or never, in the form of a MetadataAvailabilityResult.
"""
if (
not self.ds_cfg["allow_ipv4"] and not self.ds_cfg["allow_ipv6"]
) or (
not self.ds_cfg["allow_local_stage"]
and not self.ds_cfg["allow_init_stage"]
):
# if we're not allowed to fetch data, we shouldn't try
LOG.info("Configuration prohibits fetching metadata.")
return MetadataAvailabilityResult.NOT_AVAILABLE
if self.local_stage:
return self._should_fetch_data_local()
else:
return self._should_fetch_data_network()
def _should_fetch_data_local(self) -> MetadataAvailabilityResult:
"""
Returns whether metadata should be retrieved during the local stage, or
if it should wait for the init stage.
"""
if not self.ds_cfg["allow_local_stage"]:
# if this stage is explicitly disabled, don't attempt to fetch here
LOG.info("Configuration prohibits local stage setup")
return MetadataAvailabilityResult.DEFER
if not self.ds_cfg["allow_dhcp"] and not self.ds_cfg["allow_ipv6"]:
# without dhcp, we can't fetch during the local stage over IPv4.
# If we're not allowed to use IPv6 either, then we can't init
# during this stage
LOG.info(
"Configuration does not allow for ephemeral network setup."
)
return MetadataAvailabilityResult.DEFER
return MetadataAvailabilityResult.AVAILABLE
def _should_fetch_data_network(self) -> MetadataAvailabilityResult:
"""
Returns whether metadata should be fetched during the init stage.
"""
if not self.ds_cfg["allow_init_stage"]:
# if this stage is explicitly disabled, don't attempt to fetch here
LOG.info("Configuration does not allow for init stage setup")
return MetadataAvailabilityResult.DEFER
return MetadataAvailabilityResult.AVAILABLE
def _get_network_context_managers(
self,
) -> List[Tuple[Union[Any, EphemeralIPNetwork], bool]]:
"""
Returns a list of context managers which should be tried when setting
up a network context. If we're running in init mode, this return a
noop since networking should already be configured.
"""
network_context_managers: List[
Tuple[Union[Any, EphemeralIPNetwork], bool]
] = []
if self.local_stage:
# at this stage, networking isn't up yet. To support that, we need
# an ephemeral network
# find the first interface that isn't lo or a vlan interface
interfaces = get_interfaces_by_mac()
interface = None
preferred_prefixes = self.ds_cfg["preferred_mac_prefixes"]
for mac, inf in interfaces.items():
# try to match on the preferred mac prefixes
if any(
[mac.startswith(prefix) for prefix in preferred_prefixes]
):
interface = inf
break
if interface is None:
LOG.warning(
"Failed to find default interface, attempting DHCP on "
"fallback interface"
)
interface = find_fallback_nic()
network_context_managers = []
if self.ds_cfg["allow_ipv6"]:
network_context_managers.append(
(
EphemeralIPNetwork(
self.distro,
interface,
ipv4=False,
ipv6=True,
),
True,
),
)
if self.ds_cfg["allow_ipv4"] and self.ds_cfg["allow_dhcp"]:
network_context_managers.append(
(
EphemeralIPNetwork(
self.distro,
interface,
ipv4=True,
),
False,
)
)
else:
if self.ds_cfg["allow_ipv6"]:
network_context_managers.append(
(
noop(),
True,
),
)
if self.ds_cfg["allow_ipv4"]:
network_context_managers.append(
(
noop(),
False,
),
)
return network_context_managers
def _fetch_metadata(self, use_v6: bool = False) -> bool:
"""
Runs through the sequence of requests necessary to retrieve our
metadata and user data, creating a token for use in doing so, capturing
the results.
"""
try:
# retrieve a token for future requests
token_response = url_helper.readurl(
self._build_url("token", use_v6=use_v6),
request_method="PUT",
timeout=30,
sec_between=2,
retries=4,
headers={
"Metadata-Token-Expiry-Seconds": "300",
},
)
if token_response.code != 200:
LOG.info(
"Fetching token returned %s; not fetching data",
token_response.code,
)
return True
token = str(token_response)
# fetch general metadata
metadata = url_helper.readurl(
self._build_url("metadata", use_v6=use_v6),
timeout=30,
sec_between=2,
retries=2,
headers={
"Accept": "application/json",
"Metadata-Token": token,
},
)
self.metadata = json.loads(str(metadata))
# fetch user data
userdata = url_helper.readurl(
self._build_url("userdata", use_v6=use_v6),
timeout=30,
sec_between=2,
retries=2,
headers={
"Metadata-Token": token,
},
)
self.userdata_raw = str(userdata)
try:
self.userdata_raw = b64decode(self.userdata_raw)
except binascii.Error as e:
LOG.warning("Failed to base64 decode userdata due to %s", e)
except url_helper.UrlError as e:
# we failed to retrieve data with an exception; log the error and
# return false, indicating that we should retry using a different
# network if possible
LOG.warning(
"Failed to retrieve metadata using IPv%s due to %s",
"6" if use_v6 else "4",
e,
)
return False
return True
def _get_data(self) -> bool:
"""
Overrides _get_data in the DataSource class to actually retrieve data
"""
LOG.debug("Getting data from Akamai DataSource")
if not is_on_akamai():
LOG.info("Not running on Akamai, not running.")
return False
local_instance_id = get_local_instance_id()
self.metadata = {
"instance-id": local_instance_id,
}
availability = self._should_fetch_data()
if availability != MetadataAvailabilityResult.AVAILABLE:
if availability == MetadataAvailabilityResult.NOT_AVAILABLE:
LOG.info(
"Metadata is not available, returning local data only."
)
return True
LOG.info(
"Configured not to fetch data at this stage; waiting for "
"a later stage."
)
return False
network_context_managers = self._get_network_context_managers()
for manager, use_v6 in network_context_managers:
with manager:
done = self._fetch_metadata(use_v6=use_v6)
if done:
# fix up some field names
self.metadata["instance-id"] = self.metadata.get(
"id",
local_instance_id,
)
break
else:
# even if we failed to reach the metadata service this loop, we
# still have the locally-available metadata (namely the instance id
# and cloud name), and by accepting just that we ensure that
# cloud-init won't run on our next boot
LOG.warning(
"Failed to contact metadata service, falling back to local "
"metadata only."
)
return True
def check_instance_id(self, sys_cfg) -> bool:
"""
A local-only check to see if the instance id matches the id we see on
the system
"""
return sources.instance_id_matches_system_uuid(
self.get_instance_id(), "system-serial-number"
)
class DataSourceAkamaiLocal(DataSourceAkamai):
"""
A subclass of DataSourceAkamai that runs the same functions, but during the
init-local stage. This allows configuring networking via cloud-init, as
networking hasn't been configured yet.
"""
local_stage = True
datasources = [
# run in init-local if possible
(DataSourceAkamaiLocal, (sources.DEP_FILESYSTEM,)),
# if not, run in init
(
DataSourceAkamai,
(
sources.DEP_FILESYSTEM,
sources.DEP_NETWORK,
),
),
]
# cloudinit/sources/__init__.py will look for and call this when deciding if
# we're a valid DataSource for the stage its running
def get_datasource_list(depends) -> List[sources.DataSource]:
return sources.list_from_depends(depends, datasources)
|