Cloud Native Sustainability Landscape
This document was published on 23.06. We are aware that this document contains gaps that will be addressed in future releases. Contributions are very welcome!
All relevant wordings can be found explained here in the glossary. If you are missing something, feel free to submit a PR to include it.
Summary
Read the Korean translation of this document here.
Cloud computing has revolutionized the way we store and process data, enabling organizations to be more agile, efficient, and scalable. However, as companies transform their business models to meet sustainability requirements, concerns about environmental sustainability in cloud computing have also emerged. The carbon footprint of cloud computing has become a topic of discussion, as it indirectly causes enormous amounts of emissions due to its energy consumption. As a result, it has become imperative to quantify and reduce carbon emissions associated with cloud computing to mitigate the impact on the environment.
Quantifying operational carbon emissions is not as simple as deploying tooling for visibility and accountability. This is particularly true for cloud computing, as there are multiple hardware components enclosed in a server, different generations/architecture/vendors of hardware in cloud Infrastructure, dependencies of the services, services running in virtualized/containerized environments, separate fan/cooling controller in the server, missing data, telemetry & observability, AI/ML workloads, and confidential workloads. These challenges make it difficult to accurately measure carbon emissions associated with cloud computing.
In this white paper, we explore the challenges associated with carbon and energy accounting in cloud computing and provide insights into the complexities of quantifying carbon emissions in public and private clouds. Furthermore, the paper explores sector-specific challenges, such as the telecommunications industry.
Table of Contents
- Table of Contents
- Contributors
- Foundations of Sustainable Cloud Systems
- Challenges of Sustainable Cloud Systems
- Challenges of Carbon/Energy Accounting
- Layers of the solutions
- Current Industry Research and Development
- Current Sustainable Cloud Computing Landscape
- Sustainability Initiatives
Contributors
A special thank you to our contributors of this document. If you are interested in improving and enhancing the content, please file a PR on the repo and ensure you add yourself as a contributor below!
Huamin Chen, Marlow Weston, Niki Manoledaki, Eun Kyung Lee, Chen Wang, Chris Lloyd-Jones, Parul Singh, Przemysław Perycz, Christopher Cantalupo, Patricia Cahill, Jochen Joswig, Emily Fox, Leonard Pahlke
Foundations of Sustainable Cloud Systems
Sustainable cloud and cloud native systems may only be constructed when three foundations exist. These foundations form the basis of informed design, decision-making, and accountability in cloud and cloud native architectures.
Carbon Emissions of the Cloud
The carbon emissions of Information and Communication Technology (short: ICT) systems are categorized as and composed of:
- operational emissions: the amount of carbon emitted during the operational or in-use phase of an ICT system. These emissions are mostly due to burning of fossil fuels to generate the electricity required by these systems.
- embodied emissions: the amount of carbon emitted during the creation and disposal the hardware (short: HW) and physical components of an ICT system (e.g.: devices, servers, cables, buildings, etc.). Embodied emissions are also referred to as embedded emissions.
Cloud and cloud native systems are no exception to the foundation of carbon emissions to ensure environmentally sustainable computing as they are firmly under the umbrella of ICT systems.
Green Computing
Green computing refers to the architecture and design of software and systems that monitor and optimize resource consumption, reduce environmental impact, and improve sustainability, while providing useful services to its users and stakeholders.
Carbon/Energy Accounting
Carbon and energy accounting refers to systems, services, and methodologies to track and account for carbon and energy consumption.
Challenges of Sustainable Cloud Systems
The challenges associated with building and maintaining sustainable cloud systems are still being uncovered with green computing and carbon and energy accounting as two rapidly developing fields with interest from a variety of industry sectors. We anticipate more challenges and solutions to be uncovered as more adoption and maturity of these technologies increases. However, the below section attempts to capture current challenges across the foundations. A popular standard for Carbon Accounting is the Greenhouse Gas Protocol (short: GHG protocol, GHGP).
Challenges of Carbon/Energy Accounting
Quantifying Operational Carbon Emission
Observability in the performance of cloud native workloads has grown in popularity as organizations learn the value in various telemetry data points for increased visibility and understanding of how their architectures are operating. As more organizations begin to consider how to reach similar value in the resource utilization and energy consumption, they’re learning that quantifying the operational carbon emission is not as simple as deploying tooling for visibility and accountability. Quantifying the operational carbon emission is not trivial for a number of reasons, not limited to the following:
- Multiple Hardware (HW) components enclosed in a server - power modeling is required for various HW components (e.g., CPU, Memory, GPU, Storage, I/O) for accurate quantification/estimation.
- HW is used by multiple users/accounts simultaneously – power modeling per different user (e.g., multiple software thread(s)) is a totally different problem for modeling. A important issue to understand here is Energy Proportionality. The SW/HW interaction should be well-understood for power modeling.
- Different generations/architecture/vendors of HW in cloud Infrastructure - power modeling is required for different generations/architecture/vendors for example, Intel vs. AMD vs. ARM, Skylake vs. Sapphire Rapids, and ConnectX-5 vs. ConnectX-6.
- Dependencies of the services - a service may use different services. (e.g., Kubernetes uses COS service), applications may be distributed across data centers and clouds.
- Services running in virtualized/containerized environments - power modeling is required for virtualized/containerized environments, which increases the complexity of modeling
- Separate fan/cooling controller in the server – The fan and other cooling components are often controlled by a separate controller, which requires additional modeling.
- Missing data – due to the limitation of exposing internal data in the cloud, accessing the key data to calculate the operation emission is prohibited. On-premise (On-prem) data centers are sometimes lacking power measurement technology.
- Telemetry & observability – a user often uses multiple HW at the same time, reliable and high-granularity telemetry becomes more important. However, telemetry/observability overhead should be low relative to the services being executed on the server/cloud.
- AI/ML workloads – dramatic increase in using Artificial Intelligence (AI)/ Machine Learning (ML) workload leads to the strong need of dedicated GPU-based clusters. The characteristics of such workloads are different than traditional workloads and their power consumptions are significantly higher.
- Confidential workloads - evolve from VM use case to confidential container (SGX/SEV/TDX), the TEE (Trusted Execution Environment) and the usage of bounce buffer/SWIOTLB might cost more energy. However, the confidential workload is hard to be observed due to security concerns.
Quantifying embedded carbon emissions is also very challenging as manufacturing details (embodied emissions) are not being incorporated into information for holistic quantification by consumers of manufactured technology. This is out of the scope of this white paper, however this TAG encourages interested readers to suggest guidance, best practices, methods, and mechanisms to quantify these emissions by filing an issue or pull request on our repository.
Clouds
Challenges in the Public Cloud
Public cloud providers, such as AWS, Azure, and GCP are often quite restrictive with consumption and emission data, as the providers limit decisions users can make with regard to accessing sustainability-related metrics. Sustainability-related metrics include data points such as the energy consumption, hardware, electricity source, data center PUE, etc.
Providers do try to keep their day-to-day costs, energy usage, and emissions down, but the functionality exposed to users to can be quite limited. This is likely due in part to the shared responsibility model upon which cloud computing is designed - abstracting the operational complexity organizations would otherwise be responsible for in running their own data centers.
Additionally, the quantification challenges previously identified also contribute heavily to further difficulties in accounting for carbon costs by specific users, as the carbon accounting can take much longer than users have to connect to individual types of jobs. The hyperscalers mentioned above offer insight into the carbon emissions of cloud resources through carbon dashboards or APIs. Yet, these can be quite limited and/or have a considerable time lag for the carbon emission data to become available within an acceptable time window for users to take action on. In addition, the methodologies used to calculate carbon emissions can vary between cloud providers, reducing a user’s ability to compare between providers. How this information is measured or estimated is often obtuse, inconsistent, and without industry consensus. As with most emerging technology areas, the variance in underlying implementations will continue until industry centers around a collective schema or framework for both measuring and expressing those measurements.
Users of Cloud Service Providers
Users are often unaware and inexperienced in how their workloads effect their organization’s carbon footprint. Those that do care about their environmental impact have a hard time connecting their individual workloads to their carbon costs.
Challenges in the Private Clouds
These are clouds run by particular companies or universities for use of the members of those companies or universities. These clouds are often more trusting environments, as the users are accountable to the administrators or management of the cloud they are running their workloads on. Due to the special-purposes of private clouds, environmental sustainability, green computing, and accountability of emissions are not in the forefront of design, operation, or even expense, thus contributing to further challenges unique to private cloud. These are still yet unknown.
Sector Specific Challenges
Telco
Telecommunications (Telco) customers are often demanding due to their unique needs in providing high reliability and high velocity service to their customers. Telco needs their systems to be extremely stable, and the traffic to be fast and reliable. Some systems are left entirely at full-power because power modulation can affect traffic and the tolerance for this is low. Building systems that reduce power that Telco is confident will not affect their traffic is challenging.
Finance
Finance may have simulations being run, in the off-hours, and those will look like an AI/ML workload. However, for transactions and fast-traffic, finance has predictable times of day of use-when the markets are up. For this reason, time-of-day adjustments on the majority of the clusters run by finance can be limited. However, transaction times do affect real dollars, so being fast will be prioritized in these environments over power use. In order to get this set of customers to use sustainable options, care must be taken to limit the impact to their bottom line. Additional, high-security and regulatory requirements can increase the utilization and emissions from finance workloads as a result of increased logging, monitoring, and other factors.
Workload Specific Challenges
AI/ML
In addition to the challenges identified above, Artificial Intelligence (AI) and Machine Learning (ML) workloads exist either as schedulable compute at a cloud service provider or within a bare metal environment. In either case, AI/Ml clusters often have the added complexity of XPUs, or accelerators. These accelerators take significant amounts of power to run, more by an order of magnitude required from regular computer chips. Additionally, some of the workloads on these clusters are not time-sensitive, for instance training sets of information, and some are time-sensitive, for instance inference jobs for recognition systems.
Layers of the solutions
When considering solutions complimentary to the three foundations of sustainable cloud systems, we can divide solution considerations into three general areas:
- Which data center to use, if there are multiple options available.
- Where to place the workload once a data center is chosen.
- How to manage the resources on the node allocated for a workload to run on.
All of these elements can be investigated further individually.
Area | Goal | Efforts |
---|---|---|
Multi Data Centers | Intelligently choosing which data center to schedule on according to environmental factors such as whether the region is powered by renewables, the region’s Marginal Emissions Rate, Power Usage Effectiveness (PUE), time of day, etc. | Cluster Management |
Within Data Center | Scheduling effectively according to workload, availability, and urgency of workload | Power Management, K8S Scheduler Plugin |
Within a node | Optimizing resources to handle workload specifications (which may include performance parameters) while minimizing resource consumption | Node Tuning, Pod Scaling |
Current Industry Research and Development
There are a number of developments and ongoing research in the field of sustainable computing. If you know of some that aren’t captured here, we would love for you to contribute them!
Runtime System Power Measurement
📗 A summarization of topics and research up to 2016
Energy Conservation and Carbon Reduction
Tuning, Scaling, and Configuration
At runtime, energy consumed by workloads can be reduced at HW level through DVFS-based scaling, at SW level through runtime parameter tuning and re-configuration, or at the orchestration level through scale-to-zero automation.
Green System Architecture
Green HW/SW systems either improve sub-system efficiency or change the way that computation is conducted.
For instance, programs written in energy efficient languages or running on more optimized runtimes are generally “greener”.
On the other hand, architectures that address the root cause of energy waste, including idle power and data center cooling, are evaluated to be more environmentally friendly. For instance, Federated Learning spreads model training to devices that do not require expensive cooling is evaluated to reduce carbon footprint in aggregate.
Current Sustainable Cloud Computing Landscape
The diagram below illustrates the dimensions of the sustainable cloud computing landscape, which are described in detail in the following sections.
%%{init: {'theme':'neutral'}}%%
flowchart TB
root{{Sustainable Cloud Computing Landscape}} -.- dc[Data centers] & methodologies[Methodologies]
root -..- infra[Infrastructure Tooling] & obs[Observability Tooling]
dc --> smart_dc[Smart Data centers] & cooling[Cooling / BMC]
methodologies --> measurement[Measurement Methodologies] & obs_methodologies[Observability Methodologies]
infra --> scheduling[Scheduling] & scaling[Scaling] & tuning[Resource Tuning]
classDef dimensions fill:#ececff,stroke:#9572db,stroke-width:4px
class dc,methodologies,infra,obs dimensions;
Data centers
Smart Data Centers
- ECO-Qube is a holistic management system that aims to enhance energy efficiency and cooling performance by orchestrating both hardware and software components in edge computing applications ECO-Qube
- Patchwork Kilt - A blueprint for sustainable data centers.
- Open Compute Sustainability Project - Leveraging the expertise of the OCP community, we offer an open framework and resources for OCP members and data center industry – vendors, suppliers, and end users - to deploy industry best practices that promotes reusability and circularity.
Cooling / BMC
- 📰 🧊 OCP Cooling Telemetry
Improve data center cooling facility efficiency through platform power telemetry
Data center operators usually over provision facility capacity to ensure enough buffer to fulfill peak demand. Over provisioning brings great pressure to data centers’ total cost of ownership (TCO). Today, the data center management stack has been widely deployed to monitor data center runtime health status and it gathered tons of data across power, temperature, and resource utilization. These data create opportunities to optimize data center efficiency through data intelligences. In this paper, we introduced our practices in cloud environments for using power trend prediction to improve cooling efficiency. Meanwhile, this paper discussed some key challenges and design considerations while enabling IT platform data-driven facility control at hyperscale data center, e.g. telemetry collection, messaging mechanism, and management API. Effective interoperability among IT devices, facility and management systems is very critical for solution deployment, and the adoption of Open Compute Project design and Redfish API easier system-level integration and reduce deployment costs over different systems and different manufacturers. - 🧊 BMC Telemetry Exposes Baseboard Management Controller data in Prometheus format.
Methodologies
Measurement Methodologies
- Software Carbon Intensity (SCI) Standard - A specification that describes how to calculate the carbon intensity of software applications.
- Green Software Patterns - An online open-source database of software patterns reviewed and curated by the Green Software Foundation across a wide range of categories.
- SCI Guidance - The SCI Guidance project details various approaches on how to understand the different methodologies that are available for calculating energy, carbon intensity, embodied emissions, and functional unit values which are the core components of the SCI calculation.
- Runtime system power consumption estimate Run-time estimation of system and sub-system level power consumption
Observability Methodologies
- 👀 Open Telemetry
High-quality, ubiquitous, and portable telemetry to enable effective observability
A CNCF incubating project designed to create and collect telemetry data from services and software and then forward these to a variety of analysis tools. OpenTelemetry integrates with popular libraries and frameworks such as Spring, ASP.NET Core, Express, Quarkus, and others.
Observability Tooling
The diagram below illustrates the observability dimension of the sustainable cloud computing landscape.
%%{init: {'theme':'neutral'}}%%
flowchart TB
root{{Observability - Sustainable Cloud Computing Landscape}} -.- obs[Observability Tooling]
%% OBSERVABILITY
obs --> g_profiler[gProfiler]
obs --> power_api[PowerAPI\n a Python\n framework]
obs --> kepler[Kepler\n Kubernetes based\n Efficient Power\n Level Exporter]
obs --> scaphandre[Scaphandre\n Multi-platform energy consumption\n metrology agent]
obs --> green_metrics_tool[Green Metrics\n Tool]
obs --> cloud_carbon_footprint[Cloud Carbon\n Footprint]
obs --> influx_telegraf[InfluxData\n Telegraf Collector]
obs --> carbon_ql[Carbon QL]
obs --> powertop[PowerTOP]
click g_profiler "https//docs.gprofiler.io/" "Source"
click power_api "https://github.com/powerapi-ng/" "Source"
click kepler "https://github.com/sustainable-computing-io/kepler" "Source"
click scaphandre "https://github.com/hubblo-org/scaphandre" "Source"
click green_metrics_tool "https://docs.green-coding.berlin/" "Source"
click cloud_carbon_footprint "https://www.cloudcarbonfootprint.org/docs/" "Source"
click influx_telegraf "https://github.com/influxdata/telegraf" "Source"
click carbon_ql "https://github.com/Green-Software-Foundation/carbon-ql" "Source"
click powertop "https://github.com/fenrus75/powertop" "Source"
click tracarbon "https://github.com/fvaleye/tracarbon" "Source"
classDef sections fill:#ececff,stroke:#9572db,stroke-width:4px
class obs sections;
- 👀 gProfiler
OS code profiling tool to visualize applications’ execution sequences and resource usage down to the line of code level
gProfiler, is a free, self-service, and open source, enabling businesses to improve application performance through continuous profiling, thereby reducing costs and minimizing carbon footprint. Granulate users can monitor their carbon emission reduction on the gCenter dashboard, alongside cost and resource reductions, with the CO2 Savings Meter. - 👀 PowerAPI
Python framework for building software-defined power meters
PowerAPI is a middleware toolkit for building software-defined power meters. Software-defined power meters are configurable software libraries that can estimate the power consumption of software in real time. PowerAPI supports the acquisition of raw metrics from a wide diversity of sensors (eg., physical meters, processor interfaces, hardware counters, OS counters) and the delivery of power consumptions via different channels (including file system, network, web, graphical). As a middleware toolkit, PowerAPI offers the capability of assembling power meters «à la carte» to accommodate user requirements. - 👀
Kubernetes-based Efficient Power Level Exporter:
Kepler leverages eBPF programs to probe per-container energy consumption related to system counters and exports them as Prometheus metrics. These metrics help end users monitor their containers’ energy consumption and help cluster administrators make intelligent decisions toward achieving their energy conservation goals. The Kepler Model Server is an internal program that provides Kepler with ML models for estimating power consumption on Kubernetes workloads. The Kepler Model Server pre-trains its models with node energy statistics (labels) and node performance counters (features) as Prometheus metrics on a variety of different Kubernetes clusters and workloads. Once the models achieve an acceptable performance level, Kepler Model Server exports them via flask routes and Kepler can then access them to calculate per-pod energy consumption metrics given per-pod performance counters. Unlike other similar projects, the Kepler Model Server also continuously trains and tunes its pre-trained models using node data scraped by Kepler’s Power Estimate Agents from client clusters. This gives Kepler the ability to further adapt its pod energy consumption calculation capabilities to serve clients’ unique systems. - 👀 Scaphandre
Scaphandre
Scaphandre is a multi-platform monitoring agent, dedicated to power usage / energy consumption metrics and other useful data for reducing ICT software impacts. - 👀 Green Metrics Tool A holistic framework to measure the energy / co2 of your application.
- 👀 InfluxData Telegraf Collector - an open source, plugin-based agent for collecting, processing, aggregating, and writing metrics. Includes multiple input plugins that help determine energy consumption, e.g. intel_powerstat (exposes CPU & DRAM power consumption, CPU temperature, TDP, CPU and uncore frequencies, C-State residencies), ipmi_sensor (exposes IPMI sensor data), redfish (exposes CPU temperature, fan speed, power supply and voltage data as exposed by DMTF Redfish interfaces), and a high number of plugins that help determine the utilization of individual resources that in turn help identifying where the power is consumed. A rich set of available output plugins makes it easy to integrate with various metrics destinations.
- 👀 Carbon QL - The intent of this project is to build a single API codenamed carbonQL that you can use to measure your software emissions for every runtime environment.
- 👀
Cloud Carbon Footprint
This application pulls usage data (compute, storage, networking, etc.) from major cloud providers and calculates estimated energy (Watt-Hours) and greenhouse gas emissions expressed as carbon dioxide equivalents (metric tons CO2e). We display these visualizations in a dashboard for developers, sustainability leaders and other stakeholders in an organization to view and take action. It currently supports AWS, Google Cloud and Microsoft Azure. - 👀 PowerTOP - a Linux tool, which among other things allows you to monitor the power consumption per process running on the Linux machine.
- 📗 OSTI [Paper] Metrics for Evaluating Energy Saving Techniques for Resilient HPC Systems
- 📗 Carbon Aware SDK: The Carbon Aware SDK is a WebApi and Command Line Interface (CLI) to assist in building carbon aware software.
- 👀 Tracarbon: Tracarbon tracks your device’s energy consumption and calculates your carbon emissions using your location using Python.
- 👀 Green Metrics Tool: Modular energy and carbon benchmarking suite with dashboard, API, timeline and comparisons included
- 👀 Eco-CI: Show energy and carbon for CI/CD pipelines on GitHub and GitLab
- 👀 Cloud Energy: ML model to estimate energy and carbon for restricted cloud environments
Infrastructure Tooling
The diagram below illustrates the infrastructure dimension of the sustainable cloud computing landscape.
%%{init: {'theme':'neutral'}}%%
flowchart TB
root{{Infrastructure - Sustainable Cloud Computing Landscape}} -.- scheduling[Scheduling] & scaling[Scaling] & tuning[Resource Tuning]
%% SCHEDULING
scheduling --> plat_aware_scheduling[Platform aware\n scheduling]
scheduling --> intent_driven_orch[Intend driven\n orchestration]
click plat_aware_scheduling "https://github.com/intel/platform-aware-scheduling/tree/master/telemetry-aware-scheduling/docs/power" "Source"
click intent_driven_orch "https://github.com/intel/intent-driven-orchestration" "Source"
%% SCALING
scaling --> predictive_vpa[Predictive Vertical\n Pod Autoscaler]
scaling --> clever[Container Level Energy \n efficient VPA Recommender]
scaling --> keda[Kubernetes Event\n driven Autoscaling]
click predictive_vpa "https://github.com/openshift/predictive-vpa-recommenders" "Source"
click clever "https://github.com/sustainable-computing-io/clever" "Source"
click keda "https://keda.sh/" "Source"
%% RESOURCE TUNING
tuning --> tuned_on_ocp[Node tuning\n via TuneD on OCP]
tuning --> k8s_power_manager[Kubernetes\n Power Manager]
tuning --> geopm[Extensible\n Power Manager]
click tuned_on_ocp "https://docs.openshift.com/container-platform/4.10/scalability_and_performance/using-node-tuning-operator.html" "Source"
click k8s_power_manager "https://github.com/intel/kubernetes-power-manager" "Source"
click geopm "https://geopm.github.io" "Source"
classDef sections fill:#ececff,stroke:#9572db,stroke-width:4px
class scheduling,scaling,tuning sections;
Scheduling At The Cluster Level
At the cluster-level scheduling phase, energy to be consumed by the workload can be reduced by intelligent schedulers that are aware of carbon footprint in a data center, thermal temperature and cooling, caching aware, or server power efficiency. Batch scheduling according to power costs (carbon, money, et cetera).
- 🚆 Power Driven Scheduling and Scaling with CPU telemetry in K8s
Power Driven Scheduling and Scaling with CPU telemetry in Kubernetes
Telemetry Aware Scheduling, a scheduling extension, and the Kubernetes native Horizontal Pod Autoscaler (HPA) are used to enable cluster automation based on real-time information about the current state of power usage on the node. The power metrics used to drive placement and scaling decisions derive from Intel’s Running Average Power Limit (RAPL). collectd is used to gather the metrics and expose them to Prometheus which makes them available inside the cluster using the Prometheus Adapter. - 🚆
Intent Driven Orchestration
This grants a new way to do orchestration by moving from an imperative model to an intent driven model for choosing workload placement. In this model, the user expresses their intents in the form of objectives (e.g. as required latency, throughput, or reliability targets) and the orchestration stack itself determines what resources in the infrastructure are required to fulfill the objectives. This new approach will continue to benefit from community investments in scheduling (determining when & where to place workloads) and be augmented with a continuous running planning loop determining what/how to configure in the system. There is already preliminary work being done to leverage this in a power-optimal usage environment. - 📗 Carbon-aware Kubernetes scheduler A Low Carbon Kubernetes Scheduler
- 📗 Energy aware scheduling [Paper] Improving Data Center Efficiency Through Holistic Scheduling In Kubernetes
Scaling
- 🚤 Predictive VPA Predictive Vertical Pod Autoscaler (VPA) recommenders pluggable with the default VPA on OpenShift
- 🚤 CLEVER
Container Level Energy-efficient VPA Recommender for Kubernetes:
Vertical Pod Autoscalers in Kubernetes allow for automatic CPU and memory request and limit adjustment based on historical resource usage measurements. A VPA deployment has three main components: VPA Recommender, VPA Updater, and VPA Admission Controller. It is possible to replace the default VPA Recommender with a custom Recommender. CLEVER, an intelligent recommender, uses this feature to ensure the QoS or performance of the workloads are not compromised when you try to adjust the CPU frequencies of your cluster. Here’s how it works: assume you have a frequency tuner deployed in your cluster to update the frequency of the CPUs frequencies as per a target metrics or energy consumption budget. Intuitively, when you lower down the frequencies, you do save energy but the performance of workloads also decreases. To counter this you can obtain information like ClusterState and CPU frequencies for the nodes after the frequencies were changed. CLEVER recomputes the new recommendation for CPU requests for pods managed by the VPA based on the updated CPU frequencies. That’s how CLEVER guarantees a similar QoS for a workload by lowering the frequencies to reduce energy but at the same time increasing CPU allocation. - 🚤 KEDA: Kubernetes Event-driven Autoscaling enables scale-to-zero platforms.
On-Node Power Management Tuning
Once the region and node are chosen, administrators and users can further tune the node to minimize the amount of power necessary to run workloads. This can reduce power 30% or more per node.
- 🎵 Node tuning via TuneD on OCP
Manage node-level tuning by orchestrating the tuned daemon
The Node Tuning Operator helps you manage node-level tuning by orchestrating the TuneD daemon. The majority of high-performance applications require some level of kernel tuning. The Node Tuning Operator provides a unified management interface to users of node-level sysctl’s and more flexibility to add custom tuning specified by user needs. - 🎵 Kubernetes Power Manager
Kubernetes Operator designed to expose and utilize Intel specific power management technologies in a Kubernetes Environment
The allocation of CPU resources from a pool of platforms in a container orchestration engine like Kubernetes (K8s) is exclusively based on availability. In order to expose and use power management technologies in a Kubernetes context, the Kubernetes Power Manager is a Kubernetes operator created using the Operator SDK. The Kubernetes Power Manager makes use of a powerful set of power management technologies that give users more precise control over CPU performance and power usage on a per-core basis. Yet, Kubernetes is purposefully built to operate as an abstraction layer between the workload and such hardware capabilities as a workload orchestrator. Users of Kubernetes who are running performance-critical applications with particular requirements that depend on hardware capabilities have a hurdle as a result of this. By enabling the user to adjust the frequencies and determine the priority level of the cores selected by the Kubernetes Native CPU Manager, the Kubernetes Power Manager fills the gap between the hardware feature enablement and the container orchestration layer. It has been proven to work with TuneD as well to allow TuneD profiles to bem used to control the power on the nodes according to various frequency tunings. - 🎵 GEOPM
Extensible Power Manager:
Initially specific to HPC environments, but now more generalized, the Global Extensible Open Power Manager (GEOPM) is a framework for exploring power and energy optimizations on heterogeneous platforms.
The GEOPM software is split into two packages: The GEOPM Service and the GEOPM Runtime. The GEOPM Service provides user-space access to low-level hardware metrics and configuration knobs. The GEOPM Runtime leverages the GEOPM Service to tune hardware settings in reaction to hardware metrics and application feedback. The application feedback is collected through lightweight asynchronous profiling hooks injected with callbacks into middle-ware packages.
The GEOPM Runtime has a plugin architecture for selecting between optimization algorithms. Some of the built-in algorithms target energy efficiency, and others optimize performance within a power bound. The port of GEOPM to Kubernetes is ongoing. There is an experimental branch calledcloud
with implementations of new features that support Kubernetes. These features will be migrated into the maindev
branch as they each become production ready.
Additional documentation may be found in the service readme file and in the runtime k8 directory.
Sustainability Initiatives
There are a number of sustainability initiatives ongoing, if we’ve missed one please contribute to this list by filing a pull request!
Organizations
- 🐝 Green Software Foundation
Building a trusted ecosystem of people, standards, tooling and best practices for green software
The Green Software Foundation (GSF) exists to change how we build software, so there are zero harmful environmental effects, a foundation with over 42 member organizations. Key pillars are Knowledge, Tech Culture, and Tooling; which are delivered through a standards working group, an open source working group, a community working group, and a policy working group.
The GSF has created a software carbon intensity (SCI) standard, which has been submitted to ISO (International Standards Organisation) for ratification, to ensure we measure carbon consistently. This standard is being implemented in code through the Carbon Aware SDK (a tool to do more when the energy grid is green, and less when it is dirty), the Carbon Pipeline (measuring carbon in the CI/CD process, and CarbonQL - a standardized API for measuring carbon according to the SCI standard. - 🐝
LF Energy
LF Energy is an open source foundation focused on the power systems sector, hosted within The Linux Foundation. LF Energy provides a neutral, collaborative community to build the shared digital investments that will transform the world’s relationship to energy. This organization contains the repositories for the core LF Energy Foundation and many of the hosted projects and working groups. Their landscape can be found here. - 🐝 Energy Efficient High Performance Computing Working Group
Encourages implementation of energy conservation measures, energy efficient design in high performance computing (HPC)
Mission is to encourage the implementation of energy conservation measures, energy efficient design in high performance computing (HPC), and share ideas. Can find an extensive collection of papers here that can be extrapolated from in terms of patterns to be lifted into the cloud native landscape. - 🐝
Green Software Training
This initiative will teach you how to build, maintain and run greener applications irrespective of the application domain, industry, organization size or type, programming language, or framework; leading to a Green Software Certification backed by the Linux Foundation. - 🐝
Cloud Carbon Footprint
Get to know the carbon footprint of your cloud usage - and reduce it.
Cloud Carbon Footprint is an open source tool that provides visibility and tooling to measure, monitor and reduce your cloud carbon emissions. We use best practice methodologies to convert cloud utilization into estimated energy usage and carbon emissions, producing metrics and carbon savings estimates that can be shared with employees, investors, and other stakeholders. - 🐝
Open Compute Project
Almost 100% of the energy used in a processor turns into heat. Up until very recently this has been a liability and a challenge: an immense quantity of heat to be removed which requires a significant additional energy expenditure. With the proper design, data center cooling systems can be converted into a heat source and a heat management opportunity that in turn converts the heat from a liability to an asset. The OCP Heat Reuse subgroup explores these challenges and opportunities. Its goal is to suggest solutions to enable implementations of technologies designed to harness the heat to turn cost into profit.```
Conferences
- 🐝 Linux Foundation’s SustainabilityCon The first sustainability-focused track by the Linux Foundation
- 🐝 EnviroInfo: EnviroInfo 2023 is the 37th edition of the long-standing and well-established international and interdisciplinary conference series on leading environmental information and communication technologies.
Carbon Emissions Reports
- 📄 IEA Emissions - Global Energy and CO2 Status Report 2019
- 📄 European Environment Agency EU Greenhouse Emission Intensity
- 📄 electricityMap’s real-time CO2 emission data
- SCI Reporting - Creating the infrastructure, and processes to store, host, and publicly report SCI scores, and other related reporting requirements within the SCI specification.
- 📄 WattTime API Provides insight into a electricity grid’s marginal emissions rate
Net Zero / Carbon Neutrality
- 🥬 The Climate Pledge Net-Zero Carbon by 2040
- 🥬 WeTransfer WeTransfer becomes Climate Neutral
- 🥬 Adrian Cockroft, ex-Amazon VP of Sustainability Architecture “Cloud computing pioneer’s new focus is on sustainability transformation”
- 🥬 Supercritical Helping businesses achieve net zero
Programming Language Efficiency Analysis
- 🔌 Energy Efficiency of Languages The complete set of tools for energy consumption analysis of programming languages, using Computer Language Benchmark Game
Feedback
Was this page helpful?
Glad to hear it! Please tell us how we can improve.
Sorry to hear that. Please tell us how we can improve.