Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
ARTIFICIAL INTELLIGENCE BASED POWER CONSUMPTION OPTIMIZATION
Document Type and Number:
WIPO Patent Application WO/2023/075853
Kind Code:
A1
Abstract:
An optimization apparatus that receives data related to operational characteristics of a plurality of devices in a network, classifies the plurality of devices in the network into a plurality of clusters based on the data, builds a plurality of artificial intelligence (AI) models, each of the AI models corresponding to one of the plurality of clusters, determines a predicted operational characteristic for a first device based on an AI model, among the AI models, corresponding to a cluster to which the first device belongs, and outputs a recommendation for the first device based on the predicted operational characteristics.

Inventors:
KESAVAN KRISHNAKUMAR (JP)
DORIA ALEXANDER (JP)
SUTHAR MANISH (JP)
Application Number:
PCT/US2022/028474
Publication Date:
May 04, 2023
Filing Date:
May 10, 2022
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
RAKUTEN MOBILE INC (JP)
RAKUTEN MOBILE USA LLC (US)
International Classes:
G06N20/00; G06F9/50; G06Q10/06; G06N3/00
Foreign References:
US20200204628A12020-06-25
US20210109584A12021-04-15
Attorney, Agent or Firm:
KIBLAWI, Fadi N. et al. (US)
Download PDF:
Claims:
What is claimed is:

1. An apparatus comprising: a memory storing one or more instructions; and a processor configured to execute the one or more instructions to: receive data related to operational characteristics of a plurality of devices in a network, classify the plurality of devices in the network into a plurality of clusters based on the data, build a plurality of artificial intelligence (Al) models, each of the Al models corresponding to one of the plurality of clusters, determine a predicted operational characteristic for a first device based on an Al model, among the Al models, corresponding to a cluster to which the first device belongs, and output a recommendation for the first device based on the predicted operational characteristics.

2. The apparatus of claim 1, wherein the processor is further configured to execute a clustering algorithm classify the plurality of devices in the network into the plurality of clusters.

3. The apparatus of claim 1, wherein each of the plurality of Al models are tailored to one of the plurality of clusters.

4. The apparatus of claim 1, wherein the processor is further configured to control an operation parameter of a CPU of the first device based on the predicted operational characteristic.

33

5. The apparatus of claim 1, wherein the processor is further configured to set a clock frequency of a CPU of the first device based on the predicted operational characteristic.

6. The apparatus of claim 1, wherein the data comprises at least one of historical data including one of server parameters, metrics or key performance indicators.

7. The apparatus of claim 1, wherein the processor is further configured to classify the plurality of devices in the network into the plurality of clusters based on one or more patterns identified in the data.

8. The apparatus of claim 7, wherein the one or more patterns may be workload signature information, kernel statistics information, traffic pattern information, time information or location information.

9. A method comprising: receiving data related to operational characteristics of a plurality of devices in a network; classifying the plurality of devices in the network into a plurality of clusters based on the data; building a plurality of artificial intelligence (Al) models, each of the Al models corresponding to one of the plurality of clusters; determining a predicted operational characteristic for a first device based on an Al model, among the Al models, corresponding to a cluster to which the first device belongs; and

34 outputting a recommendation for the first device based on the predicted operational characteristics.

10. The method of claim 9, further comprising executing a clustering algorithm classify the plurality of devices in the network into the plurality of clusters.

11. The method of claim 9, wherein each of the plurality of Al models are tailored to one of the plurality of clusters.

12. The method of claim 9, further comprising controlling an operation parameter of a CPU of the first device based on the predicted operational characteristic.

13. The method of claim 9, further comprising setting a clock frequency of a CPU of the first device based on the predicted operational characteristic.

14. The method of claim 9, wherein the data comprises at least one of historical data including one of server parameters, metrics or key performance indicators.

15. The method of claim 9, further classifying the plurality of devices in the network into the plurality of clusters based on one or more patterns identified in the data.

16. The method of claim 15, wherein the one or more patterns may be workload signature information, kernel statistics information, traffic pattern information, time information or location information.

Description:
ARTIFICIAL INTELLIGENCE BASED POWER CONSUMPTION OPTIMIZATION

[Technical Field]

[0001] The disclosure relates to an optimization apparatus, an optimization system, an optimization method, and a storage medium. More particularly, it relates to an optimization apparatus, an optimization system, an optimization method, and a storage medium for optimizing power consumption based on artificial intelligence. However, the disclosure is not limited to optimizing power consumption. For instance, one or more aspects of the disclosure may be applied in optimization of other features in an electronic device or a system.

[Related Art]

[0002] In large networks, such as communication networks, numerous servers and/or devices may consume large amounts of power. This power consumption not only affects the functioning of the servers and the devices, but it also increases the cost for operating and maintaining the servers and devices.

[0003] Accordingly, there is a need for optimizing the power consumption of the servers and devices, particularly in large networks.

[Summary]

[0004] In a related art technology, one approach is to build a single model for power optimization for all servers. However, such an approach is not very ideal, since implementing a single model for all the servers does not take into account the differences between the features and functionalities of all the servers. According to another approach, an individual model may be built separately for each server. However, such an approach would not scalable. In some other cases, a rule based approach has be implemented, in which, rule-based algorithms (i.e. , “put server X to sleep during midnight of every day”). However, such an approach is cumbersome and is not efficient. [0005] As such, there is a need for an improved manner of optimizing one or more aspects of servers provided in large networks.

[0006] According to an aspect of the disclosure, there are provided apparatuses, methods and systems for implementing scalable, efficient and lightweight Al models to optimize server operation characteristics such as power consumption.

[0007] According to an aspect of the disclosure, there is provided an apparatus comprising: a memory storing one or more instructions; and a processor configured to execute the one or more instructions to: receive data related to operational characteristics of a plurality of devices in a network, classify the plurality of devices in the network into a plurality of clusters based on the data, build a plurality of artificial intelligence (Al) models, each of the Al models corresponding to one of the plurality of clusters, determine a predicted operational characteristic for a first device based on an Al model, among the Al models, corresponding to a cluster to which the first device belongs, and output a recommendation to operation based on the predict operation characteristics for the first device based on the predicted operational characteristics.

[0008] The processor is further configured to execute a clustering algorithm classify the plurality of devices in the network into the plurality of clusters.

[0009] Each of the plurality of Al models are tailored to one of the plurality of clusters.

[0010] The processor is further configured to control an operation parameter of a CPU of the first device based on the predicted operational characteristic.

[0011] The processor is further configured to set a clock frequency of a CPU of the first device based on the predicted operational characteristic.

[0012] The data comprises at least one of historical data including one of server parameters, metrics or key performance indicators.

[0013] The processor is further configured to classify the plurality of devices in the network into the plurality of clusters based on one or more patterns identified in the data. [0014] The one or more paterns may be workload signature information, kernel statistics information, traffic patern information, time information or location information.

[0015] According to another aspect of the disclosure, there is provided a method comprising: receiving data related to operational characteristics of a plurality of devices in a network; classifying the plurality of devices in the network into a plurality of clusters based on the data; building a plurality of artificial intelligence (Al) models, each of the Al models corresponding to one of the plurality of clusters; determining a predicted operational characteristic for a first device based on an Al model, among the Al models, corresponding to a cluster to which the first device belongs; and outputing a recommendation for the first device based on the predicted operational characteristics.

[0016] The method further comprising executing a clustering algorithm classify the plurality of devices in the network into the plurality of clusters.

[0017] Each of the plurality of Al models are tailored to one of the plurality of clusters.

[0018] The method further comprising controlling an operation parameter of a CPU of the first device based on the predicted operational characteristic.

[0019] The method further comprising seting a clock frequency of a CPU of the first device based on the predicted operational characteristic.

[0020] The data comprises at least one of historical data including one of server parameters, metrics or key performance indicators.

[0021] The method further comprising classifying the plurality of devices in the network into the plurality of clusters based on one or more paterns identified in the data.

[0022] The one or more paterns may be workload signature information, kernel statistics information, traffic patern information, time information or location information.

[Brief Description of the Drawings] [0023] The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.

[0024] These and/or other aspects will become apparent and more readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings in which:

[0025] FIG. 1A illustrates a network including a plurality of servers according to an example embodiment of the disclosure;

[0026] FIG. IB illustrates a detailed diagram of a server including according to an example embodiment of the disclosure;

[0027] FIG. 2A illustrates an apparatus according to an example embodiment of the disclosure;

[0028] FIG. 2B illustrates a connection between an apparatus and a plurality of servers according to another example embodiment of the disclosure;

[0029] FIG. 2C illustrates a detailed diagram of an apparatus according to an example embodiment of the disclosure;

[0030] FIG. 3 is a chart illustrating clusters of servers according to an example embodiment of the disclosure;

[0031] FIG. 4 illustrates operating states of the servers according to an example embodiment;

[0032] FIG. 5 illustrates a method of optimization according to an example embodiment of the disclosure;

[0033] FIG. 6 illustrates a process flow according to an example embodiment of the disclosure; and

[0034] FIGS.7 and 8 are graphs illustrating a level of accuracy of the prediction according to example embodiments. [Description of Example Embodiments]

[0035] Example embodiments will now be described below in more detail with reference to the accompanying drawings. The following detailed descriptions are provided to assist the reader in gaining a comprehensive understanding of the methods, apparatuses, and/or systems described herein. However, the example embodiment provided in the disclosure should not be considered as limiting the scope of the disclosure. Accordingly, various changes, modifications, and equivalents of the systems, apparatuses and/or methods described herein will be suggested to those of ordinary skill in the art.

[0036] The terms used in the description are intended to describe embodiments only, and shall by no means be restrictive. Unless clearly used otherwise, expressions in a singular form include a meaning of a plural form. In the present description, an expression such as “including” is intended to designate a characteristic, a number, a step, an operation, an element, a part or combinations thereof, and shall not be construed to preclude any presence or possibility of one or more other characteristics, numbers, steps, operations, elements, parts or combinations thereof.

[0037] One or more example embodiments of the disclosure will be described below with reference to the drawings. Throughout the drawings, the same components or corresponding components are labeled with the same reference numerals, and, accordingly, the description thereof may be omitted or simplified.

[0038] FIG. 1A illustrates a network 1 including a plurality of servers 101. According to an example embodiment, the network 1 may be a communication network for facilitating communication between the plurality of servers 101. For instance, the network 1 may be a large network serving millions of electronic devices, such as user equipment (UE). As an example, the network 1 may be part of a cellular radio system or an internet service provider system in a large metropolitan area, which uses hundreds of servers transmission of information or data. Although a plurality of servers are illustrated in FIG. 1A, the disclosure is not limited thereto, and as such, according to another example embodiment, the network may include telecommunication devices, such as base stations, or other electronic devices such as servers, computers, mobile devices etc.,

[0039] According to an example embodiment, the plurality of servers in the network may be located at different geographical regions. For instance, as illustrated in FIG. 1 A, servers 101 A, may be located at location A, servers 10 I B, may be located at location B, and servers 101 C, may be located at location C. According to an example embodiment, locations A, B and C may be physical locations. However, the disclosure is not limited thereto, and as such, according to another example embodiment, the plurality of servers 101 may be cloud-based virtual machines (VMs).

[0040] FIG. IB illustrates the cloud of servers including, among many servers, server 101 1, server 101_2 and server 101 3. Internal representative hardware of a servers 101 1, 101_2 and 101_3 are illustrated. Each of these servers 101 1 , 101_2 and 101 3 may include a CPU, and the CPU may include a plurality of cores. For instance, the CPU may include core 1, core 2, core 3, ... core n (where is an integer). Each core of the CPU can perform operations separately from the other cores. Or, multiple cores of the CPU may work together to perform parallel operations on a shared set of data in the CPU's memory cache (e.g., a portion of memory). According to an example embodiment, the server 101 1 may have, for example, 80 cores. However, the disclosure is not limited thereto, and as such, different number of cores may be provided. The server 101 1 may also include one or more fans which provide airflow, FPGA chips, and interrupt hardware. The components illustrated in FIG. IB are exemplary, and as such, other servers of the disclosure may add other components and/or or omit one or more of the components illustrated in FIG. IB.

[0041] Since network 1 employs large numbers of servers 101, there is a need for optimizing power consumption of the servers 101. However, related art power optimization systems fail to provide a scalable, efficient and lightweight system optimize server power consumption. According to an example embodiment, there is provided a scalable, efficient and lightweight system, implemented by artificial intelligence (Al) models, to optimize server power consumption. For instance, according to an example embodiment, Al models are generated by taking into account differences in features and functionalities between the servers 101. For instance, a servers 101 A at location A may have one or more first characteristics different from one or more second characteristics of a servers 101 B at location B. Therefore, the operation and the power consumption characteristics may vary. However, the disclosure is not limited thereto, and as such, according to another example embodiment, there may be characteristic differences between the servers 101 A at location A. For example, the servers 101 different workloads running different protocols. As such, the operation and the power consumption characteristics may vary between the servers 101 A at location A.

[0042] According to an example embodiment, an optimization apparatus performs a clustering operation to capture the patterns across multiple servers 101, across different geographical regions and/or multiple workloads running different protocols based on their workload signatures, time of the day patterns, network traffic patterns, kernel statistics, etc. Based on the captured patterns, the optimization apparatus clusters the multiple servers 101 according to the captured patterns. Thereafter, the optimization apparatus builds an Al model for each cluster of servers to take advantage of patterns that are specific to each cluster. Accordingly, a plurality of Al models are deployed, each of the Al models corresponding to each of the respective servers in each of the respective clusters, such that, a same Al model is used for each sever in a respective cluster. For instance, a first Al model corresponding to a first cluster is deployed with respect to a first server in the first cluster and a second Al model corresponding to a second cluster is deployed with respect to a second server in the second cluster. [0043] According to an example embodiment, the Al models may predict one or more future characteristics of the servers 101. For instance, the first Al model may predict one or more characteristics of one or more servers in the first cluster in the future, and the second Al model may predict one or more characteristics of one or more servers in the second cluster in the future. According to an example embodiment, one or more characteristics may be traffic on each of the servers over a period of time in the future. According to an example embodiment, one or more characteristics may be traffic on each core of the servers. For instance, the first Al model may predict the traffic on each core of the one or more servers in the first cluster over the next ten minutes. However, the disclosure is not limited thereto, and as such, according to other example embodiments, one or more characteristics may be different from the traffic and the period of time may be different from ten minutes. For instance, according to another example embodiment, the one or more characteristics may be a processing load on each core of the one or more servers in the future. According to an example embodiment, the core of the server may be a Central Processing Unit (CPU) of the server. However, the disclosure is not limited thereto, and as such, one or more characteristics other types processors, or other electronic circuitry may be predicted.

[0044] According to an example embodiment, the optimization apparatus may output setting information corresponding to one or more features and/or functionalities of the servers based on the predicted one or more characteristics of the servers. For instance, the setting information may correspond to a CPU frequency based on the predicted one or more characteristics of the servers.

[0045] However, the disclosure is not limited thereto, and as such, according to another example embodiment, the setting information may indicate a state of one or more servers based on the predicted one or more characteristics of the servers. For example, the setting information may indicate an operation state of the servers. According to an example embodiment, the setting information may indicate that the one or more servers operate in a certain state, among a plurality of operation states. The operation state indicated in the setting information being determined based on the predicted one or more characteristics of the servers. According to an example embodiment, the operation state may be related to the processing frequency of the CPU. For instance, the operation states may be a first state, in which, the CPU frequency is set to 2.6 GHz, a second state, in which, the CPU frequency is set to or 2 GHZ or a third state, in which, the CPU frequency is set to 1.6 GHz. However, the disclosure is not limited thereto, and the operation states may be related to other features or functionalities of the servers.

[0046] According to an example embodiment, the optimization apparatus may control one or more servers based on the predicted one or more characteristics of the servers. For instance, optimization apparatus may output instruction to control the core of the one or more servers to operate at a specific frequency. According to another example embodiment, the optimization apparatus may output a recommendation to operate the one or more servers in a particular manner based on the predicted one or more characteristics of the servers.

[0047] FIG. 2A illustrates an apparatus 200 according to an example embodiment of the disclosure. The apparatus 200 may be configured to build scalable, efficient and lightweight Al models to manage, control and/or optimize one or more servers 100 of the network 1. According to an example embodiment, the apparatus 200 may include a processor 210, a memory 220, a storage 230 and a communication interface 240. However, the disclosure is not limited to the arrangement of components illustrated in FIG. 2 A. For instance, according to another example embodiment, according to an example embodiment, the apparatus may further include a display, a input/output (I/O) interface, or a bus line that connects the components of the apparatus 200. As such, according to another example embodiment, the other components or may be included in the apparatus 200 or omitted from the apparatus 200. [0048] According to an example embodiment, the processor 210 may be CPU, a graphic processing unit (GPU) or other processing circuity. According to an example embodiment, the memory 220 may include a random access memory (RAM) or other types of memory. According to an example embodiment, the storage 230 may be formed of a storage medium such as a non-volatile memory, a hard disk drive, or the like and functions as a storage unit. According to an example embodiment, the communication interface 240 may include a transceiver configured to transmit and receive data from one or more devices external to the apparatus 200. According to an example embodiment, the communication interface 240 may include electronic components and/or circuitry to perform wireless communication with the one or more external devices.

[0049] According to an example embodiment, the storage 230 stores a program for performing one or more operations to build Al models to manage, control and/or optimize one or more servers 100 of the network 1. According to an example embodiment, the program may include one or more instructions or computer codes. According to an example embodiment, the processor 210 may function as a control unit that operates by executing the program stored in the storage 230.

[0050] Moreover, according to an example embodiment, the processor 230 may execute the one or more instructions or computer codes to implement one or more modules to build Al models to manage, control and/or optimize one or more servers 100 of the network 1. According to an example embodiment, the processor 210 may control the operation of the apparatus 210. According to an example embodiment, the memory 220 may provide a memory field necessary for the operation of the processor 210. According to an example embodiment, the communication interface 240 may be connected to other devices, such as servers 101, in the network 1. According to an example embodiment, data may be transmitted or received from other devices in the network through the communication interface 240.

[0051] According to an example embodiment, the processor 210 may receive data from one or more servers 101 in the network 1. According to an example embodiment, the processor 210 may receive the data from a management server, which has collected the data about the one or more servers 101 in the network 1. According to another example embodiment, the processor 210 may receive and collect the data directly from the one or more servers 101 in the network 1. According to an example embodiment, the data may be relate to a characteristics of the one or more servers 101. For instance, the data may be server parameters related to the hardware components of servers 101 or the functionalities of the server 101. In some example embodiments, the server parameter includes a field programmable gate array (FPGA) parameter, a CPU parameter, a memory parameter, and/or an interrupt parameter.

In some embodiments, the FPGA parameter is message queue, the CPU parameter is load and/or processes, the memory parameter is IRQ (interrupt request) or DISKIO (disk input/output operations), and the interrupt parameter is IPMI (intelligent Platform Management Interface) and/or IOWAIT (i.e., idle time).

[0052] The server parameters may include the following parameter show in Table 1 below. _ _

[0053] However, the disclosure is not limited to the server parameters listed above. For instance, according to another example of the disclosure, the data may include other parameter, metrics or performance indicators. For instance, the data may include key performance indicators (KPI). As such, processor may receive large number of data points.

[0054] According to an example embodiment, the processor 210 may perform a clustering operation on the data. For instance, the processor 210 may apply a clustering algorithm on the data to identify patterns across multiple servers 101. According to an example embodiment, the clustering algorithm may implement machine learning to group data points in the data into similar clusters based on features of the data points. For instance, the processor 210 may cluster the servers 101 operating across different geographical regions and/or performing multiple workloads running different protocols based on their workload signatures, time of the day patterns, network traffic patterns, kernel statistics, etc.

[0055] Referring to FIG. 3, the processor 210 may cluster the data points into clusters C1-C8 based on features of the data points. For instance, each of the clusters C1-C8 may include a group of servers 101. According to an example embodiment, each dot inside a cluster may represent a server having certain pattern that is same or similar to other servers in the cluster. As shown in FIG. 3, cluster Cl may include a plurality of first servers that have same workload signatures or similar workload signatures. However, the disclosure is not limited thereto, and as such, each clusters C2-C8 may include a plurality of servers that have same respective patterns or similar respective patterns. For instance, a server satisfying a specific criteria or threshold with respect to a particular pattern of a cluster may be considers as part of the cluster.

[0056] According to an example embodiment, the processor 210 may build an Al model for each cluster of servers to take advantage of patterns that are specific to each cluster. For instance, the processor 210 may build a first Al model (Al Model 1) corresponding to a first cluster Cl. In particular, the processor 210 may build the first Al model (Al Model 1) corresponding to the servers in the first cluster Cl. Also, the processor 210 may build the second Al model (Al Model 2) corresponding to the servers in the second cluster C2. Each of the Al models, such as Al Model 1 and Al Model 2, are built or trained using test data. The test data may be historical data collected from the servers.

[0057] According to an example embodiment, the model training may be performed by: (1) loading data for training (i.e., historical data for servers); (2) setting targets based on a condition of the servers (obtain labels by labelling nodes based on the condition using the data), (3) computing statistical features of the data, and adding the statistical features to the data object, (4) identifying leading indicators for the condition, this identification is based on the data and the labels, (5) training an Al model with the leading indicators, the data, and the labels, and (6) optimizing the Al model by performing hyperparameter tuning and model validation. The output from operations (l)-(6) may be optimize the Al model by performing hyperparameter tuning and model validation (some of the historical data has been used for training, some has been reserved for testing at this stage). The output of the above approach is the Al model. According to another example embodiment, the training of the Al model may be performed by unlabeled data.

[0058] According to an example embodiment, in operation (2), the targets may be set based on the clusters. For instance, the model may be trained by taking into account the specific patterns identified for the servers in each of the clusters, such that the trained Al models are tailored for each cluster. For instance, the Al model for cluster Cl may be trained by setting the targets based on a workload signature. However, the disclosure is not limited thereto, and as such, other patterns, such as time of the day patterns, network traffic patterns, kernel statistics etc., may be used as targets for training the model.

[0059] According to an example embodiment, the processor 210 may deploy a plurality of Al models. Accordingly, each of the Al models corresponding to each of the respective servers in each of the respective clusters may be deployed, such that, a same Al model is used for each sever in a respective cluster. For instance, a first Al model (Al Model 1) corresponding to a first cluster Cl may be deployed with respect to a first server SI in the first cluster Cl. Also, a second Al model (Al Model 2) corresponding to a second cluster C2 may be deployed with respect to a second server S2 in the second cluster C2. Moreover, the first Al model is deployed for all the servers in cluster Cl, and the second Al model is deployed for all the servers in cluster C2. Also, the processor 210 may build a third Al model corresponding to servers in cluster C3, a fourth Al model corresponding to servers in cluster C4, a fifth Al model corresponding to servers in cluster C5, a sixth Al model corresponding to servers in cluster C6, a seventh Al model corresponding to servers in cluster C7, and an eight Al model corresponding to servers in cluster C8. However, the disclosure is not limited to the clusters in FIG. 3 and the Al models corresponding to the clusters. As such, according to another example embodiment, different number of clusters and Al models may be provided.

[0060] According to an example embodiment, the Al models may predict one or more future characteristics of the servers 101. For instance, the first Al model may predict one or more characteristics of one or more servers in the first cluster Cl. Also, the second Al model may predict one or more characteristics of one or more servers in the second cluster C2. According to an example embodiment, one or more characteristics may be traffic on each of the servers over a period of time in the future. According to an example embodiment, one or more characteristics may be traffic on each core of the servers. For instance, the first Al model may predict the traffic on each core of the one or more servers in the first cluster over the next ten minutes. However, the disclosure is not limited thereto, and as such, according to other example embodiments, one or more characteristics may be different from the traffic and the period of time may be different from ten minutes.

[0061] According to an example embodiment, the optimization apparatus may output setting information corresponding to one or more features and/or functionalities of the servers based on the predicted one or more characteristics of the servers. For instance, the setting information may correspond to a CPU frequency based on the predicted one or more characteristics of the servers. However, the disclosure is not limited thereto, and as such, according to another example embodiment, the setting information may indicate a state of one or more servers based on the predicted one or more characteristics of the servers. For example, the setting information may indicate an operation state of the servers.

[0062] According to an example embodiment, the setting information may indicate that the one or more servers operate in a certain state, among a plurality of operation states. The operation state indicated in the setting information being determined based on the predicted one or more characteristics of the servers. According to an example embodiment, the operation state may be related to the processing frequency of the CPU. For instance, the operation states may be a first state, in which, the CPU frequency is set to 2.6 GHz, a second state, in which, the CPU frequency is set to or 2 GHZ or a third state, in which, the CPU frequency is set to 1.6 GHz. However, the disclosure is not limited thereto, and the operation states may be related to other features or functionalities of the servers. [0063] According to an example embodiment, the optimization apparatus may control one or more servers based on the predicted one or more characteristics of the servers. For instance, optimization apparatus may output instruction to control the core of the one or more servers to operate at a specific frequency. According to another example embodiment, the optimization apparatus may output a recommendation to operate the one or more servers in a particular manner based on the predicted one or more characteristics of the servers.

[0064] FIG. 4 illustrates operating states of the servers according to an example embodiment. For instance, row 1 may correspond to servers in the first cluster Cl, row 2 may correspond to servers in the second cluster C2, row 3 may correspond to servers in the third cluster C3 and row 4 may correspond to servers in the fourth cluster C4. According to an example embodiment, the current state of all the servers in all the clusters may be P0. According to an example embodiment, state P0 may represent a CPU frequency of 2.6 GHz or a maximum frequency. According to an example embodiment, based on a predicted using the Al model described in the disclosure, the servers in the first cluster Cl may have a recommended state of CO, which is a normal operating state.

[0065] According to an example embodiment, based on a predicted using the Al model described in the disclosure, the servers in the second cluster C2 may have a recommended state, in which, the servers operate at state P2 eighty percent (80%) of the time and operate at state P0 twenty percent (20%) of the time. Here, P2 may represent a CPU frequency of 1.6 GHz.

[0066] According to an example embodiment, based on a predicted using the Al model described in the disclosure, the servers in the third cluster C3 may have a recommended state, in which, the servers operate at state Pl fifty percent (50%) of the time and operate at state P0 fifty percent (50%) of the time. Here, Pl may represent a CPU frequency of 2 GHz.

[0067] According to an example embodiment, based on a predicted using the Al model described in the disclosure, the servers in the fourth cluster C4 may have a recommended state, in which, the servers operate at state P2 twenty five percent (25%) of the time, operate at state Pl twenty five percent (25%) of the time and operate at state P0 fifty percent (50%) of the time. [0068] Although FIG. 9 illustrates four recommended states, the disclosure is not limited thereto, and as such according to another example embodiment, other recommended states may be determined and output. According to an example embodiment, the servers may be controlled to operate based on the recommended states.

[0069] FIG. 2B illustrates an example embodiment of an apparatus 200 connected to a plurality of servers in network. According to an example embodiment, the optimization apparatus 200 may be connected to the servers 101 1 , 101_2 and 101 3 through a management server. For example, the management server may be an edge node of the servers. According to an example embodiment of the disclosure, the optimization apparatus 200 may transmit the setting information for the servers 101 1, 101_2 and 101 3 to the management server based on the predicted one or more characteristics of the servers using the Al models.

[0070] FIG. 2C illustrates a detailed diagram of an apparatus 200 according to an example embodiment. In FIG. 2C, the apparatus 200 may include the same components illustrated in FIG. 2A. However, the diagram of the apparatus 200 in FIG. 2C may further illustrate the modules implemented by the processor 210. According to an example embodiment, the processor 210 may execute one or more instructions (or program codes) to implement a clustering module 211, a model builder 212, a predictor 213 and an output module 214.

[0071] According to an example embodiment, the clustering module 211 may classify the plurality of devices in the network into a plurality of clusters based on the data. According to an example embodiment, the clustering module 211 may capture the patterns across multiple servers, and cluster the plurality of servers based on the captured patterns. According to an example embodiment, the classification operation may be performed using machine learning. [0072] According to an example embodiment, the model builder 212 may build a plurality of artificial intelligence (Al) models, each of the Al models corresponding to one of the plurality of clusters. For instance, an Al model may be built respectively for each cluster of servers to take advantage of patterns that are specific to each cluster.

[0073] According to an example embodiment, the predictor 213 may deploy a plurality of Al models and determine a predicted operational characteristic for a first device based on the deployed Al model. That is, the predictor 213 may deploy the Al models to predict one or more future characteristics of one or more of the servers. According to an example embodiment, one of the characteristics may be traffic on each of the servers over a period of time in the future. According to an example embodiment, one or more characteristics may be traffic on each core of the servers. However, the disclosure is not limited thereto, and as such, according to other example embodiments, other characteristics such as a processing load on the servers or the memory usage of the servers.

[0074] According to an example embodiment, the output module 214 may output a recommendation for the first device based on the predicted operational characteristics. The output setting information may correspond to one or more features and/or functionalities of the servers based on the predicted one or more characteristics of the servers. For instance, the setting information may correspond to a CPU frequency based on the predicted one or more characteristics of the servers.

[0075] According to an example embodiment, the apparatus 200 illustrated in FIGS.

2A, 2B and 2C may be an operating console computer, which further include a display and a user interface.

[0076] FIG. 5 illustrates a flow chart of operations in an optimization method according to an example embodiment. The operations illustrated in FIG. 5 may be performed one or more processor. For instance, the operations illustrated in FIG. 5 may be performed by a single processor or by two or more processors working in combination. [0077] According to an example embodiment, the method includes receiving data related to operational characteristics of a plurality of devices a network (SI 10). For instance, the data may be received from one or more servers in a network. According to an example embodiment, the data may be relate to a characteristics of the one or more servers, i.e., server parameters related to the hardware components of servers or the functionalities of the server.

[0078] According to an example embodiment, the method includes classifying the plurality of devices in the network into a plurality of clusters based on the data (S120). According to an example embodiment, the classifying operation may be a clustering operation to capture the patterns across multiple servers, across different geographical regions, and/or multiple workloads running different protocols based on their workload signatures, time of the day patterns, network traffic patterns, kernel statistics, etc. Accordingly, the plurality of servers are clustered based on the captured patterns. According to an example embodiment, the classification operation may be performed using machine learning.

[0079] According to an example embodiment, the method includes building a plurality of artificial intelligence (Al) models, each of the Al models corresponding to one of the plurality of clusters (S130). For instance, an Al model may be built respectively for each cluster of servers to take advantage of patterns that are specific to each cluster. Accordingly, a plurality of Al models are deployed, each of the Al models corresponding to each of the respective servers in each of the respective clusters, such that, a same Al model is used for each sever in a respective cluster.

[0080] According to an example embodiment, the method includes determining a predicted operational characteristic for a first device based on an Al model corresponding to a cluster to which the first device belongs (S140). That is, the Al models may predict one or more future characteristics of one or more of the servers. According to an example embodiment, one of the characteristics may be traffic on each of the servers over a period of time in the future. According to an example embodiment, one or more characteristics may be traffic on each core of the servers. However, the disclosure is not limited thereto, and as such, according to other example embodiments, other characteristics such as a processing load on the servers or the memory usage of the servers.

[0081] According to an example embodiment, the method includes outputting a recommendation for the first device based on the predicted operational characteristics (SI 50). The output setting information may correspond to one or more features and/or functionalities of the servers based on the predicted one or more characteristics of the servers. For instance, the setting information may correspond to a CPU frequency based on the predicted one or more characteristics of the servers.

[0082] According to an example embodiment, the setting information may indicate a state of one or more servers based on the predicted one or more characteristics of the servers. For example, the setting information may indicate an operation state of the servers being determined based on the predicted one or more characteristics of the servers. According to an example embodiment, the operation states may be a first state, in which, the CPU frequency is set to 2.6 GHz, a second state, in which, the CPU frequency is set to or 2 GHZ or a third state, in which, the CPU frequency is set to 1.6 GHz. However, the disclosure is not limited thereto, and the operation states may be related to other features or functionalities of the servers.

[0083] According to an example embodiment, the method may include transmitting a control signal to one or more servers based on the predicted one or more characteristics of the servers. For instance, the method may include outputting instructions to control the core of the one or more servers to operate at a certain frequency on the predicted one or more characteristics of the servers. According to another example embodiment, the method may include output instructions to control the one or more servers to operate at an increased or a reduced speed. According to another example embodiment, the method may include output instructions to control the one or more servers to operate using less resources. However, the disclosure is not limited thereto, and as such, according to another example embodiment, other output or control setting are possible based on frequency on the predicted one or more characteristics of the servers.

[0084] FIG. 6 illustrates a process flow according to an example embodiment of the disclosure. According to an example embodiment, the optimization apparatus receive data from telegraf server and/or foresight (5G/LTE) servers. According to an example embodiment, the data may include 2 billion data points made of 535 metrics and/or 200 key performance indicators (KPIs). However, the disclosure is not limited thereto, and as such, different amount of data may be received and processed by the optimization apparatus.

[0085] According to an example embodiment, the optimization apparatus classifies the plurality of servers in the network into a plurality of clusters based on the data. Based on the plurality of clusters, the optimization apparatus builds a plurality of artificial intelligence (Al) models, each of the Al models corresponding to one of the plurality of clusters. The optimization apparatus may predict future CPU load based on the Al models and recommend CPU frequency based on the predicted future CPU load. The recommend CPU frequency may be one or a combination of the following states: CO, P0, Pl, and P2. However, the disclosure is not limited thereto, and as such, other states are possible.

[0086] According to an example embodiment, the method includes determining a predicted operational characteristic for a first device based on an Al model corresponding to a cluster to which the first device belongs (S140). That is, the Al models may predict one or more future characteristics of one or more of the servers. According to an example embodiment, one of the characteristics may be traffic on each of the servers over a period of time in the future. According to an example embodiment, one or more characteristics may be traffic on each core of the servers. However, the disclosure is not limited thereto, and as such, according to other example embodiments, other characteristics such as a processing load on the servers or the memory usage of the servers. [0087] According to an example embodiment, the method includes outputting a recommendation for the first device based on the predicted operational characteristics (SI 50). The output setting information may correspond to one or more features and/or functionalities of the servers based on the predicted one or more characteristics of the servers. For instance, the setting information may correspond to a CPU frequency based on the predicted one or more characteristics of the servers.

[0088] FIGS.7 and 8 are graphs illustrating a level of accuracy of the prediction according to example embodiments. For instance, FIG. 7 shows that the prediction based on the Al models build for each clusters and applied at the compute nodes was 97% accurate. That is, the prediction has an Fl score of .97 for the compute node according to an example embodiment. Moreover, FIG. 8 shows that the prediction based on the Al models build for each clusters and applied at the management node was 97% accurate. That is, the prediction has an Fl score of .99 for the management node according to an example embodiment.

[0089] The scope of one or more example embodiments also includes a processing method of storing, in a storage medium, a program that causes the configuration of the example embodiment to operate to implement the function of the example embodiment described above, reading out as a code the program stored in the storage medium, and executing the code in a computer. That is, a computer readable storage medium is also included in the scope of each example embodiment. Further, not only the storage medium in which the program described above is stored but also the program itself is included in each example embodiment. Further, one or more components included in the example embodiments described above may be a circuit such as an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA), or the like configured to implement the function of each component.

[0090] As the storage medium, for example, a floppy (registered trademark) disk, a hard disk, an optical disk, a magneto-optical disk, a Compact Disk (CD)-ROM, a magnetic tape, a nonvolatile memory card, or a ROM can be used. Further, the scope of each of the example embodiments includes an example that operates on Operating System (OS) to perform a process in cooperation with another software or a function of an add-in board without being limited to an example that performs a process by an individual program stored in the storage medium.

[0091] Note that all the example embodiments described above are mere examples of embodiments in implementing the disclosure, and the technical scope of the disclosure should not be construed in a limiting sense by these example embodiments. That is, the disclosure can be implemented in various forms without departing from the technical concept thereof or the primary feature thereof.