[go: up one dir, main page]

CN120353681A - Evaluation method and device for server performance - Google Patents

Evaluation method and device for server performance

Info

Publication number
CN120353681A
CN120353681A CN202510838379.1A CN202510838379A CN120353681A CN 120353681 A CN120353681 A CN 120353681A CN 202510838379 A CN202510838379 A CN 202510838379A CN 120353681 A CN120353681 A CN 120353681A
Authority
CN
China
Prior art keywords
model
difference value
operation data
network
performance
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202510838379.1A
Other languages
Chinese (zh)
Inventor
阮志龙
许博
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Suzhou Metabrain Intelligent Technology Co Ltd
Original Assignee
Suzhou Metabrain Intelligent Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Suzhou Metabrain Intelligent Technology Co Ltd filed Critical Suzhou Metabrain Intelligent Technology Co Ltd
Priority to CN202510838379.1A priority Critical patent/CN120353681A/en
Publication of CN120353681A publication Critical patent/CN120353681A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/30Monitoring
    • G06F11/34Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment
    • G06F11/3409Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment for performance assessment
    • G06F11/3428Benchmarking

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Hardware Design (AREA)
  • Quality & Reliability (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Debugging And Monitoring (AREA)

Abstract

本发明公开了一种服务器性能的评估方法和装置,涉及电数字数据处理技术领域,包括采集目标服务器部件在多个环境中的实际运行数据;将实际运行数据输入至预先训练的动作模型中,以输出实际运行数据和对应理论运行数据之间的第一差异值;将第一差异值输入至预先建立的奖励模型中,以输出部件的性能分数,并确定目标服务器的性能评估结果,解决了静态阈值法动态适应性差,误判率较高;而基准测试工具资源消耗大,场景泛化能力弱,实时性较差、测试周期较长,无法支持在线动态评估技术问题,达到了通过多维度整合,简化测试流程,减少人为干预,适用场景更加广泛,准确率更高,此外,通过简化模型表达缩短测试时间,提升测试效率的技术效果。

The present invention discloses a server performance evaluation method and device, which relate to the technical field of electronic digital data processing, including collecting actual operation data of target server components in multiple environments; inputting the actual operation data into a pre-trained action model to output a first difference value between the actual operation data and the corresponding theoretical operation data; inputting the first difference value into a pre-established reward model to output a performance score of the component, and determining the performance evaluation result of the target server, thereby solving the problems that the static threshold method has poor dynamic adaptability and a high misjudgment rate; and the benchmark test tool has large resource consumption, weak scenario generalization ability, poor real-time performance, a long test cycle, and cannot support online dynamic evaluation technology. The method achieves the technical effect of simplifying the test process, reducing human intervention, and making the applicable scenarios more extensive and the accuracy higher through multi-dimensional integration. In addition, the test time is shortened and the test efficiency is improved by simplifying the model expression.

Description

Evaluation method and device for server performance
Technical Field
The invention relates to the technical field of electric digital data processing, in particular to a server performance evaluation method and device.
Background
In the related art, server performance evaluation mainly depends on two methods, namely a static threshold method and a benchmark test tool. The static threshold method is used for realizing performance evaluation by presetting a fixed threshold (for example, the processor utilization rate is greater than 80 percent to trigger an alarm) and combining hardware indexes acquired by a Linux performance tool, and a benchmark test tool is used for simulating a pressure scene through test tools such as JMeter, loadRunner and the like and generating a performance report by utilizing a predefined test case.
However, in the related art, the static threshold method has poor dynamic adaptability, cannot sense load fluctuation, has single index, causes higher misjudgment rate, depends on manpower, and has low efficiency and stronger subjectivity, while the benchmark test tool has large resource consumption, weak scene generalization capability, poor instantaneity and longer test period, cannot support online dynamic evaluation, influences service operation, and is difficult to cope with the real-time evaluation requirement of a large-scale server cluster, so that improvement is needed.
Disclosure of Invention
The invention provides a server performance evaluation method and device, which at least solve the problems of poor dynamic adaptability, single index, higher misjudgment rate, large reference test tool resource consumption, weak scene generalization capability, poor instantaneity, longer test period, incapability of supporting online dynamic evaluation and the like of a static threshold method in the related technology.
The invention provides a server performance evaluation method which comprises the steps of collecting actual operation data of at least one component of a target server in a plurality of environments, inputting the actual operation data into a pre-trained action model to output a first difference value between the actual operation data and corresponding theoretical operation data, inputting the first difference value into a pre-established rewarding model to output performance scores of the at least one component in the corresponding environments, and determining a performance evaluation result of the target server based on the performance scores.
The invention provides a server performance evaluation device which comprises a first acquisition module, a first output module and a first determination module, wherein the first acquisition module is used for acquiring actual operation data of at least one component of a target server in a plurality of environments, the first output module is used for inputting the actual operation data into a pre-trained action model so as to output a first difference value between the actual operation data and corresponding theoretical operation data, and the first determination module is used for inputting the first difference value into a pre-established reward model so as to output performance scores of the at least one component in the corresponding environments and determining a performance evaluation result of the target server based on the performance scores.
The invention also provides a server which comprises a memory for storing a computer program and a processor for realizing the steps of any server performance evaluation method when executing the computer program.
The present invention also provides a computer-readable storage medium having a computer program stored therein, wherein the computer program when executed by a processor implements the steps of any one of the server performance assessment methods described above.
The invention also provides a computer program product comprising a computer program which, when executed by a processor, implements the steps of any of the above methods for evaluating server performance.
According to the invention, the collected actual operation data of at least one component of the target server in a plurality of environments can be input into a pre-trained action model, and then a first difference value between the actual operation data and corresponding theoretical operation data is output, and the first difference value is input into a pre-established rewarding model, and then the performance score of at least one component in the corresponding environments is output, so that the performance evaluation result of the target server is determined, therefore, the problems of poor dynamic adaptability of a static threshold method, single index, higher misjudgment rate, large consumption of reference testing tool resources, weak scene generalization capability, poor instantaneity, longer testing period and incapability of supporting on-line dynamic evaluation are solved, the technical effects of integrating the overall performance of the server through multiple dimensions, simplifying the testing flow, reducing human intervention, improving the stability and reliability of the server performance test, having wider applicable scenes, having higher accuracy, designing different models to evaluate and calculate, simplifying model expression, enabling the model to be easier to train, and improving the testing time and testing efficiency are achieved.
Drawings
For a clearer description of embodiments of the present invention, the drawings that are required to be used in the embodiments will be briefly described, it being apparent that the drawings in the following description are only some embodiments of the present invention, and other drawings may be obtained according to the drawings without inventive effort for those skilled in the art.
FIG. 1 is a flowchart of a method for evaluating server performance according to an embodiment of the present invention;
FIG. 2 is a block diagram of a first action network provided in accordance with one embodiment of the present invention;
FIG. 3 is a block diagram of a second action network provided in accordance with one embodiment of the present invention;
FIG. 4 is a flow chart illustrating the execution of an action network model provided in accordance with one embodiment of the present invention;
FIG. 5 is a flow chart for training a pre-established motion model provided in accordance with one embodiment of the present invention;
FIG. 6 is a flow chart illustrating the operation principle of a server performance evaluation method according to an embodiment of the present invention;
fig. 7 is a block diagram of an evaluation apparatus for server performance according to an embodiment of the present invention.
Reference numerals:
Wherein, 10-the evaluation device of the server performance; 100-first acquisition module, 200-first output module, 300-first determination module.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. Based on the embodiments of the present invention, all other embodiments obtained by a person of ordinary skill in the art without making any inventive effort are within the scope of the present invention.
It should be noted that in the description of the present invention, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. The terms "first," "second," and the like in this specification are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order.
The present invention will be further described in detail below with reference to the drawings and detailed description for the purpose of enabling those skilled in the art to better understand the aspects of the present invention.
The embodiment of the invention provides a server performance evaluation method, which is described in detail by combining an execution flow of the server performance evaluation method.
Specifically, fig. 1 is a flowchart of a method for evaluating server performance according to an embodiment of the present invention.
As shown in fig. 1, the method for evaluating the performance of the server includes the following steps:
In step S101, actual operational data of at least one component of the target server in a plurality of environments is collected.
It should be understood that, in the embodiment of the present invention, at least one component of the target server may include, but is not limited to, a processor (Central Processing Unit, abbreviated as CPU), a memory, a hard disk, a network, a computing card, and the present invention is not limited to the specific one.
In addition, it should be noted that, in the embodiment of the present invention, the related tools, such as Zabbix, nagios, etc., may be used, and the present invention is not limited to specific examples, and the actual operation data may be collected by the performance monitor or the command line tool.
In the actual execution process, the embodiment of the invention can acquire the actual operation data of the processor, the memory, the hard disk, the network, the computing card and the like in the target server in a plurality of environments in real time. Wherein a plurality of environments may be understood as the environment temperature at which the target server operates being different.
By way of example, the embodiment of the invention can adjust the environment temperature, and perform the performance test of the target server under different environment temperatures, thereby obtaining different actual operation data.
Optionally, in one embodiment of the present invention, acquiring actual operation data of at least one component of the target server in a plurality of environments includes acquiring at least one of first data of a processor in the target server, second data of a memory, third data of a hard disk, fourth data of a network, fifth data of a computing card, and sixth data of a different environment, and determining the actual operation data based on at least one of the first data, the second data, the third data, the fourth data, the fifth data, and the sixth data.
It should be understood that in the embodiment of the present invention, the actual running data may include, but is not limited to, the first data of the processor, the second data of the memory, the third data of the hard disk, the fourth data of the network, the fifth data of the computing card, the sixth data of the different environments, and the like, and the present invention is not limited to the specific one.
Further, in the embodiment of the present invention, the first data may include, but is not limited to, a core number, a frequency, a thread number, a thread frequency, etc. of the processor, the present invention is not limited to, the second data may include, but is not limited to, a capacity, a transmission rate, a bandwidth, etc. of the memory, the present invention is not limited to, the third data may include, but is not limited to, a capacity, a transmission rate, a bandwidth, etc. of the hard disk, the present invention is not limited to, the fourth data may include, but is not limited to, a transmission rate, etc. of the network, the present invention is not limited to, the fifth data may include, but is not limited to, a calculation force, a transmission rate, etc. of the calculation force card, the present invention is not limited to, the sixth data may include, but is not limited to, an ambient temperature, an ambient humidity, etc., and the present invention is not limited to.
In some embodiments, the embodiment of the present invention may determine the actual running data by acquiring the first data of the processor, the second data of the memory, the third data of the hard disk, the fourth data of the network, the fifth data of the power card, and the sixth data of the different environments in the target server.
In some embodiments, the embodiment of the present invention may determine the actual running data by acquiring the first data of the processor, the second data of the memory, the third data of the hard disk, the fourth data of the network, and the fifth data of the computing card in the target server.
In some embodiments, the embodiment of the present invention may determine the actual running data by acquiring the first data of the processor, the second data of the memory, and the third data of the hard disk in the target server.
The embodiment of the invention can determine the actual operation data by acquiring the number and frequency of the processor, the capacity and transmission rate of the memory, the capacity and transmission rate of the hard disk, the transmission rate of the network, the computing power and transmission rate of the computing card, the environment temperature and the like, and can be specifically set by a person skilled in the art according to the actual situation, and the invention is not particularly limited.
The actual running data in the embodiment of the invention comprises various data, such as the first data of a processor, the second data of a memory, the third data of a hard disk, the fourth data of a network, the fifth data of a computing card, the sixth data of different environments and the like, can realize multidimensional data association analysis, construct full-dimensional performance images, further improve environmental adaptability and predictive maintenance capability.
Optionally, before the actual operation data is input into the pre-trained action model, the method further comprises the steps of determining at least one of first structural information, second structural information with batch normalization and function information of an activation function of a fully-connected neural network in the action model based on the actual operation data, connecting the fully-connected neural network, the batch normalization and the activation function in series based on the at least one of the first structural information, the second structural information and the function information to obtain a first action network in the action model, combining the at least two first action networks and the first action network in series to obtain a first characteristic of the actual operation data, obtaining a second action network in the action model based on the first characteristic and the first action network, and building the action model based on the first action network and the second action network.
It will be appreciated that in the embodiment of the present invention, the action model may include, but is not limited to, a first action network and a second action network, and specifically may be set by those skilled in the art according to the actual situation, and the present invention is not limited thereto.
The first action network FBR network is shown in connection with fig. 2, where the FBR network may include, but is not limited to, a fully connected neural network (Fully Connected Neural Network, abbreviated as FNN), a batch normalization (Batch Normalization, abbreviated as BN), and an activation function Relu, and the FBR network is obtained by connecting the fully connected neural network, the batch normalization, and the activation function Relu in series.
The second action network FC network is shown in connection with fig. 3, where the FC network may combine the features of two FBR networks and one FBR network connected in series to obtain a first feature, and the first feature is processed by the FBR network to obtain the FC network.
Further, the embodiment of the invention can be described with reference to an execution flow of an action network model shown in fig. 4, and the main content of the method can be that actual operation data is acquired, the actual operation data is processed through an FBR network to obtain initial characteristics, the initial characteristics are processed through the FBR network to obtain second characteristics, the second characteristics are processed through an FC network to obtain third characteristics, the third characteristics are processed through the FBR network to obtain fourth characteristics, the fourth characteristics are processed through the FC network to obtain fifth characteristics, the fifth characteristics are processed through the FBR network to obtain characteristics F1, the fifth characteristics are processed through the FBR network to obtain characteristics F2, the actual operation data is processed through the FBR network to obtain characteristics F3, the characteristics F3 are processed through the FBR network to obtain characteristics F4, the characteristics F4 and the characteristics F2 are processed through the FBR network to obtain characteristics F5, the characteristics F5 and the characteristics F1 are processed through the FBR network to obtain characteristics F6, the characteristics F7 and the characteristics F3 are combined, the sixth characteristics are obtained, the sixth characteristics are processed through the FC network, the fifth characteristics are processed through the fifth characteristics and the FBR network to obtain the characteristics F1, the characteristics F2 is processed through the FBR network to obtain the characteristics F9, the characteristics F9 and the characteristics of the eighth characteristics are processed through the FBR network to obtain the characteristics, the characteristics F9, the characteristics is processed through the characteristics and the characteristics of the FBR 9, and the eighth characteristics is processed through the FBR network to obtain the characteristics, and the characteristics.
In some embodiments, the embodiment of the present invention may determine, based on actual operation data, first structural information of a fully connected neural network, second structural information of batch normalization, and function information of an activation function in an action model, and further connect the fully connected neural network, the batch normalization, and the activation function in series to obtain a first action network, and combine at least two first action networks and the first action network that are connected in series to obtain a first feature of the actual operation data, thereby obtaining a second action network, and further construct the action model.
For example, the embodiment of the present invention may construct a first action network in conjunction with fig. 2, construct a second action network in conjunction with fig. 3, and construct an action model in conjunction with fig. 4.
According to the embodiment of the invention, the action model is constructed through network modularization, so that independent optimization or replacement is realized, further, the maintenance cost is reduced, the low-delay reasoning requirement is met through series enhancement feature fusion, the instantaneity is ensured, the model structure is regulated according to actual operation data, the dynamic regulation of the structure is realized, and the decision quality is improved.
Optionally, in one embodiment of the present invention, before inputting the actual operation data into the pre-trained motion model, the method further includes collecting training operation data of at least one component of the server in a plurality of environments, inputting the training operation data into the pre-established motion model to output a second difference value of the training operation data and the corresponding theoretical operation data, calculating a training performance score of the at least one component in the corresponding environment by using the pre-established reward model based on the second difference value, detecting whether the at least one component meets a preset performance condition based on the training performance score, training parameter information of the pre-established motion model by using discrete estimation to obtain a trained motion model, and generating a component meeting the preset performance condition based on the trained motion model.
In some embodiments, the embodiment of the present invention may train a pre-established motion model before inputting actual operation data into the pre-trained motion model, thereby obtaining a trained motion model. The embodiment of the present invention may be described with reference to fig. 5, where the content of training a pre-established action model may be:
Step S501 is to collect training operation data of at least one component of a server in a plurality of environments.
In the embodiment of the invention, the training operation data is that, wherein,Is the number of cores of the processor; Is the frequency of the processor; is the capacity of the memory; is the transmission rate of the memory; Is the capacity of the hard disk; the transmission rate of the hard disk is the transmission rate of the hard disk; The transmission rate of the network port is the transmission rate of the network port; The calculation value is the calculation value of the calculation card; Is the current temperature value.
Step S502, inputting the training operation data into a pre-established action model to output a second difference value between the training operation data and the corresponding theoretical operation data.
Wherein, the embodiment of the invention can utilize the pre-established action modelCalculating the difference value between the training data of processor and the corresponding theoretical running data by usingCalculating the difference value between the memory training data and the corresponding theoretical operation data by usingCalculating the difference value between the hard disk training data and the corresponding theoretical operation data by usingCalculating the difference value between the network training data and the corresponding theoretical operation data by usingAnd calculating the difference value between the power card training data and the corresponding theoretical operation data.
Further, in an embodiment of the present invention,The maximum thread and frequency (such as the turbo frequency, etc.) of the main test processor are not particularly limited by the present invention; the capacity and transmission rate of the memory (such as between the memory and the hard disk, between the memory and the accelerator card, etc.) are mainly tested, and the invention is not limited in particular; The capacity and transmission rate of the hardware are mainly tested (such as between the hardware and the hard disk, between the hard disk and the memory, etc.), and the invention is not limited in particular; the transmission rate of the equipment network is mainly tested; The present invention is not particularly limited, and the calculation force value and the data transmission rate of the calculation force card (such as between the memory and the acceleration card, between the acceleration card and the acceleration card, etc.) are mainly tested.
Step S503, calculating training performance scores by using a pre-established rewarding model.
The training performance scores of the processor, the memory, the hard disk, the network, the power card and the energy consumption under the corresponding environments can be calculated by utilizing the pre-established rewarding model based on the second difference value.
Step S504, judging whether the performance of the server meets a certain performance condition.
In the embodiment of the present invention, if a certain performance condition is satisfied, step S506 is performed, and otherwise, step S505 is performed. Wherein, certain performance conditions can be set by a person skilled in the art according to practical situations, and the invention is not particularly limited.
Step S505, retraining the pre-established motion model with the discrete estimation.
The method and the device can train parameter information of the pre-established action model by utilizing discrete estimation, further obtain a trained action model, and generate a component meeting certain performance conditions based on the trained action model.
And S506, determining a performance evaluation result of the server.
According to the embodiment of the invention, the scene coverage can be improved by collecting training data under different environments, the accuracy of model prediction is quantized by comparing actual and theoretical data, the evaluation and the optimization are facilitated, certain performance conditions are set to ensure that the performance of the component reaches the expected standard, and an optimization flow is triggered when the performance of the component does not reach the standard, so that the model effect is improved.
Optionally, in one embodiment of the invention, the parameter information of the pre-established motion model is trained by using discrete estimation, wherein the parameter information comprises probability distribution of motion selection in training parameter information based on gradient information of the discrete estimation, parameter update amplitude in training parameter information based on learning rate information of the discrete estimation, parameter adjustment sensitivity in training parameter information based on logarithmic information of the discrete estimation, and motion value in training parameter information based on a dominance function of the discrete estimation.
In some embodiments, the embodiments of the present invention may feed back the value of the pre-established reward model to the pre-trained action model through unbiased discrete estimation, and further retrain the parameter information of the pre-established action model, where the expression of the discrete estimation may be, but is not limited to:
,
wherein, the Determining probability distribution of action selection for discrete estimated parameters, and updating by gradient rise to enable the strategy to trend to actions with high dominant values; for the learning rate (step length) of discrete estimation, controlling the updating amplitude of parameters, wherein too high learning rate can cause strategy oscillation, and too low learning rate can be slow in convergence; Reflecting the sensitivity of the current action to parameter adjustment for the gradient of the logarithmic probability of the strategy function, and back-propagating the dominant signal to the strategy network through a chain rule; measuring motion as a dominance function In stateThe following relative values, the expressions of which may be, but are not limited to:
,
,
,
wherein, the Is expressed in state as action cost functionExecute action downwardsThe desired accumulation of the obtained values is obtained,Is in state ofCost function, in stateThe average expectation of the policy to be followed is that,For time sequence difference errors, the instantaneous difference between the function estimated value and the actual value,For the current prize value,Is the parameter of the ultrasonic wave to be used as the ultrasonic wave,And marking the current time.
The embodiment of the invention can control the action selection process more accurately through discrete estimation of training probability distribution, improve the exploration and utilization balance, provide more stable learning rate information, improve the stability of the model in the training process, help the model to adjust parameters more finely, and further help the model to make better decisions.
In step S102, the actual operation data is input into a pre-trained motion model to output a first difference value between the actual operation data and the corresponding theoretical operation data.
In some embodiments, the embodiment of the present invention may input actual operation data into a pre-trained motion model, so as to obtain a first difference value between the actual operation data and corresponding theoretical operation data.
By way of example, the embodiment of the invention can calculate the first difference value between the actual operation data and the corresponding theoretical operation data through a pre-trained action model.
Optionally, in one embodiment of the invention, the actual operation data is input into a pre-trained action model to output a first difference value between the actual operation data and the corresponding theoretical operation data, wherein the first difference value comprises a processor difference value between the actual operation data of the processor and the theoretical operation data of the processor based on at least one of thread information and frequency information of the processor in the actual operation data, the pre-trained action model is utilized to output the actual operation data of the processor and the processor difference value of the theoretical operation data of the processor, at least one of capacity information and transmission rate information of a memory in the actual operation data is utilized to output a memory difference value between the actual operation data of the memory and the theoretical operation data of the memory based on the pre-trained action model, at least one of capacity information and transmission rate information of a hard disk in the actual operation data is utilized to output a hard disk difference value of the actual operation data of the hard disk and the theoretical operation data of the hard disk based on the pre-trained action model, a calculation card in the actual operation data and the transmission rate information of the actual operation data is utilized to output a network difference value of the actual operation data of the network based on the transmission rate information of the actual operation data, and a calculation card in the actual operation data of the hard disk and the theoretical operation data of the network is utilized to calculate a calculation card force and a difference value of the actual operation card in the actual operation data of the hard disk and the theoretical operation data of the actual operation data of the hard operation data.
It may be appreciated that the pre-trained motion model in the embodiment of the present invention increases motion disturbance, such as a processor thread step, a steep increase in memory transmission, and a hard disk read/write rate, which is not particularly limited by the present invention.
Further, in some embodiments, the embodiment of the present invention may output the processor difference value between the actual running data and the theoretical running data of the processor by using the processor thread step action disturbance in the pre-trained action model based on the thread information and the frequency information of the processor in the actual running data, or may be based on other action disturbance, which is not limited in particular.
In some embodiments, the embodiments of the present invention may use the memory transmission steep action disturbance in the pre-trained action model to output the memory difference value between the actual operation data and the theoretical operation data of the memory based on the capacity information and the transmission rate information of the memory in the actual operation data, or may use other action disturbance, and the present invention is not limited in particular.
In some embodiments, the embodiments of the present invention may output a hard disk difference value between actual operation data and theoretical operation data of a hard disk by using a hard disk read-write rate action disturbance in a pre-trained action model based on capacity information and transmission rate information of the hard disk in the actual operation data, or may be based on other action disturbances, which is not particularly limited by the present invention.
In some embodiments, the embodiments of the present invention may output the network difference value between the actual operation data and the theoretical operation data of the network by using a pre-trained motion model based on the transmission rate information of the network in the actual operation data.
In some embodiments, the embodiment of the invention can output the difference value of the calculation card between the actual operation data and the theoretical operation data of the calculation card by utilizing a pre-trained action model based on the calculation data and the transmission rate information of the calculation card in the actual operation data.
Further, the embodiment of the invention can obtain the first difference value based on the processor difference value, the memory difference value, the hard disk difference value, the network difference value, the computing card difference value and the like.
According to the embodiment of the invention, the processor difference value, the memory difference value, the hard disk difference value, the network difference value, the computing power card difference value and the like can be analyzed, so that the capability of diagnosing the fine performance is realized, the resource optimization accuracy is improved, the predictive maintenance is enhanced, and the stability of the model is improved.
Optionally, in one embodiment of the present invention, before the first difference value is input into the pre-established rewards model, the method further comprises the steps of establishing a processor rewards model in the rewards model by using task information, thread information and instruction information in the processor difference value based on the processor difference value in the first difference value, establishing a memory rewards model in the rewards model by using data information and bandwidth information in the memory difference value based on the memory difference value in the first difference value, establishing a hard disk rewards model in the rewards model by using bandwidth information in the hard disk difference value based on the hard disk difference value in the first difference value, establishing a network rewards model in the rewards model by using transmission data information in the network difference value based on the network difference value in the first difference value, establishing a credit card rewards model in the rewards model by using credit information in the credit card difference value based on the calculated difference value, establishing an energy consumption rewards model in the rewards model based on the energy consumption information in the first difference value, establishing a memory rewards model based on the processor rewards model, the hard disk, the network, the card and the energy consumption model in the rewards model. Wherein the expression of the processor reward model may be, but is not limited to,:
,
wherein, the The method comprises the steps of (1) making instruction numbers for tasks; for the number of threads to be theoretical, Is the theoretical thread frequency; the number of the instructions can be processed for each period of the chip actually; Is the actual completion time; is the number of required threads.
The expression of the memory reward model may be, but is not limited to:
,
wherein, the Data total; is the theoretical bandwidth; Is the actual bandwidth; Is the theoretical capacity.
The expression of the hard disk reward model may be, but is not limited to,:
,
wherein, the Data total; is the theoretical bandwidth; Is the actual bandwidth; Is the theoretical capacity.
The expression of the network rewards model may be, but is not limited to,:
,
wherein, the For the pre-transmission data size; Is the actual received data size.
The expression of the card reward model may be, but is not limited to,:
,
wherein, the Calculating the amount for the model; the force value is calculated for the actual.
The expression of the energy consumption rewards model may be, but is not limited to,:
,
wherein, the Is theoretical energy consumption; is the actual energy consumption value.
In some embodiments, the reward model that may be built by the embodiments of the present invention may include, but is not limited to, a processor reward model, a memory reward model, a hard disk reward model, a network reward model, a credit card reward model, and an energy consumption reward model, and the present invention is not limited in particular.
The embodiment of the invention can establish a processor rewarding model by utilizing task information, thread information and instruction information in the processor difference value, and the expression can be but is not limited to:
,
wherein, the The method comprises the steps of (1) making instruction numbers for tasks; for a theoretical/nominal number of threads, Is the theoretical/nominal thread frequency; the number of the instructions can be processed for each period of the chip actually; Is the actual completion time; is the number of required threads.
In some embodiments, the embodiment of the present invention may use the data information and the bandwidth information in the memory discrepancy value to build a memory reward model, and the expression may be, but is not limited to:
,
wherein, the Data total; Is theoretical/nominal bandwidth; Is the actual bandwidth; is theoretical/nominal capacity.
In some embodiments, the embodiments of the present invention may use bandwidth information in the hard disk difference value to build a hard disk reward model, where the expression may be, but is not limited to:
,
wherein, the Data total; Is theoretical/nominal bandwidth; For the actual bandwidth, the value is from the beginning of data transmission to the end of data transmission, and is the data group; is theoretical/nominal capacity.
In some embodiments, the embodiments of the present invention may use the transmission data information in the network discrepancy value to build a network rewards model, and the expression may be, but is not limited to,:
,
wherein, the For the pre-transmission data size; For the size of the data actually received, packet loss is mainly prevented.
In some embodiments, the embodiments of the present invention may use the computing power information in the computing power card difference value to build a computing power card rewards model, and the expression may be, but is not limited to,:
,
wherein, the Calculating the amount for the model; the force value is calculated for the actual.
In some embodiments, the embodiment of the present invention may establish an energy consumption rewards model based on the energy consumption information in the first difference value, and the expression may be, but is not limited to,:
,
wherein, the Is theoretical/nominal energy consumption; is the actual energy consumption value.
According to the embodiment of the invention, the task scheduling and thread management of the processor can be optimized through the processor rewarding model, the processing efficiency is improved, the use and allocation of the memory are optimized through the memory rewarding model, the memory bottleneck is reduced, the read-write operation of the hard disk is optimized through the hard disk rewarding model, the data access speed is improved, the allocation of network resources is optimized through the network rewarding model, the network delay is reduced, the use of the power card is optimized through the power card rewarding model, the computing efficiency is improved, the energy consumption is reduced while the performance is ensured through the energy consumption rewarding model, the green computing is realized, the advantages of the rewarding models of all components are integrated, and the performance, the resource management and the energy efficiency of the server are comprehensively optimized.
In step S103, the first difference value is input into a pre-established bonus model to output a performance score of at least one component in a corresponding environment, and a performance evaluation result of the target server is determined based on the performance score.
In some embodiments, the first difference value may be input into a pre-established reward model, so that performance scores of different components under corresponding environments are determined, thereby determining a performance evaluation result of the server.
The embodiment of the invention can obtain the performance score of the processorMemory performance scoreHard disk performance scorePerformance score of the power cardNetwork performance scoreAnd energy consumption performance scoreEtc., the present invention is not particularly limited.
Further, the embodiment of the invention can determine the performance evaluation result based on the processor performance score, the memory performance score, the hard disk performance score, the network performance score and the power card performance score, and the calculation formula can be but is not limited to:
,
wherein, the Is a super parameter; the performance evaluation result of the server is obtained; a processor performance score; is the memory performance score; the hard disk performance score; Is a network performance score; calculating the performance score of the force card; Is an energy consumption performance fraction.
The embodiment of the invention can comprehensively consider the processor performance score, the memory performance score, the hard disk performance score, the network performance score, the power card performance score and the energy consumption performance score, has more accurate evaluation results and wider application scenes.
Optionally, in one embodiment of the present invention, inputting the first differential value into a pre-established rewards model to output a performance score of the at least one component under the corresponding environment includes calculating a processor performance score of a processor in the server with the processor rewards model in the pre-established rewards model based on the processor differential value in the first differential value, calculating a memory performance score of a memory in the server with the memory rewards model in the pre-established rewards model based on the memory differential value in the first differential value, calculating a hard disk performance score of a hard disk in the server with the hard disk rewards model in the pre-established rewards model based on the hard disk differential value in the first differential value, calculating a network performance score of the network in the server with the network rewards model in the pre-established rewards model based on the network differential value in the first differential value, calculating a power card performance score of a power card in the server with the power card in the pre-established model based on the power card differential value in the first differential value, and calculating a power consumption of the server with the pre-established power consumption model in the power consumption model based on the power consumption differential value in the first differential value.
In some embodiments, the embodiments of the present invention may calculate the processor performance score using a processor reward model of the pre-established reward models based on the processor difference value of the first difference values.
In some embodiments, the memory performance score may be calculated by using a memory reward model in a pre-established reward model based on a memory difference value in the first difference value.
In some embodiments, the hard disk performance score may be calculated by using a hard disk reward model in a pre-established reward model based on the hard disk difference value in the first difference value.
In some embodiments, the embodiments of the present invention may calculate the network performance score using a network rewards model of the pre-established rewards model based on the network discrepancy value of the first discrepancy values.
In some embodiments, the embodiments of the present invention may calculate the power card performance score using a power card reward model of the pre-established reward models based on the power card differential value of the first differential values.
In some embodiments, the energy consumption performance score may be calculated by using an energy consumption rewarding model in a pre-established rewarding model based on the energy consumption difference value in the first difference value.
The embodiment of the invention can utilize each component to evaluate independently, more accurately position the performance bottleneck, facilitate maintenance and expansion, enhance the interpretation of the result, comprehensively consider a plurality of components and energy consumption, help to find the balance between the performance and the energy efficiency, adapt to different loads and environmental changes and make more intelligent decisions.
Alternatively, in one embodiment of the invention, determining the performance evaluation result of the target server based on the performance score includes counting a total number of performance scores based on the performance score, calculating an initial performance evaluation result of the target server based on the performance score, and determining the performance evaluation result based on the total number and the initial performance evaluation result.
It can be understood that, in order to ensure that the performances of the processor, the memory, the hardware and the like meet the requirements, the embodiment of the invention can perform multiple evaluations, such as twice, three times and the like, and the invention is not particularly limited, so as to obtain different performance scores.
As a possible implementation manner, the embodiment of the invention can count the total number of the performance scores, further calculate the initial performance evaluation result of the target server, and determine the performance evaluation result based on the total number and the initial performance evaluation result.
For example, the embodiment of the invention can calculate the initial performance evaluation result by using a calculation formula of the performance evaluation result, and take a flat distance value of the initial performance evaluation result as the performance evaluation result.
According to the embodiment of the invention, the performance of a plurality of components is comprehensively considered by counting the total number of the performance scores, the one-sided performance of single component evaluation is avoided, the overall performance of the server is more accurately reflected, the evaluation result is more in line with the requirements of actual application scenes, the performance bottleneck can be rapidly positioned, a targeted optimization strategy is formulated, the operation and maintenance cost is reduced, and the resource utilization rate is improved.
The working principle of the server performance evaluation method according to the embodiment of the present invention is described below with reference to a specific embodiment.
Fig. 6 is a flowchart of an operation principle of a server performance evaluation method according to an embodiment of the present invention.
And step S601, collecting actual operation data of a target server in different environments.
In the embodiment of the invention, different environments may be different environmental temperatures.
Step S602, inputting actual operation data into a pre-trained action model to obtain a first difference value between the actual operation data and corresponding theoretical operation data.
The embodiment of the invention can calculate the first difference value between the actual operation data and the theoretical operation data of the processor, the memory, the hardware, the network, the power card, the energy consumption and the like by utilizing the pre-trained action model.
And step S603, inputting the first difference value into a pre-established rewarding model to obtain the performance score under the corresponding environment.
The embodiment of the invention can calculate the processor performance score, the memory performance score, the hard disk performance score, the power card performance score, the network performance score, the energy consumption performance score and the like by utilizing a pre-established model, and is not particularly limited.
Step S604, determining an initial performance evaluation result of the server.
The embodiment of the invention can substitute the processor performance score, the memory performance score, the hard disk performance score, the power card performance score, the network performance score, the energy consumption performance score and the like into a calculation formula of the performance evaluation result to calculate the initial performance evaluation result.
And step S605, determining a performance evaluation result of the server.
The embodiment of the invention can calculate the corresponding average value based on the total number of the initial performance evaluation results, so as to determine the performance evaluation result.
From the description of the above embodiments, it will be clear to a person skilled in the art that the method according to the above embodiments may be implemented by means of software plus the necessary general hardware platform, but of course also by means of hardware, but in many cases the former is a preferred embodiment.
According to the evaluation method for the server performance, the collected actual operation data of at least one component of the target server in a plurality of environments can be input into the pre-trained action model, the first difference value between the actual operation data and the corresponding theoretical operation data is output, the first difference value is input into the pre-established rewarding model, and the performance score of at least one component in the corresponding environment is output, so that the performance evaluation result of the target server is determined, and therefore the problems that the static threshold method is poor in dynamic adaptability, single in index and high in misjudgment rate can be solved, the reference test tool is high in resource consumption, weak in scene generalization capability, poor in instantaneity and long in test period, the technical problem of online dynamic evaluation cannot be supported is solved, the overall performance of the server is judged through multi-dimensional integration, the test flow is simplified, the human intervention is reduced, the stability and reliability of the server performance test are improved, the applicable scene is wider, the accuracy is higher, the different models are designed for evaluation calculation, the model expression is simplified, the model is easier to train, the test time is shortened, and the test efficiency is improved.
The embodiment of the invention also provides a server performance evaluation device.
Fig. 7 is a block diagram of an evaluation apparatus for server performance according to an embodiment of the present invention.
As shown in fig. 7, the server performance evaluation apparatus 10 includes a first acquisition module 100, a first output module 200, and a first determination module 300.
Wherein the first acquisition module 100 is configured to acquire actual operation data of at least one component of the target server in a plurality of environments.
The first output module 200 is configured to input actual operation data into a pre-trained motion model, so as to output a first difference value between the actual operation data and corresponding theoretical operation data.
The first determining module 300 is configured to input the first difference value into a pre-established rewards model, to output a performance score of at least one component in a corresponding environment, and determine a performance evaluation result of the target server based on the performance score.
Optionally, in one embodiment of the present invention, the method further comprises a second determining module, a first generating module, a second generating module, a third generating module and a first constructing module.
The second determining module is used for determining at least one of the first structural information of the fully-connected neural network, the second structural information of batch normalization and the function information of the activation function in the action model based on the actual operation data before the actual operation data is input into the pre-trained action model.
The first generation module is used for connecting the fully-connected neural network, the batch normalization and the activation function in series based on at least one of the first structure information, the second structure information and the function information so as to obtain a first action network in the action model.
And the second generation module is used for merging the at least two first action networks connected in series with the first action network so as to obtain the first characteristic of the actual operation data.
And the third generation module is used for obtaining a second action network in the action model based on the first characteristic and the first action network.
The first construction module is used for constructing an action model based on the first action network and the second action network.
Optionally, in one embodiment of the present invention, the system further comprises a second acquisition module, a second output module, a calculation module, a detection module and a fourth generation module.
Wherein the second acquisition module is used for acquiring training operation data of at least one component of the server in a plurality of environments before inputting actual operation data into the pre-trained action model.
And the second output module is used for inputting the training operation data into a pre-established action model so as to output a second difference value of the training operation data and the corresponding theoretical operation data.
And the calculation module is used for calculating training performance scores of at least one component in corresponding environments by using a pre-established rewarding model based on the second difference value.
And the detection module is used for detecting whether at least one component meets the preset performance condition or not based on the training performance score.
And the fourth generation module is used for training the parameter information of the pre-established action model by utilizing discrete estimation under the condition that at least one component does not meet the preset performance condition so as to obtain a trained action model, and generating the component meeting the preset performance condition based on the trained action model.
Optionally, in one embodiment of the present invention, the fourth generating module includes a first training unit, a second training unit, a third training unit, and a fourth training unit.
The first training unit is used for training probability distribution of action selection in the parameter information based on gradient information of discrete estimation.
And the second training unit is used for training the parameter updating amplitude in the parameter information based on the discretely estimated learning rate information.
And a third training unit for training the parameter adjustment sensitivity in the parameter information based on the discrete estimated logarithmic information.
And the fourth training unit is used for training the action value in the parameter information based on the dominance function of the discrete estimation.
Alternatively, in one embodiment of the present invention, the first acquisition module 100 includes an acquisition unit and a first determination unit.
The system comprises an acquisition unit, a storage unit and a storage unit, wherein the acquisition unit is used for acquiring at least one of first data of a processor, second data of a memory, third data of a hard disk, fourth data of a network, fifth data of a computing card and sixth data of different environments in a target server.
And a first determining unit configured to determine actual operation data based on at least one of the first data, the second data, the third data, the fourth data, the fifth data, and the sixth data.
Alternatively, in one embodiment of the present invention, the first output module 200 includes a first output unit, a second output unit, a third output unit, a fourth output unit, a fifth output unit, and a generation unit.
The first output unit is used for outputting processor difference values of the actual operation data of the processor and the theoretical operation data of the processor by utilizing a pre-trained action model based on at least one of the thread information and the frequency information of the processor in the actual operation data.
And the second output unit is used for outputting the memory difference value of the actual operation data of the memory and the theoretical operation data of the memory by utilizing a pre-trained action model based on at least one of capacity information and transmission rate information of the memory in the actual operation data.
And the third output unit is used for outputting hard disk difference values of the actual operation data of the hard disk and the theoretical operation data of the hard disk by utilizing a pre-trained action model based on at least one of capacity information and transmission rate information of the hard disk in the actual operation data.
And the fourth output unit is used for outputting the network difference value of the actual operation data of the network and the theoretical operation data of the network by utilizing the pre-trained action model based on the transmission rate information of the network in the actual operation data.
And a fifth output unit for outputting a calculation card difference value of actual operation data of the calculation card and theoretical operation data of the calculation card by using a pre-trained action model based on at least one of calculation data and transmission rate information of the calculation card in the actual operation data.
The generating unit is used for obtaining a first difference value based on at least one of the processor difference value, the memory difference value, the hard disk difference value, the network difference value and the power card difference value.
Optionally, in one embodiment of the present invention, the method further comprises a second building module, a third building module, a fourth building module, a fifth building module, a sixth building module, a seventh building module, and an eighth building module.
The second construction module is used for constructing the processor rewarding model in the rewarding model by utilizing the task information, the thread information and the instruction information in the processor discrepancy value based on the processor discrepancy value in the first discrepancy value before the first discrepancy value is input into the pre-established rewarding model.
And the third construction module is used for establishing a memory rewarding model in the rewarding model by utilizing the data information and the bandwidth information in the memory difference value based on the memory difference value in the first difference value.
And the fourth construction module is used for constructing a hard disk rewarding model in the rewarding model by utilizing the bandwidth information in the hard disk difference value based on the hard disk difference value in the first difference value.
And a fifth construction module, configured to establish a network rewarding model in the rewarding model by using the transmission data information in the network discrepancy value based on the network discrepancy value in the first discrepancy value.
And a sixth construction module, configured to establish a card rewarding model in the rewarding model by using the power information in the card difference value based on the power card difference value in the first difference value.
And a seventh construction module, configured to establish an energy consumption rewarding model in the rewarding models based on the energy consumption information in the first difference value.
And an eighth building module for building a rewards model based on at least one of a processor rewards model, a memory rewards model, a hard disk rewards model, a network rewards model, a power card rewards model and an energy consumption rewards model.
Alternatively, in one embodiment of the present invention, the first determining module 300 includes a first calculating unit, a second calculating unit, a third calculating unit, a fourth calculating unit, a fifth calculating unit, and a sixth calculating unit.
The first calculating unit is used for calculating the processor performance score of the processor in the server by using the processor rewarding model in the pre-established rewarding model based on the processor difference value in the first difference value.
And the second calculation unit is used for calculating the memory performance score of the memory in the server by using the memory reward model in the pre-established reward model based on the memory difference value in the first difference value.
And a third calculation unit for calculating the hard disk performance score of the hard disk in the server by using the hard disk rewarding model in the pre-established rewarding model based on the hard disk difference value in the first difference value.
And a fourth calculation unit for calculating a network performance score of the network in the server using a network rewards model in the pre-established rewards model based on the network difference value in the first difference value.
And a fifth calculation unit, configured to calculate a power card performance score of the power card in the server using the power card rewarding model in the pre-established rewarding model based on the power card difference value in the first difference value.
And a sixth calculation unit for calculating the energy consumption performance score of the server by using the energy consumption rewarding model in the pre-established rewarding model based on the energy consumption difference value in the first difference value.
Alternatively, in one embodiment of the present invention, wherein,
The expression of the processor reward model may be, but is not limited to,:
,
wherein, the The method comprises the steps of (1) making instruction numbers for tasks; for the number of threads to be theoretical, Is the theoretical thread frequency; the number of the instructions can be processed for each period of the chip actually; Is the actual completion time; is the number of required threads.
The expression of the memory reward model may be, but is not limited to:
,
wherein, the Data total; is the theoretical bandwidth; Is the actual bandwidth; Is the theoretical capacity.
The expression of the hard disk reward model may be, but is not limited to,:
,
wherein, the Data total; is the theoretical bandwidth; Is the actual bandwidth; Is the theoretical capacity.
The expression of the network rewards model may be, but is not limited to,:
,
wherein, the For the pre-transmission data size; Is the actual received data size.
The expression of the card reward model may be, but is not limited to,:
,
wherein, the Calculating the amount for the model; the force value is calculated for the actual.
The expression of the energy consumption rewards model may be, but is not limited to,:
,
wherein, the Is theoretical energy consumption; is the actual energy consumption value.
Optionally, in one embodiment of the present invention, the first determining module 300 includes a statistics unit, a seventh calculation unit, and a second determining unit.
Wherein, the statistics unit is used for counting the total number of the performance scores based on the performance scores.
And a seventh calculation unit for calculating an initial performance evaluation result of the target server based on the performance score.
And a second determining unit configured to determine a performance evaluation result based on the total number and the initial performance evaluation result.
The description of the features in the embodiment corresponding to the evaluation device of the server performance may refer to the related description of the embodiment corresponding to the evaluation method of the server performance, which is not described herein in detail.
According to the evaluation device for the server performance, which is provided by the embodiment of the invention, the collected actual operation data of at least one component of the target server in a plurality of environments can be input into a pre-trained action model, the first difference value between the actual operation data and the corresponding theoretical operation data is output, the first difference value is input into a pre-established rewarding model, and the performance score of at least one component in the corresponding environment is output, so that the performance evaluation result of the target server is determined, therefore, the problems of poor dynamic adaptability, single index and higher misjudgment rate of a static threshold method can be solved, the reference test tool has the advantages of high resource consumption, weak scene generalization capability, poor instantaneity and longer test period, the technical problem of on-line dynamic evaluation cannot be supported, the problem of integrating through multiple dimensions, the judgment of the overall performance of the unified server is achieved, the test flow is simplified, the human intervention is reduced, the stability and reliability of the server performance test are improved, the applicable scene is wider, the accuracy is higher, the evaluation calculation is performed by designing different models, the model expression is simplified, the model is easier to train, and the test efficiency is improved.
An embodiment of the invention also provides a server comprising a memory in which a computer program is stored and a processor arranged to run the computer program to perform the steps of any of the above described embodiments of the method of evaluating server performance.
An embodiment of the present invention also provides a computer readable storage medium having a computer program stored therein, wherein the computer program is configured to perform, when run, the steps of any of the above embodiments of the method for evaluating server performance.
In an exemplary embodiment, the computer readable storage medium may include, but is not limited to, a U disk, a Read-Only Memory (ROM), a random access Memory (Random Access Memory, RAM), a removable hard disk, a magnetic disk, or an optical disk, etc. various media in which a computer program may be stored.
The embodiment of the invention also provides a computer program product, which comprises a computer program, and the computer program is executed by a processor to implement the steps in the embodiment of any one of the server performance evaluation methods.
Embodiments of the present invention also provide another computer program product comprising a non-volatile computer readable storage medium storing a computer program which, when executed by a processor, implements the steps of any of the above embodiments of the server performance assessment method.
Those of skill would further appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both, and that the various illustrative elements and steps are described above generally in terms of functionality in order to clearly illustrate the interchangeability of hardware and software. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
The method for evaluating the server performance provided by the invention is described in detail above. The principles and embodiments of the present invention have been described herein with reference to specific examples, the description of which is intended only to facilitate an understanding of the method of the present invention and its core ideas. It should be noted that it will be apparent to those skilled in the art that various modifications and adaptations of the invention can be made without departing from the principles of the invention and these modifications and adaptations are intended to be within the scope of the invention as defined in the following claims.

Claims (15)

1.一种服务器性能的评估方法,其特征在于,包括以下步骤:1. A method for evaluating server performance, comprising the following steps: 采集目标服务器的至少一个部件在多个环境中的实际运行数据;Collecting actual operation data of at least one component of the target server in multiple environments; 将所述实际运行数据输入至预先训练的动作模型中,以输出所述实际运行数据和对应理论运行数据之间的第一差异值;Inputting the actual operation data into a pre-trained motion model to output a first difference value between the actual operation data and the corresponding theoretical operation data; 将所述第一差异值输入至预先建立的奖励模型中,以输出所述至少一个部件在对应环境下的性能分数,并基于所述性能分数确定所述目标服务器的性能评估结果。The first difference value is input into a pre-established reward model to output a performance score of the at least one component in a corresponding environment, and a performance evaluation result of the target server is determined based on the performance score. 2.根据权利要求1所述的服务器性能的评估方法,其特征在于,在将所述实际运行数据输入至预先训练的动作模型之前,还包括:2. The method for evaluating server performance according to claim 1, characterized in that before inputting the actual operation data into the pre-trained action model, it also includes: 基于所述实际运行数据,确定动作模型中全连接神经网络的第一结构信息、批量归一化的第二结构信息和激活函数的函数信息中的至少一个;Based on the actual operation data, determining at least one of first structural information of a fully connected neural network, second structural information of batch normalization, and function information of an activation function in the action model; 基于所述第一结构信息、所述第二结构信息和所述函数信息中的至少一个,将所述全连接神经网络、所述批量归一化和所述激活函数进行串联,以得到所述动作模型中的第一动作网络;Based on at least one of the first structure information, the second structure information and the function information, the fully connected neural network, the batch normalization and the activation function are connected in series to obtain a first action network in the action model; 将串联的至少两个所述第一动作网络和所述第一动作网络进行合并,以得到所述实际运行数据的第一特征;Merging at least two of the first action networks connected in series with the first action network to obtain a first feature of the actual operation data; 基于所述第一特征和所述第一动作网络,得到所述动作模型中的第二动作网络;Based on the first feature and the first action network, obtaining a second action network in the action model; 基于所述第一动作网络和所述第二动作网络,构建所述动作模型。The action model is constructed based on the first action network and the second action network. 3.根据权利要求1所述的服务器性能的评估方法,其特征在于,在将所述实际运行数据输入至预先训练的动作模型之前,还包括:3. The method for evaluating server performance according to claim 1, characterized in that before inputting the actual operation data into the pre-trained action model, it also includes: 采集服务器的至少一个部件在多个环境中的训练运行数据;Collecting training operation data of at least one component of the server in multiple environments; 将所述训练运行数据输入至预先建立的动作模型中,以输出所述训练运行数据和对应理论运行数据的第二差异值;Inputting the training operation data into a pre-established motion model to output a second difference value between the training operation data and the corresponding theoretical operation data; 基于所述第二差异值,利用预先建立的奖励模型计算所述至少一个部件在对应环境下的训练性能分数;Based on the second difference value, calculating a training performance score of the at least one component in the corresponding environment using a pre-established reward model; 基于所述训练性能分数,检测所述至少一个部件是否满足预设性能条件;Based on the training performance score, detecting whether the at least one component meets a preset performance condition; 在所述至少一个部件不满足所述预设性能条件的情况下,利用离散估计训练所述预先建立的动作模型的参数信息,以得到训练后的动作模型,并基于所述训练后的动作模型生成满足所述预设性能条件的部件。In the case that at least one component does not meet the preset performance conditions, the parameter information of the pre-established motion model is trained using discrete estimation to obtain a trained motion model, and a component that meets the preset performance conditions is generated based on the trained motion model. 4.根据权利要求3所述的服务器性能的评估方法,其特征在于,所述利用离散估计训练所述预先建立的动作模型的参数信息,包括:4. The method for evaluating server performance according to claim 3, wherein the step of training the parameter information of the pre-established motion model by discrete estimation comprises: 基于所述离散估计的梯度信息,训练所述参数信息中动作选择的概率分布;Based on the discrete estimated gradient information, training the probability distribution of action selection in the parameter information; 基于所述离散估计的学习率信息,训练所述参数信息中的参数更新幅度;Based on the discrete estimated learning rate information, training the parameter update amplitude in the parameter information; 基于所述离散估计的对数信息,训练所述参数信息中的参数调整敏感度;Based on the logarithmic information of the discrete estimate, training the parameter adjustment sensitivity in the parameter information; 基于所述离散估计的优势函数,训练所述参数信息中的动作价值。The action values in the parameter information are trained based on the discrete estimated advantage function. 5.根据权利要求1所述的服务器性能的评估方法,其特征在于,所述采集目标服务器的至少一个部件在多个环境中的实际运行数据,包括:5. The method for evaluating server performance according to claim 1, wherein the collecting actual operation data of at least one component of the target server in multiple environments comprises: 获取所述目标服务器中处理器的第一数据、内存的第二数据、硬盘的第三数据、网路的第四数据、算力卡的第五数据和不同环境的第六数据中的至少一个;Obtain at least one of first data of a processor, second data of a memory, third data of a hard disk, fourth data of a network, fifth data of a computing power card, and sixth data of different environments in the target server; 基于所述第一数据、所述第二数据、所述第三数据、所述第四数据、所述第五数据和所述第六数据中的至少一个,确定所述实际运行数据。The actual operation data is determined based on at least one of the first data, the second data, the third data, the fourth data, the fifth data, and the sixth data. 6.根据权利要求1所述的服务器性能的评估方法,其特征在于,所述将所述实际运行数据输入至预先训练的动作模型中,以输出所述实际运行数据和对应理论运行数据之间的第一差异值,包括:6. The method for evaluating server performance according to claim 1, wherein the step of inputting the actual operation data into a pre-trained motion model to output a first difference value between the actual operation data and the corresponding theoretical operation data comprises: 基于所述实际运行数据中处理器的线程信息和频率信息中的至少一个,利用所述预先训练的动作模型输出所述处理器的实际运行数据和所述处理器的理论运行数据的处理器差异值;Based on at least one of thread information and frequency information of the processor in the actual operation data, outputting a processor difference value between the actual operation data of the processor and theoretical operation data of the processor using the pre-trained motion model; 基于所述实际运行数据中内存的容量信息和传输速率信息中的至少一个,利用所述预先训练的动作模型输出所述内存的实际运行数据和所述内存的理论运行数据的内存差异值;Based on at least one of the capacity information and the transmission rate information of the memory in the actual operation data, outputting a memory difference value between the actual operation data of the memory and the theoretical operation data of the memory using the pre-trained motion model; 基于所述实际运行数据中硬盘的容量信息和传输速率信息中的至少一个,利用所述预先训练的动作模型输出所述硬盘的实际运行数据和所述硬盘的理论运行数据的硬盘差异值;Based on at least one of the capacity information and the transmission rate information of the hard disk in the actual operation data, outputting a hard disk difference value between the actual operation data of the hard disk and the theoretical operation data of the hard disk using the pre-trained motion model; 基于所述实际运行数据中网络的传输速率信息,利用所述预先训练的动作模型输出所述网络的实际运行数据和所述网络的理论运行数据的网络差异值;Based on the transmission rate information of the network in the actual operation data, using the pre-trained action model to output a network difference value between the actual operation data of the network and the theoretical operation data of the network; 基于所述实际运行数据中算力卡的算力数据和传输速率信息中的至少一个,利用所述预先训练的动作模型输出所述算力卡的实际运行数据和所述算力卡的理论运行数据的算力卡差异值;Based on at least one of the computing power data and the transmission rate information of the computing power card in the actual operation data, outputting a computing power card difference value between the actual operation data of the computing power card and the theoretical operation data of the computing power card using the pre-trained action model; 基于所述处理器差异值、所述内存差异值、所述硬盘差异值、所述网络差异值和所述算力卡差异值中的至少一个,得到所述第一差异值。The first difference value is obtained based on at least one of the processor difference value, the memory difference value, the hard disk difference value, the network difference value, and the computing power card difference value. 7.根据权利要求1所述的服务器性能的评估方法,其特征在于,在将所述第一差异值输入至预先建立的奖励模型中之前,还包括:7. The method for evaluating server performance according to claim 1, characterized in that before inputting the first difference value into a pre-established reward model, it further comprises: 基于所述第一差异值中的处理器差异值,利用所述处理器差异值中的任务信息、线程信息和指令信息建立奖励模型中的处理器奖励模型;Based on the processor difference value in the first difference value, establishing a processor reward model in the reward model using task information, thread information and instruction information in the processor difference value; 基于所述第一差异值中的内存差异值,利用所述内存差异值中的数据信息和带宽信息建立所述奖励模型中的内存奖励模型;Based on the memory difference value in the first difference value, establishing a memory reward model in the reward model by using data information and bandwidth information in the memory difference value; 基于所述第一差异值中的硬盘差异值,利用所述硬盘差异值中的带宽信息建立所述奖励模型中的硬盘奖励模型;Based on the hard disk difference value in the first difference value, a hard disk reward model in the reward model is established by using bandwidth information in the hard disk difference value; 基于所述第一差异值中的网络差异值,利用所述网络差异值中的传输数据信息建立所述奖励模型中的网络奖励模型;Based on the network difference value in the first difference value, establishing a network reward model in the reward model using the transmission data information in the network difference value; 基于所述第一差异值中的算力卡差异值,利用所述算力卡差异值中的算力信息建立所述奖励模型中的算力卡奖励模型;Based on the computing power card difference value in the first difference value, establishing a computing power card reward model in the reward model using computing power information in the computing power card difference value; 基于所述第一差异值中的能耗信息,建立所述奖励模型中的能耗奖励模型;Based on the energy consumption information in the first difference value, establishing an energy consumption reward model in the reward model; 基于所述处理器奖励模型、所述内存奖励模型、所述硬盘奖励模型、所述网络奖励模型、所述算力卡奖励模型和所述能耗奖励模型中的至少一个,建立所述奖励模型。The reward model is established based on at least one of the processor reward model, the memory reward model, the hard disk reward model, the network reward model, the computing power card reward model and the energy consumption reward model. 8.根据权利要求7所述的服务器性能的评估方法,其特征在于,所述将所述第一差异值输入至预先建立的奖励模型中,以输出所述至少一个部件在对应环境下的性能分数,包括:8. The method for evaluating server performance according to claim 7, wherein the step of inputting the first difference value into a pre-established reward model to output a performance score of the at least one component under a corresponding environment comprises: 基于所述第一差异值中的处理器差异值,利用所述预先建立的奖励模型中的处理器奖励模型计算所述服务器中处理器的处理器性能分数;Calculating a processor performance score of a processor in the server using a processor reward model in the pre-established reward model based on a processor difference value in the first difference value; 基于所述第一差异值中的内存差异值,利用所述预先建立的奖励模型中的内存奖励模型计算所述服务器中内存的内存性能分数;Based on the memory difference value in the first difference value, using the memory reward model in the pre-established reward model to calculate the memory performance score of the memory in the server; 基于所述第一差异值中的硬盘差异值,利用所述预先建立的奖励模型中的硬盘奖励模型计算所述服务器中硬盘的硬盘性能分数;Based on the hard disk difference value in the first difference value, using the hard disk reward model in the pre-established reward model to calculate the hard disk performance score of the hard disk in the server; 基于所述第一差异值中的网络差异值,利用所述预先建立的奖励模型中的网络奖励模型计算所述服务器中网络的网络性能分数;Based on the network difference value in the first difference value, using the network reward model in the pre-established reward model to calculate the network performance score of the network in the server; 基于所述第一差异值中的算力卡差异值,利用所述预先建立的奖励模型中的算力卡奖励模型计算所述服务器中算力卡的算力卡性能分数;Based on the computing power card difference value in the first difference value, using the computing power card reward model in the pre-established reward model to calculate the computing power card performance score of the computing power card in the server; 基于所述第一差异值中的能耗差异值,利用所述预先建立的奖励模型中的能耗奖励模型计算所述服务器的能耗性能分数。Based on the energy consumption difference value in the first difference value, the energy consumption performance score of the server is calculated using the energy consumption reward model in the pre-established reward model. 9.根据权利要求7所述的服务器性能的评估方法,其特征在于,其中,9. The method for evaluating server performance according to claim 7, wherein: 所述处理器奖励模型的表达式为:The processor reward model is expressed as: , 其中,为任务做指令数;为理论线程数 为理论线程频率;为芯片实际每周期可处理指令数;为实际完成时间;为需求线程数;in, Do the number of instructions for the task; is the theoretical number of threads , is the theoretical thread frequency; The number of instructions that the chip can actually process per cycle; is the actual completion time; is the number of threads required; 所述内存奖励模型的表达式为:The expression of the memory reward model is: , 其中,为数据总量;为理论带宽;为实际带宽;为理论容量;in, is the total amount of data; is the theoretical bandwidth; is the actual bandwidth; is the theoretical capacity; 所述硬盘奖励模型的表达式为:The expression of the hard disk reward model is: , 其中,为数据总量;为理论带宽;为实际带宽;为理论容量;in, is the total amount of data; is the theoretical bandwidth; is the actual bandwidth; is the theoretical capacity; 所述网络奖励模型的表达式为:The expression of the network reward model is: , 其中,为传输前数据大小;为实际收到的数据大小;in, is the data size before transmission; is the actual received data size; 所述算力卡奖励模型的表达式为:The expression of the computing power card reward model is: , 其中,为模型计算量;为实际算力值;in, is the model calculation amount; is the actual computing power value; 所述能耗奖励模型的表达式为:The expression of the energy consumption reward model is: , 其中,为理论能耗;为实际能耗值。in, is the theoretical energy consumption; is the actual energy consumption value. 10.根据权利要求1所述的服务器性能的评估方法,其特征在于,所述基于所述性能分数确定所述目标服务器的性能评估结果,包括:10. The method for evaluating server performance according to claim 1, wherein determining the performance evaluation result of the target server based on the performance score comprises: 基于所述性能分数,统计所述性能分数的总数;Based on the performance scores, counting the total number of the performance scores; 基于所述性能分数,计算所述目标服务器的初始性能评估结果;Based on the performance score, calculating an initial performance evaluation result of the target server; 基于所述总数和所述初始性能评估结果,确定所述性能评估结果。Based on the total number and the initial performance evaluation result, the performance evaluation result is determined. 11.一种服务器性能的评估装置,其特征在于,包括:11. A server performance evaluation device, comprising: 第一采集模块,用于采集目标服务器的至少一个部件在多个环境中的实际运行数据;A first collection module, used to collect actual operation data of at least one component of the target server in multiple environments; 第一输出模块,用于将所述实际运行数据输入至预先训练的动作模型中,以输出所述实际运行数据和对应理论运行数据之间的第一差异值;a first output module, configured to input the actual operation data into a pre-trained motion model to output a first difference value between the actual operation data and corresponding theoretical operation data; 第一确定模块,用于将所述第一差异值输入至预先建立的奖励模型中,以输出所述至少一个部件在对应环境下的性能分数,并基于所述性能分数确定所述目标服务器的性能评估结果。The first determination module is used to input the first difference value into a pre-established reward model to output a performance score of the at least one component in a corresponding environment, and determine a performance evaluation result of the target server based on the performance score. 12.根据权利要求11所述的服务器性能的评估装置,其特征在于,还包括:12. The server performance evaluation device according to claim 11, further comprising: 第二确定模块,用于在将所述实际运行数据输入至预先训练的动作模型之前,基于所述实际运行数据,确定动作模型中全连接神经网络的第一结构信息、批量归一化的第二结构信息和激活函数的函数信息中的至少一个;A second determination module is used to determine at least one of first structural information of a fully connected neural network, second structural information of batch normalization, and function information of an activation function in the action model based on the actual operation data before inputting the actual operation data into the pre-trained action model; 第一生成模块,用于基于所述第一结构信息、所述第二结构信息和所述函数信息中的至少一个,将所述全连接神经网络、所述批量归一化和所述激活函数进行串联,以得到所述动作模型中的第一动作网络;A first generating module is used for connecting the fully connected neural network, the batch normalization and the activation function in series based on at least one of the first structural information, the second structural information and the function information to obtain a first action network in the action model; 第二生成模块,用于将串联的至少两个所述第一动作网络和所述第一动作网络进行合并,以得到所述实际运行数据的第一特征;A second generating module, configured to merge at least two of the first action networks connected in series with the first action network to obtain a first feature of the actual operation data; 第三生成模块,用于基于所述第一特征和所述第一动作网络,得到所述动作模型中的第二动作网络;A third generating module, configured to obtain a second action network in the action model based on the first feature and the first action network; 第一构建模块,用于基于所述第一动作网络和所述第二动作网络,构建所述动作模型。The first construction module is used to construct the action model based on the first action network and the second action network. 13.根据权利要求11所述的服务器性能的评估装置,其特征在于,还包括:13. The server performance evaluation device according to claim 11, further comprising: 第二采集模块,用于在将所述实际运行数据输入至预先训练的动作模型之前,采集服务器的至少一个部件在多个环境中的训练运行数据;A second collection module, for collecting training operation data of at least one component of the server in multiple environments before inputting the actual operation data into the pre-trained motion model; 第二输出模块,用于将所述训练运行数据输入至预先建立的动作模型中,以输出所述训练运行数据和对应理论运行数据的第二差异值;A second output module, used for inputting the training operation data into a pre-established motion model to output a second difference value between the training operation data and the corresponding theoretical operation data; 计算模块,用于基于所述第二差异值,利用预先建立的奖励模型计算所述至少一个部件在对应环境下的训练性能分数;a calculation module, configured to calculate a training performance score of the at least one component in a corresponding environment based on the second difference value using a pre-established reward model; 检测模块,用于基于所述训练性能分数,检测所述至少一个部件是否满足预设性能条件;a detection module, configured to detect whether the at least one component meets a preset performance condition based on the training performance score; 第四生成模块,用于在所述至少一个部件不满足所述预设性能条件的情况下,利用离散估计训练所述预先建立的动作模型的参数信息,以得到训练后的动作模型,并基于所述训练后的动作模型生成满足所述预设性能条件的部件。The fourth generation module is used to train the parameter information of the pre-established action model using discrete estimation to obtain a trained action model when at least one component does not meet the preset performance conditions, and generate a component that meets the preset performance conditions based on the trained action model. 14.一种服务器,其特征在于,包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述程序,以实现如权利要求1-10任一项所述的服务器性能的评估方法。14. A server, characterized in that it comprises: a memory, a processor, and a computer program stored in the memory and executable on the processor, wherein the processor executes the program to implement the server performance evaluation method according to any one of claims 1 to 10. 15.一种计算机可读存储介质,其上存储有计算机程序,其特征在于,该程序被处理器执行,以用于实现如权利要求1-10任一项所述的服务器性能的评估方法。15. A computer-readable storage medium having a computer program stored thereon, wherein the program is executed by a processor to implement the server performance evaluation method according to any one of claims 1 to 10.
CN202510838379.1A 2025-06-20 2025-06-20 Evaluation method and device for server performance Pending CN120353681A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202510838379.1A CN120353681A (en) 2025-06-20 2025-06-20 Evaluation method and device for server performance

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202510838379.1A CN120353681A (en) 2025-06-20 2025-06-20 Evaluation method and device for server performance

Publications (1)

Publication Number Publication Date
CN120353681A true CN120353681A (en) 2025-07-22

Family

ID=96410006

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202510838379.1A Pending CN120353681A (en) 2025-06-20 2025-06-20 Evaluation method and device for server performance

Country Status (1)

Country Link
CN (1) CN120353681A (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030033122A1 (en) * 2001-03-15 2003-02-13 Dempsey Derek M. Performance assessment of data classifiers
CN116594713A (en) * 2023-05-30 2023-08-15 苏州浪潮智能科技有限公司 Method, device, equipment and medium for determining server configuration parameters
CN119493714A (en) * 2023-08-21 2025-02-21 华为技术有限公司 Server energy consumption prediction method, device, computing equipment and storage medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030033122A1 (en) * 2001-03-15 2003-02-13 Dempsey Derek M. Performance assessment of data classifiers
CN116594713A (en) * 2023-05-30 2023-08-15 苏州浪潮智能科技有限公司 Method, device, equipment and medium for determining server configuration parameters
CN119493714A (en) * 2023-08-21 2025-02-21 华为技术有限公司 Server energy consumption prediction method, device, computing equipment and storage medium

Similar Documents

Publication Publication Date Title
CN111913803B (en) Service load fine granularity prediction method based on AKX hybrid model
CN119862029B (en) Edge computing resource allocation optimization method and system based on reinforcement learning
CN109347668A (en) A training method and device for a service quality assessment model
CN102184746B (en) Storage performance testing system and method based on particle swarm optimization parameter
CN118569814B (en) A smart gas enterprise user supervision method, Internet of Things system and medium
CN110191015B (en) CPI index-based cloud service performance intelligent prediction method and device
CN118885291A (en) Real-time CPU load balancing method and system for cloud computing platform based on reinforcement learning
CN116909712A (en) Intelligent task scheduling system and method based on machine learning
CN113326132B (en) An information adjustment method, device and storage medium
CN120104352A (en) Computing power node load balancing method, device, equipment and storage medium
CN119759592B (en) Dynamic load balancing method and system for multi-core heterogeneous ASIC computing motherboard
CN119988017A (en) A computing power analysis method, system and storage medium for shared servers
CN119645778A (en) Log parsing adaptive optimization method and system based on swarm intelligence
CN119537036B (en) A parallel processing architecture optimization method and system for test equipment
CN119292892B (en) Model effect evaluation system and method based on quantitative reasoning
CN120353681A (en) Evaluation method and device for server performance
de Almeida et al. Using machine learning and in-band network telemetry for service metrics estimation
Mays et al. Decentralized data allocation via local benchmarking for parallelized mobile edge learning
CN118069369A (en) A method for integrating multiple business systems using cloud computing server
CN118953111A (en) A charging pile control method and system under dynamic load
CN118195160B (en) A method, device and computer equipment for training a power grid section control model
CN119766736B (en) A network congestion control method, system, device and medium
CN119561829A (en) Multi-agent based deep reinforcement learning method and system
CN120196440A (en) Normalized measurement method for multi-dimensional index system of heterogeneous computing resources
US20220390933A1 (en) Pattern recognition device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination