Nothing Special   »   [go: up one dir, main page]

CN110263618B - Iteration method and device of nuclear body model - Google Patents

Iteration method and device of nuclear body model Download PDF

Info

Publication number
CN110263618B
CN110263618B CN201910361579.7A CN201910361579A CN110263618B CN 110263618 B CN110263618 B CN 110263618B CN 201910361579 A CN201910361579 A CN 201910361579A CN 110263618 B CN110263618 B CN 110263618B
Authority
CN
China
Prior art keywords
data
body model
training
model
nuclear body
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910361579.7A
Other languages
Chinese (zh)
Other versions
CN110263618A (en
Inventor
曹佳炯
丁菁汀
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Advanced New Technologies Co Ltd
Advantageous New Technologies Co Ltd
Original Assignee
Advanced New Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Advanced New Technologies Co Ltd filed Critical Advanced New Technologies Co Ltd
Priority to CN201910361579.7A priority Critical patent/CN110263618B/en
Publication of CN110263618A publication Critical patent/CN110263618A/en
Application granted granted Critical
Publication of CN110263618B publication Critical patent/CN110263618B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides an iteration method and device of a nuclear body model, wherein the iteration method of the nuclear body model comprises the following steps: desensitizing the received data, and extracting and screening characteristic data from the desensitized data; selecting super parameters from the characteristic data according to the state of the nuclear body model, and training the nuclear body model; and evaluating the trained nuclear body model, and after the evaluation is qualified, putting the iterative nuclear body model on line. The problem that the performance of a deep learning algorithm is affected by the change of data caused by the change of the distribution of test data and the distribution of training data due to different time is solved; meanwhile, the problem that the distribution of test data and the distribution of training data change due to different scenes, so that the performance of a deep learning algorithm is influenced due to the change of the data is solved; the method provided by the invention enables the updated model to be suitable for each scene, exerts the extreme performance of the algorithm, and improves the accuracy of verification of the kernel body of the algorithm.

Description

Iteration method and device of nuclear body model
Technical Field
The present invention relates to the field of computer technologies, and in particular, to an iteration method and apparatus for a kernel body model.
Background
Currently, a nuclear body system, such as a common face recognition system, comprises a plurality of modules including face detection, face calibration and face comparison. In these modules, almost all algorithms are implemented by deep learning models. Deep learning is a data-driven learning algorithm, i.e., the performance of the deep learning algorithm will perform better when the distribution of test data is close to the distribution of training data.
In practical application, with different application scenes of the algorithm, the distribution of the test data and the distribution of the training data form a larger or smaller gap, so that the practical performance of the algorithm is affected. On the other hand, even in the same scene, the distribution of the test data and the distribution of the training data change over time, which results in that the data change affects the performance of the deep learning algorithm and further affects the accuracy of the face recognition system.
Disclosure of Invention
The invention aims to provide an iteration method and device of a nuclear body model, which are used for solving the problem that the model verification is inaccurate due to the fact that the distribution of test data and the distribution of training data form a gap according to different scenes or time.
According to a first aspect of the present invention, there is provided an iterative method of a kernel body model, comprising:
desensitizing the received data, and extracting and screening characteristic data from the desensitized data;
selecting super parameters from the characteristic data according to the state of the nuclear body model, and training the nuclear body model;
and evaluating the trained nuclear body model, and after the evaluation is qualified, putting the iterative nuclear body model on line.
Further, the method of the present invention, after desensitizing the received data, further comprises: timing after desensitizing the received data, and judging whether the current service is in an idle state or not when the timing time is greater than or equal to a preset backflow time interval;
if yes, carrying out data reflow so as to execute the steps of extracting and screening characteristic data from the data after the desensitization treatment.
Further, the method of the present invention extracts and screens feature data from the desensitized data, including:
reprocessing the desensitized data;
extracting the reprocessed characteristic data;
and carrying out post-processing on the reprocessed characteristic data, and filtering out abnormal characteristic data to obtain a training sample.
Further, according to the method of the present invention, according to the state of the core body model, a hyper-parameter is selected from the feature data, and the core body model is trained, including:
and selecting super parameters from the training samples according to the state of the nuclear body model, and training the nuclear body model.
Further, the method of the present invention performs post-processing on the reprocessed feature data, filters out abnormal feature data, and obtains a training sample, including:
adopting a clustering algorithm, and filtering data which are far away from a clustering center by a preset distance as abnormal characteristic data;
and sampling the data with the abnormal characteristic data filtered according to a preset sample selection strategy to obtain a training sample.
Further, the method of the present invention, the evaluating the trained core body model includes:
and evaluating the trained nuclear body model according to the algorithm evaluation set and the algorithm precision index.
Further, the method of the present invention, wherein the reprocessing the data after the desensitization treatment includes:
and if the desensitized data is image data, decoding, scale adjustment and image normalization are carried out on the image data.
According to a second aspect of the present invention, there is provided an iteration apparatus of a nuclear body model, comprising:
the desensitization processing module is used for carrying out desensitization processing on the received data;
the characteristic data screening module is used for extracting and screening characteristic data from the data subjected to the desensitization treatment;
the model training module is used for selecting super parameters from the characteristic data according to the state of the nuclear body model and training the nuclear body model;
and the model online module is used for evaluating the trained nuclear body model and online the iterative nuclear body model after the evaluation is qualified.
Further, the device of the present invention, the desensitizing processing module, includes:
the judging unit is used for timing after the desensitization treatment is carried out on the received data, and judging whether the current service is in an idle state or not when the timing time is greater than or equal to a preset backflow time interval;
and the data reflow unit is used for carrying out data reflow when the current service is in an idle state so as to execute the steps of extracting and screening the characteristic data from the desensitized data.
Further, the device of the present invention, the feature data screening module includes:
a reprocessing unit for reprocessing the desensitized data;
a feature data extraction unit for extracting the reprocessed feature data;
the training sample acquisition unit is used for carrying out post-processing on the reprocessed characteristic data and filtering out abnormal characteristic data to obtain a training sample.
Further, the device of the invention, the model training module is used for:
and selecting super parameters from the training samples according to the state of the nuclear body model, and training the nuclear body model.
Further, the device of the present invention, the training sample obtaining unit is configured to:
adopting a clustering algorithm, and filtering data which are far away from a clustering center by a preset distance as abnormal characteristic data;
and sampling the data with the abnormal characteristic data filtered according to a preset sample selection strategy to obtain a training sample.
Further, according to the device disclosed by the invention, the model online module is used for:
and evaluating the trained nuclear body model according to the algorithm evaluation set and the algorithm precision index.
Further, the device of the present invention, the reprocessing unit is configured to:
and if the desensitized data is image data, decoding, scale adjustment and image normalization are carried out on the image data.
According to a third aspect of the present invention there is provided a storage medium storing computer program instructions for execution in accordance with the method of the present invention.
According to a fourth aspect of the present invention there is provided a computing device comprising: a memory for storing computer program instructions and a processor for executing the computer program instructions, wherein the computer program instructions, when executed by the processor, trigger the computing device to perform the method of the invention.
According to the iteration method and the iteration device of the nuclear body model, the received data are subjected to desensitization treatment, and the characteristic data are extracted and screened from the preprocessed data; the problem that the performance of a deep learning algorithm is affected by the change of data caused by the change of the distribution of test data and the distribution of training data due to different time is solved; according to the state of the nuclear body model, super parameters are selected from the characteristic data, and the nuclear body model is trained, so that the problem that the distribution of test data and the distribution of training data change due to different scenes, and the performance of a deep learning algorithm is influenced due to the change of the data is solved; the method provided by the invention enables the updated model to be suitable for each scene, exerts the extreme performance of the algorithm, and improves the accuracy of verification of the kernel body of the algorithm.
Drawings
Other features, objects and advantages of the present invention will become more apparent upon reading of the detailed description of non-limiting embodiments, made with reference to the accompanying drawings in which:
FIG. 1 is a flowchart illustrating an iterative method of a kernel model according to an embodiment of the present invention;
FIG. 2 is a flowchart illustrating an iterative method of a kernel model according to another embodiment of the present invention;
FIG. 3 is a flowchart illustrating an iterative method of a kernel model according to another embodiment of the present invention;
FIG. 4 is a schematic structural diagram of an iteration apparatus of a core-shell model according to an embodiment of the present invention;
fig. 5 is a schematic structural diagram of an iteration apparatus of a core-body model according to another embodiment of the present invention.
The same or similar reference numbers in the drawings refer to the same or similar parts.
Detailed Description
The invention is described in further detail below with reference to the accompanying drawings.
Fig. 1 is a flow chart of an iteration method of a core-body model according to an embodiment of the present invention, as shown in fig. 1, where the iteration method of a core-body model according to an embodiment of the present invention includes:
101. desensitizing the received data, and extracting and screening characteristic data from the desensitized data;
in the desensitization processing in this embodiment, in order to perform desensitization processing on received data, that is, data from different scenes, it may be understood that sensitive information in the received data is removed, for example, data desensitization processing, and risk of theft of user information is guaranteed through the desensitization processing.
102. Selecting super parameters from the characteristic data according to the state of the nuclear body model, and training the nuclear body model;
the state of the nuclear body model is various, including various numerical information such as loss function values, gradient values, and the like;
the super-parameters in the feature data include learning rate, momentum, etc.
103. And evaluating the trained nuclear body model, and after the evaluation is qualified, putting the iterative nuclear body model on line.
Because the performance of the model after automatic training is not evaluated, whether the model reaches the on-line performance standard is unknown, the automatic evaluation needs to be configured and corresponding parameter indexes need to be set for judgment.
Different performance indexes such as payment and login can be performed according to different service scenes, so that the trained core body model is online after the performance indexes under the required service scenes are evaluated to be qualified.
The method comprises the steps of performing desensitization treatment on received data, and extracting and screening characteristic data from the preprocessed data; the problem that the performance of a deep learning algorithm is affected by the change of data caused by the change of the distribution of test data and the distribution of training data due to different time is solved; according to the state of the nuclear body model, super parameters are selected from the characteristic data, and the nuclear body model is trained, so that the problem that the distribution of test data and the distribution of training data change due to different scenes, and the performance of a deep learning algorithm is influenced due to the change of the data is solved; the method provided by the invention enables the updated model to be suitable for each scene, exerts the extreme performance of the algorithm, and improves the accuracy of verification of the kernel body of the algorithm.
Fig. 2 is a flow chart of an iteration method of a core-body model according to an embodiment of the present invention, as shown in fig. 2, where the iteration method of a core-body model according to an embodiment of the present invention includes:
201. desensitizing the received data;
the desensitization processing in this embodiment is a reflow operation of desensitizing the received data.
The data desensitization, namely, the scene data may contain sensitive information of the user, including the face, the user name, transaction information and the like of the user, so that the data needs to be desensitized before the data is reflowed. The desensitization treatment includes, but is not limited to, the following: the method comprises (1) watermarking encryption processing of pictures; (2) anonymously processing key information such as user names; (3) transaction information deletion. After desensitization of the data, it is stored in the local device waiting for reflow.
Generally, after desensitizing the received data, the method further comprises: timing after desensitizing the received data, and judging whether the current service is in an idle state or not when the timing time is greater than or equal to a preset backflow time interval;
if yes, carrying out data reflux so as to execute the subsequent step, namely, the step of extracting and screening the characteristic data from the data after the desensitization treatment.
In addition, the data reflow has two modes of real-time reflow and asynchronous reflow, and the method of the embodiment adopts asynchronous data reflow. Mainly because automated iteration has no high requirement on data instantaneity. And the real-time data reflux can occupy the bandwidth greatly, so that the user experience of the main service is affected. Asynchronous data reflow can flexibly set a reflow time interval, and occupy redundant bandwidth to reflow data when the main service is idle.
Different time intervals may be used for automated deployment for different scenarios, and specific time intervals are not illustrated in detail in this embodiment.
202. Reprocessing the desensitized data, and extracting the reprocessed characteristic data;
reprocessing the desensitized data as described in step 202 above may include the steps of:
and if the desensitized data is image data, decoding, scale adjustment and image normalization are carried out on the image data.
The general above step 202 includes the sub-steps of:
the scene data is streamed back to the model server.
It will be appreciated that the model server is provided on the side receiving the scene data, and may be provided with the device receiving the scene data or may be provided on a different end.
The model server preprocesses the scene data, and when the scene data is image data, the model server preprocesses the scene data including image decoding, image scale adjustment, image normalization and the like.
And carrying out automatic feature extraction on the preprocessed image by a model server.
The whole flow of the method is automatically completed by a model server, and an algorithm developer cannot acquire image data and desensitization information of a user. The reprocessed data is used as the screening of the subsequent characteristic data, and is mainly used for guaranteeing the uniformity of the subsequent characteristic data.
203. And carrying out post-processing on the reprocessed characteristic data, and filtering out abnormal characteristic data to obtain a training sample.
The post-processing is performed on the reprocessed feature data in the step 203, and the abnormal feature data is filtered to obtain a training sample, which includes the following sub-steps:
2031. adopting a clustering algorithm, and filtering data which are far away from a clustering center by a preset distance as abnormal characteristic data;
after the scene data feature extraction is completed, post-processing is required to be carried out on the feature data, noise is filtered, and a sample is selected.
Because some noise exists in the scene data, a clustering algorithm is adopted to detect abnormal data. Specifically, when the feature data is distant from the cluster center by a certain distance, it is determined as an abnormal feature. The outlier feature will not be employed as a subsequent training sample.
2032. And sampling the data with the abnormal characteristic data filtered according to a preset sample selection strategy to obtain a training sample.
Since the reflowed scene data is quite large in sample number and is a simple sample for the existing model, the disadvantage of directly adding the scene data to the training data is greater than the advantage of training the model. The strategy of sample selection is mainly that the lines are sampled according to the algorithm corresponding to the features. Taking a face comparison algorithm as an example, the algorithm will sample from the dimension of the user id, for example every 10 features of the user.
204. Selecting super parameters from the characteristic data according to the state of the nuclear body model, and training the nuclear body model;
and selecting super parameters from the training samples according to the state of the nuclear body model, and training the nuclear body model.
The traditional model training method often needs to manually select a plurality of super parameters, and has the defects of low efficiency and high labor cost. Therefore, the method adopts an automatic training method based on reinforcement learning.
When super parameters are selected from training samples, reinforcement learning can adjust the super parameters (including learning rate, momentum and the like) of the model according to the states (including loss function values, gradient values and the like) of the model in the model training process. The method can find a better set of hyper-parameters in a huge hyper-parameter search space, and has proven to be more effective than manually adjusting parameters.
205. And evaluating the trained nuclear body model, and after the evaluation is qualified, putting the iterative nuclear body model on line.
The foregoing evaluation training core body model includes:
and evaluating the trained nuclear body model according to the algorithm evaluation set and the algorithm precision index.
Regarding model automation evaluation, since the performance of the model after automation training is not evaluated, whether the model reaches the on-line performance standard is unknown, the automatic evaluation needs to be configured and corresponding parameter indexes need to be set for judgment. The model automation evaluation mainly comprises two aspects of evaluation set selection and algorithm precision index.
Regarding the algorithm evaluation set: the algorithm evaluation set comprises a basic test set and a scene test set. The basic test set mainly tests the generalization capability of the algorithm, and the scene test set mainly tests the performance of the algorithm on the corresponding service scene.
Regarding algorithm accuracy indicators, the choice of algorithm accuracy indicators is related to a specific business scenario and a specific algorithm type. For example, for face alignment algorithms, the false recognition rate is a major performance indicator. Different business scenarios, such as payment and login, may have different performance indicators. And if the algorithm precision of the model obtained by automatic training reaches the index, an automatic online process is performed.
According to the method provided by the embodiment of the invention, the received data is subjected to desensitization treatment, and the characteristic data is extracted and screened from the preprocessed data; the problem that the performance of a deep learning algorithm is affected by the change of data caused by the change of the distribution of test data and the distribution of training data due to different time is solved; according to the state of the nuclear body model, super parameters are selected from the characteristic data, and the nuclear body model is trained, so that the problem that the distribution of test data and the distribution of training data change due to different scenes, and the performance of a deep learning algorithm is influenced due to the change of the data is solved; the method provided by the invention enables the updated model to be suitable for each scene, exerts the extreme performance of the algorithm, and improves the accuracy of verification of the kernel body of the algorithm.
Among the 5 steps described above, as shown in fig. 3, including 5 processes, the following can be simplified:
301. scene data desensitization reflow: different scenes collect different data, and after data desensitization, the data is refluxed to the model server.
302. Scene data feature extraction: the model server uses the current version of the model to perform automatic feature extraction on the reflow data for subsequent model training.
303. Scene data screening: since there is noise in the reflowed data features and not all samples are needed for the model, a data filter is set to filter the data.
304. Model automation training: the model super-parameters are automatically selected and the model is trained through an automatic machine learning technology.
305. Model automation evaluation is online: and (3) finishing automatic evaluation of the model by setting a performance evaluation experiment, and determining whether to be on line.
Through the five stages, the method can automatically mine the defect of the current algorithm from the desensitized scene data, and then make up for the defect through automatic model optimization iteration. Finally, the updated model can be better adapted to each scene, and the extreme performance of the algorithm is exerted. In addition, the method can be used for automatically updating the model by configuring reasonable time intervals, so that the influence of time on data distribution is overcome.
Fig. 4 is a schematic structural diagram of an iteration apparatus of a core-body model according to an embodiment of the present invention, as shown in fig. 4, where the iteration apparatus of a core-body model according to an embodiment of the present invention includes:
a desensitization processing module 41 for desensitizing the received data;
in the desensitization processing in this embodiment, in order to perform desensitization processing on received data, that is, data from different scenes, it may be understood that sensitive information in the received data is removed, for example, data desensitization processing, and risk of theft of user information is guaranteed through the desensitization processing.
A feature data screening module 42, configured to extract and screen feature data from the desensitized data;
the model training module 43 is configured to select a hyper-parameter from the feature data according to the state of the core-body model, and train the core-body model;
the state of the nuclear body model is various, including various numerical information such as loss function values, gradient values, and the like;
the super-parameters in the feature data include learning rate, momentum, etc.
The model online module 44 is configured to evaluate the trained core body model, and online the iterated core body model after the evaluation is qualified.
Because the performance of the model after automatic training is not evaluated, whether the model reaches the on-line performance standard is unknown, the automatic evaluation needs to be configured and corresponding parameter indexes need to be set for judgment.
Different performance indexes such as payment and login can be performed according to different service scenes, so that the trained core body model is online after the performance indexes under the required service scenes are evaluated to be qualified.
In the iteration device of the nuclear body model, the desensitization processing module is used for carrying out desensitization processing on the received data, and the characteristic data screening module is used for extracting and screening the characteristic data from the preprocessed data; the problem that the performance of a deep learning algorithm is affected by the change of data caused by the change of the distribution of test data and the distribution of training data due to different time is solved; the model training module is used for selecting super parameters from the characteristic data according to the state of the nuclear body model, so that the nuclear body model is trained, and the problem that the performance of a deep learning algorithm is affected due to the change of data caused by the change of distribution of test data and distribution of training data due to different scenes is solved; the method provided by the invention enables the updated model to be suitable for each scene, exerts the extreme performance of the algorithm, and improves the accuracy of verification of the kernel body of the algorithm.
Fig. 5 is a schematic structural diagram of an iteration apparatus of a core-body model according to an embodiment of the present invention, where, as shown in fig. 5, the iteration apparatus of a core-body model according to an embodiment of the present invention includes:
the desensitization processing module 51 includes:
the judging unit is used for timing after the desensitization treatment is carried out on the received data, and judging whether the current service is in an idle state or not when the timing time is greater than or equal to a preset backflow time interval;
and the data reflow unit is used for carrying out data reflow when the current service is in an idle state so as to execute the steps of extracting and screening the characteristic data from the desensitized data.
A feature data screening module 52, configured to reprocess the desensitized data, and extract the reprocessed feature data;
wherein the feature data screening module 52 includes,
a reprocessing unit 521 for reprocessing the desensitized data;
a feature data extraction unit 522, configured to extract the reprocessed feature data;
the training sample obtaining unit 523 is configured to post-process the reprocessed feature data, and filter the abnormal feature data to obtain a training sample.
In one embodiment of the present invention, the training sample acquiring unit 523 is configured to:
adopting a clustering algorithm, and filtering data which are far away from a clustering center by a preset distance as abnormal characteristic data;
and sampling the data with the abnormal characteristic data filtered according to a preset sample selection strategy to obtain a training sample.
The model training module 53 is configured to select a hyper-parameter from the feature data according to the state of the core body model, and train the core body model;
the model online module 54 is configured to:
and evaluating the trained nuclear body model according to the algorithm evaluation set and the algorithm precision index.
In one embodiment of the invention, the reprocessing unit is configured to:
and if the desensitized data is image data, decoding, scale adjustment and image normalization are carried out on the image data.
The apparatus shown in fig. 4 and fig. 5 of the present embodiment is an implementation apparatus of the method shown in fig. 1 and fig. 2 of the present embodiment, and the specific principle thereof is the same as that of the method shown in fig. 1 and fig. 2 of the present embodiment, and will not be repeated here.
In one embodiment of the invention, there is also provided a storage medium storing computer program instructions for execution in accordance with the method of an embodiment of the invention.
In one typical configuration of the invention, the computing devices each include one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
The memory may include volatile memory in a computer-readable medium, random Access Memory (RAM) and/or nonvolatile memory, etc., such as Read Only Memory (ROM) or flash memory (flashRAM). Memory is an example of computer-readable media.
In one embodiment of the present invention, there is also provided a computing device including: a memory for storing computer program instructions and a processor for executing the computer program instructions, wherein the computer program instructions, when executed by the processor, trigger the computing device to perform the method of embodiments of the invention.
Computer-readable storage media include both non-transitory and non-transitory, removable and non-removable media, and information storage may be implemented by any method or technology. The information may be computer readable instructions, data structures, program devices, or other data. Examples of storage media for a computer include, but are not limited to, phase change memory (PRAM), static Random Access Memory (SRAM), dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), read Only Memory (ROM), electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic tape storage or other magnetic storage devices, or any other non-transmission medium which can be used to store information that can be accessed by a computing device.
It should be noted that the present invention may be implemented in software and/or a combination of software and hardware, e.g., using Application Specific Integrated Circuits (ASIC), a general purpose computer or any other similar hardware device. In some embodiments, the software program of the present invention may be executed by a processor to implement the above steps or functions. Likewise, the software programs of the present invention (including associated data structures) may be stored on a computer readable recording medium, such as RAM memory, magnetic or optical drive or diskette and the like. In addition, some steps or functions of the present invention may be implemented in hardware, for example, as circuitry that cooperates with the processor to perform various steps or functions.
It will be evident to those skilled in the art that the invention is not limited to the details of the foregoing illustrative embodiments, and that the present invention may be embodied in other specific forms without departing from the spirit or essential characteristics thereof. The present embodiments are, therefore, to be considered in all respects as illustrative and not restrictive, the scope of the invention being indicated by the appended claims rather than by the foregoing description, and all changes which come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein. Any reference sign in a claim should not be construed as limiting the claim concerned. Furthermore, it is evident that the word "comprising" does not exclude other elements or steps, and that the singular does not exclude a plurality. A plurality of units or means recited in the apparatus claims can also be implemented by means of one unit or means in software or hardware. The terms first, second, etc. are used to denote a name, but not any particular order.

Claims (16)

1. An iterative method of a kernel model, comprising:
desensitizing received image data of users from different scenes, and extracting and screening feature data from the desensitized data when a preset reflux time interval is reached and the current service is idle;
post-processing is carried out on the characteristic data, and abnormal characteristic data is filtered to obtain a training sample;
selecting super parameters from the training samples according to the state of the nuclear body model, and automatically training the nuclear body model by adopting reinforcement learning; the super-parameters selected from the training samples are the super-parameters selected from the characteristic data contained in the training samples are adjusted according to the state of the core body model in the process of training the core body model by reinforcement learning;
and evaluating the trained nuclear body model, and after the evaluation is qualified, putting the iterative nuclear body model on line.
2. The method of claim 1, wherein after desensitizing the received data, the method further comprises:
timing after desensitizing the received data, and judging whether the current service is in an idle state or not when the timing time is greater than or equal to a preset backflow time interval;
if yes, carrying out data reflow so as to execute the steps of extracting and screening characteristic data from the data after the desensitization treatment.
3. The method of claim 1, wherein the extracting and screening feature data from the desensitized data comprises:
reprocessing the desensitized data;
extracting the reprocessed characteristic data;
and carrying out post-processing on the reprocessed characteristic data, and filtering out abnormal characteristic data to obtain a training sample.
4. A method according to claim 3, wherein selecting a hyper-parameter from the feature data according to the state of the core model, and training the core model comprises:
and selecting super parameters from the training samples according to the state of the nuclear body model, and training the nuclear body model.
5. A method according to claim 3, wherein post-processing the reprocessed feature data and filtering out abnormal feature data to obtain training samples comprises:
adopting a clustering algorithm, and filtering data which are far away from a clustering center by a preset distance as abnormal characteristic data;
and sampling the data with the abnormal characteristic data filtered according to a preset sample selection strategy to obtain a training sample.
6. The method of claim 1, wherein said evaluating the trained core body model comprises:
and evaluating the trained nuclear body model according to the algorithm evaluation set and the algorithm precision index.
7. A method according to claim 3, wherein said reprocessing of said desensitised data comprises:
and if the desensitized data is image data, decoding, scale adjustment and image normalization are carried out on the image data.
8. An iteration apparatus of a nuclear body model, comprising:
the desensitization processing module is used for carrying out desensitization processing on the received image data of the users from different scenes;
the characteristic data screening module is used for extracting and screening characteristic data from the desensitized data when a preset reflux time interval is reached and the current service is idle;
post-processing is carried out on the characteristic data, and abnormal characteristic data is filtered to obtain a training sample;
the model training module is used for selecting super parameters from the training samples according to the state of the nuclear body model, and automatically training the nuclear body model by reinforcement learning; the super-parameters selected from the training samples are the super-parameters selected from the characteristic data contained in the training samples are adjusted according to the state of the core body model in the process of training the core body model by reinforcement learning;
and the model online module is used for evaluating the trained nuclear body model and online the iterative nuclear body model after the evaluation is qualified.
9. The apparatus of claim 8, wherein the desensitizing processing module comprises:
the judging unit is used for timing after the desensitization treatment is carried out on the received data, and judging whether the current service is in an idle state or not when the timing time is greater than or equal to a preset backflow time interval;
and the data reflow unit is used for carrying out data reflow when the current service is in an idle state so as to execute the steps of extracting and screening the characteristic data from the desensitized data.
10. The apparatus of claim 8, wherein the feature data screening module comprises:
a reprocessing unit for reprocessing the desensitized data;
a feature data extraction unit for extracting the reprocessed feature data;
the training sample acquisition unit is used for carrying out post-processing on the reprocessed characteristic data and filtering out abnormal characteristic data to obtain a training sample.
11. The apparatus of claim 10, wherein the model training module is configured to:
and selecting super parameters from the training samples according to the state of the nuclear body model, and training the nuclear body model.
12. The apparatus of claim 10, wherein the training sample acquisition unit is configured to:
adopting a clustering algorithm, and filtering data which are far away from a clustering center by a preset distance as abnormal characteristic data;
and sampling the data with the abnormal characteristic data filtered according to a preset sample selection strategy to obtain a training sample.
13. The apparatus of claim 8, wherein the model online module is configured to:
and evaluating the trained nuclear body model according to the algorithm evaluation set and the algorithm precision index.
14. The apparatus of claim 10, wherein the reprocessing unit is configured to:
and if the desensitized data is image data, decoding, scale adjustment and image normalization are carried out on the image data.
15. A storage medium storing computer program instructions that are executed according to the method of any one of claims 1 to 7.
16. A computing device, comprising: a memory for storing computer program instructions and a processor for executing the computer program instructions, wherein the computer program instructions, when executed by the processor, trigger the computing device to perform the method of any of claims 1 to 7.
CN201910361579.7A 2019-04-30 2019-04-30 Iteration method and device of nuclear body model Active CN110263618B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910361579.7A CN110263618B (en) 2019-04-30 2019-04-30 Iteration method and device of nuclear body model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910361579.7A CN110263618B (en) 2019-04-30 2019-04-30 Iteration method and device of nuclear body model

Publications (2)

Publication Number Publication Date
CN110263618A CN110263618A (en) 2019-09-20
CN110263618B true CN110263618B (en) 2023-10-20

Family

ID=67913965

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910361579.7A Active CN110263618B (en) 2019-04-30 2019-04-30 Iteration method and device of nuclear body model

Country Status (1)

Country Link
CN (1) CN110263618B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110879865B (en) * 2019-10-31 2022-08-12 支付宝(杭州)信息技术有限公司 Recommendation method and device for nuclear products
CN112541067A (en) * 2020-12-15 2021-03-23 平安科技(深圳)有限公司 Knowledge base problem mining method and device, electronic equipment and storage medium

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101655908A (en) * 2009-08-03 2010-02-24 深圳市朗科科技股份有限公司 Subscriber verification method and apparatus
KR101254181B1 (en) * 2012-12-13 2013-04-19 위아코퍼레이션 주식회사 Face recognition method using data processing technologies based on hybrid approach and radial basis function neural networks
CN106204779A (en) * 2016-06-30 2016-12-07 陕西师范大学 The check class attendance method learnt based on plurality of human faces data collection strategy and the degree of depth
CN107506193A (en) * 2017-08-18 2017-12-22 平安科技(深圳)有限公司 A kind of development approach of subsystem page, storage medium and server
CN107992888A (en) * 2017-11-29 2018-05-04 深圳市智物联网络有限公司 The recognition methods of operation of industrial installation and server
CN108875530A (en) * 2018-01-12 2018-11-23 北京旷视科技有限公司 Vivo identification method, vivo identification equipment, electronic equipment and storage medium
CN108875404A (en) * 2018-05-30 2018-11-23 中国联合网络通信集团有限公司 Data desensitization method, device and storage medium based on machine learning
CN108920863A (en) * 2018-07-20 2018-11-30 湖南大学 A kind of method for building up of robot servo system energy consumption estimation model
CN109040091A (en) * 2018-08-17 2018-12-18 中科物栖(北京)科技有限责任公司 The encryption method and device of deep neural network model
CN109255340A (en) * 2018-10-29 2019-01-22 东北大学 It is a kind of to merge a variety of face identification methods for improving VGG network
CN109635112A (en) * 2018-12-13 2019-04-16 平安医疗健康管理股份有限公司 Abnormal dialysis data screening method, apparatus, equipment and storage medium
CN109634724A (en) * 2018-10-16 2019-04-16 深圳壹账通智能科技有限公司 Collecting method, device, equipment and computer storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2465782B (en) * 2008-11-28 2016-04-13 Univ Nottingham Trent Biometric identity verification

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101655908A (en) * 2009-08-03 2010-02-24 深圳市朗科科技股份有限公司 Subscriber verification method and apparatus
KR101254181B1 (en) * 2012-12-13 2013-04-19 위아코퍼레이션 주식회사 Face recognition method using data processing technologies based on hybrid approach and radial basis function neural networks
CN106204779A (en) * 2016-06-30 2016-12-07 陕西师范大学 The check class attendance method learnt based on plurality of human faces data collection strategy and the degree of depth
CN107506193A (en) * 2017-08-18 2017-12-22 平安科技(深圳)有限公司 A kind of development approach of subsystem page, storage medium and server
CN107992888A (en) * 2017-11-29 2018-05-04 深圳市智物联网络有限公司 The recognition methods of operation of industrial installation and server
CN108875530A (en) * 2018-01-12 2018-11-23 北京旷视科技有限公司 Vivo identification method, vivo identification equipment, electronic equipment and storage medium
CN108875404A (en) * 2018-05-30 2018-11-23 中国联合网络通信集团有限公司 Data desensitization method, device and storage medium based on machine learning
CN108920863A (en) * 2018-07-20 2018-11-30 湖南大学 A kind of method for building up of robot servo system energy consumption estimation model
CN109040091A (en) * 2018-08-17 2018-12-18 中科物栖(北京)科技有限责任公司 The encryption method and device of deep neural network model
CN109634724A (en) * 2018-10-16 2019-04-16 深圳壹账通智能科技有限公司 Collecting method, device, equipment and computer storage medium
CN109255340A (en) * 2018-10-29 2019-01-22 东北大学 It is a kind of to merge a variety of face identification methods for improving VGG network
CN109635112A (en) * 2018-12-13 2019-04-16 平安医疗健康管理股份有限公司 Abnormal dialysis data screening method, apparatus, equipment and storage medium

Also Published As

Publication number Publication date
CN110263618A (en) 2019-09-20

Similar Documents

Publication Publication Date Title
CN110059468B (en) Applet risk identification method and device
US9811760B2 (en) Online per-feature descriptor customization
CN112017056B (en) Intelligent double-recording method and system
CN109145981B (en) Deep learning automatic model training method and equipment
CN109086734A (en) The method and device that pupil image is positioned in a kind of pair of eye image
CN110263618B (en) Iteration method and device of nuclear body model
CN110490304B (en) Data processing method and device
CN114282586A (en) Data annotation method, system and electronic equipment
CN111753870B (en) Training method, device and storage medium of target detection model
CN111258905B (en) Defect positioning method and device, electronic equipment and computer readable storage medium
CN112508039B (en) Image detection method and device
CN110852443A (en) Feature stability detection method, device and computer readable medium
CN113240050A (en) Metal printing molten pool detection method with adjustable feature fusion weight
CN110428156B (en) Method and device for checking service data and electronic equipment
CN112561713A (en) Method and device for anti-fraud recognition of claim settlement in insurance industry
CN113935748B (en) Screening method, device, equipment and medium for spot check objects
CN117033205A (en) Flow playback comparison test method and device
CN116204648A (en) Sample determination method, sample determination device, computer equipment and computer readable storage medium
CN112148764B (en) Feature screening method, device, equipment and storage medium
CN116340127A (en) Interface testing method and device
CN113128594A (en) Optimization method and equipment of evaluation model based on cross-domain data
CN114666142B (en) Object authentication method, device and system
CN118132805A (en) Data management method, device, equipment and storage medium
Wu et al. Estimate the Precision of Defects Based on Reports Duplication in Crowdsourced Testing
CN116541713A (en) Bearing fault diagnosis model training method based on local time-frequency characteristic transfer learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20200923

Address after: Cayman Enterprise Centre, 27 Hospital Road, George Town, Grand Cayman Islands

Applicant after: Innovative advanced technology Co.,Ltd.

Address before: Cayman Enterprise Centre, 27 Hospital Road, George Town, Grand Cayman Islands

Applicant before: Advanced innovation technology Co.,Ltd.

Effective date of registration: 20200923

Address after: Cayman Enterprise Centre, 27 Hospital Road, George Town, Grand Cayman Islands

Applicant after: Advanced innovation technology Co.,Ltd.

Address before: A four-storey 847 mailbox in Grand Cayman Capital Building, British Cayman Islands

Applicant before: Alibaba Group Holding Ltd.

TA01 Transfer of patent application right
GR01 Patent grant
GR01 Patent grant