Ministerio de Industria, Turismo y Comercio LogoMinisterior
 

Redes Neuronales

Resultados 35 resultados
LastUpdate Última actualización 27/07/2025 [07:20:00]
pdfxls
Solicitudes publicadas en los últimos 30 días / Applications published in the last 30 days
Resultados 1 a 25 de 35 nextPage  

PARTITIONING-BASED SCALABLE WEIGHTED AGGREGATION COMPOSITION FOR KNOWLEDGE GRAPH EMBEDDING

NºPublicación:  WO2025151950A1 24/07/2025
Solicitante: 
MASTERCARD TECH CANADA ULC [CA]
MASTERCARD TECHNOLOGIES CANADA ULC
WO_2025151950_PA

Resumen de: WO2025151950A1

The disclosure relates to methods and systems of partitioning-based scalable weighted aggregation composition for embeddings learned from knowledge graphs for training neural networks to perform downstream machine-learning tasks. For example, a system may access a knowledge graph comprising a plurality of nodes and partition the knowledge graph into a plurality of partitions based on edge densities between nodes of the knowledge graph. The system may perform partition-wise encoding using compositional message passing between nodes that enables learning from neighboring nodes. The system may generate an embedding for each node and each relation type in each partition based on the partition-wise encoding using compositional message passing. The system may concatenate the generated embeddings from the plurality of partitions. The system may train a global neural network for a downstream prediction task based on the concatenated embeddings using one or more weight matrices.

INTELLIGENT CONTROL WITH HIERARCHICAL STACKED NEURAL NETWORKS

NºPublicación:  US2025238673A1 24/07/2025
Solicitante: 
COMMONS MICHAEL LAMPORT [US]
Commons Michael Lamport
US_2024220797_PA

Resumen de: US2025238673A1

A system and method of detecting an aberrant message is provided. An ordered set of words within the message is detected. The set of words found within the message is linked to a corresponding set of expected words, the set of expected words having semantic attributes. A set of grammatical structures represented in the message is detected, based on the ordered set of words and the semantic attributes of the corresponding set of expected words. A cognitive noise vector comprising a quantitative measure of a deviation between grammatical structures represented in the message and an expected measure of grammatical structures for a message of the type is then determined. The cognitive noise vector may be processed by higher levels of the neural network and/or an external processor.

SYSTEMS AND METHODS TO IDENTIFY NEURAL NETWORK BRITTLENESS BASED ON SAMPLE DATA AND SEED GENERATION

NºPublicación:  US2025238288A1 24/07/2025
Solicitante: 
CAPITAL ONE SERVICES LLC [US]
Capital One Services, LLC
US_2024411619_PA

Resumen de: US2025238288A1

Systems and methods for determining neural network brittleness are disclosed. For example, the system may include one or more memory units storing instructions and one or more processors configured to execute the instructions to perform operations. The operations may include receiving a modeling request comprising a preliminary model and a dataset. The operations may include determining a preliminary brittleness score of the preliminary model. The operations may include identifying a reference model and determining a reference brittleness score of the reference model. The operations may include comparing the preliminary brittleness score to the reference brittleness score and generating a preferred model based on the comparison. The operations may include providing the preferred model.

DATA PROCESSING METHOD AND APPARATUS, AND STORAGE MEDIUM

NºPublicación:  US2025238658A1 24/07/2025
Solicitante: 
HUAWEI TECH CO LTD [CN]
HUAWEI TECHNOLOGIES CO., LTD
EP_4589473_PA

Resumen de: US2025238658A1

This application relates to a data processing method and apparatus, and a storage medium. The method includes: extracting a feature sequence of target data, where the feature sequence includes T input features, T is a positive integer, and t∈1, T; obtaining T hidden state vectors based on a recurrent neural network, where a tth hidden state vector is determined based on a (t−1)th input feature, a (t−1)th hidden state vector, and a (t−1)th extended state vector, and the (t−1)th extended state vector is obtained by performing lightweight processing based on the (t−1)th hidden state vector; and obtaining a processing result of the target data based on the T hidden state vectors by using a downstream task network.

GENERATIVE PIPELINE USING SINGLE SELFIE INPUT

NºPublicación:  US2025238902A1 24/07/2025
Solicitante: 
SNAP INC [US]
Snap Inc

Resumen de: US2025238902A1

The subject technology receives an input image, the input image comprising a selfie. The subject technology transforms, using a neural network, the input image to a latent representation of an identity. The subject technology transforms, using a diffusion model, a text condition to a second latent representation compatible with the latent representation of the identity. The subject technology transforms a pose template to a set of latent features for the diffusion model. The subject technology generates an intermediate image based on the latent representation of the identity, the second latent representation, and the set of latent features. The subject technology modifies, using a face enhancement network, the intermediate image based on the input image. The subject technology generates, using a face restoration network, a final output image based on the modified intermediate image. The subject technology provides for display the final output image on a display of a client device.

INDEPENDENT SPLIT MODEL INFERENCE IN SPLIT NEURAL NETWORK FOR ESTIMATING NETWORK PARAMETERS

NºPublicación:  EP4587963A1 23/07/2025
Solicitante: 
ERICSSON TELEFON AB L M [SE]
Telefonaktiebolaget LM Ericsson (publ)
WO_2024056547_PA

Resumen de: WO2024056547A1

A split neural network includes a tail network model (706) that receives a first plurality of activations and a second plurality of activations at a cut layer of the split neural network, and that generates a model output in response to the first plurality of activations and the second plurality of activations; a head network model (704) that receives a plurality of input feature values and generates the first plurality of activations in response to the plurality of input feature values and provides the first plurality of activations to the tail network model at the cut layer; and a translator model (708) that receives the first plurality of activations, that generates estimated values of the second plurality of activations in response to the first plurality of activations, and that provides the estimated values of the second plurality of activations to the tail network model at the cut layer.

ADAPTIVE VISUAL SPEECH RECOGNITION

NºPublicación:  US2025232762A1 17/07/2025
Solicitante: 
DEEPMIND TECH LIMITED [GB]
DeepMind Technologies Limited
JP_2025102754_PA

Resumen de: US2025232762A1

Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing video data using an adaptive visual speech recognition model. One of the methods includes receiving a video that includes a plurality of video frames that depict a first speaker; obtaining a first embedding characterizing the first speaker; and processing a first input comprising (i) the video and (ii) the first embedding using a visual speech recognition neural network having a plurality of parameters, wherein the visual speech recognition neural network is configured to process the video and the first embedding in accordance with trained values of the parameters to generate a speech recognition output that defines a sequence of one or more words being spoken by the first speaker in the video.

MULTIOBJECTIVE COEVOLUTION OF DEEP NEURAL NETWORK ARCHITECTURES

NºPublicación:  US2025232173A1 17/07/2025
Solicitante: 
COGNIZANT TECH SOLUTIONS U S CORPORATION [US]
Cognizant Technology Solutions U.S. Corporation
CA_3215345_A1

Resumen de: US2025232173A1

An evolutionary AutoML framework called LEAF optimizes hyperparameters, network architectures and the size of the network. LEAF makes use of both evolutionary algorithms (EAs) and distributed computing frameworks. A multiobjective evolutionary algorithm is used to maximize the performance and minimize the complexity of the evolved networks simultaneously by calculating the Pareto front given a group of individuals that have been evaluated for multiple objectives.

MULTIOBJECTIVE COEVOLUTION OF DEEP NEURAL NETWORK ARCHITECTURES

NºPublicación:  US2025232174A1 17/07/2025
Solicitante: 
COGNIZANT TECH SOLUTIONS U S CORPORATION [US]
Cognizant Technology Solutions U.S. Corporation
CA_3215345_A1

Resumen de: US2025232174A1

An evolutionary AutoML framework called LEAF optimizes hyperparameters, network architectures and the size of the network. LEAF makes use of both evolutionary algorithms (EAs) and distributed computing frameworks. A multiobjective evolutionary algorithm is used to maximize the performance and minimize the complexity of the evolved networks simultaneously by calculating the Pareto front given a group of individuals that have been evaluated for multiple objectives.

ACCELERATED TRAINING OF A NEURAL NETWORK VIA REGULARIZATION

NºPublicación:  US2025232175A1 17/07/2025
Solicitante: 
D5AI LLC [US]
D5AI LLC
US_2025068909_PA

Resumen de: US2025232175A1

A system and method for controlling a nodal network. The method includes estimating an effect on the objective caused by the existence or non-existence of a direct connection between a pair of nodes and changing a structure of the nodal network based at least in part on the estimate of the effect. A nodal network includes a strict partially ordered set, a weighted directed acyclic graph, an artificial neural network, and/or a layered feed-forward neural network.

Sparse Highly Connected Artificial Neural Network Architectures Involving Hybrid Scale Structure

NºPublicación:  US2025232176A1 17/07/2025
Solicitante: 
DRIBUS BENJAMIN FORREST [US]
Dribus Benjamin Forrest
US_2021073642_A1

Resumen de: US2025232176A1

A method of constructing geometry-induced sparse hybrid highly connected artificial neural network architectures comprising, selecting a geometry defined in terms of a manifold, selecting a direction of data flow in the geometry, selecting a node set as a finite subset of the geometry, partitioning the node set into layers with respect to the geometry and the direction of data flow, selecting an edge set consisting of edges between each node in each non-input layer of the layers and nodes in preceding layers of the layers, selecting one or more subgraphs of the resulting digraph, where each subgraph defines an individual geometry-induced sparse hybrid highly connected artificial neural network architecture with a hierarchy of edge length scales, implementing the sparse hybrid highly connected artificial neural network architectures with hierarchies of edge length scales concretely, and training the sparse hybrid highly connected artificial neural network architectures.

ELECTRONIC DEVICE FOR ACQUIRING IMAGE AND OPERATION METHOD THEREFOR

NºPublicación:  WO2025150734A1 17/07/2025
Solicitante: 
SAMSUNG ELECTRONICS CO LTD [KR]
\uC0BC\uC131\uC804\uC790 \uC8FC\uC2DD\uD68C\uC0AC
WO_2025150734_A1

Resumen de: WO2025150734A1

In an embodiment, an electronic device may comprise at least one processor, and a memory for storing one or more instructions. The memory may further store a first sub-neural network model and a second sub-neural network model. The one or more instructions may be executed by the at least one processor so as to cause the electronic device to: acquire N image frames; acquire feature information for each of the image frames by inputting the image frames one by one to the first sub-neural network model; and acquire an enhanced image by inputting, to the second sub-neural network model, integrated feature information obtained by merging the feature information for each of the image frames. The first sub-neural network model may be configured to receive one image frame and output feature information. The second sub-neural network model may be configured to output an enhanced image on the basis of input feature information.

LANGUAGE MODEL NEURAL NETWORKS FOR DOMAIN-SPECIFIC CONVERSATIONAL AGENTS

NºPublicación:  WO2025151795A1 17/07/2025
Solicitante: 
GOOGLE LLC [US]
GOOGLE LLC
WO_2025151795_PA

Resumen de: WO2025151795A1

Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training and use of a language model neural network as a domain-specific conversational agent for a particular domain.

INFERENCE METHOD AND APPARATUS FOR NEURAL NETWORK MODEL, AND RELATED DEVICE

NºPublicación:  EP4586109A1 16/07/2025
Solicitante: 
HUAWEI TECH CO LTD [CN]
Huawei Technologies Co., Ltd
EP_4586109_PA

Resumen de: EP4586109A1

This application provides an inference method for a neural network model. The method is applied to a computing cluster. The computing cluster includes a plurality of inference servers and a memory pool. Each inference server includes at least one inference card and a local memory. The method includes: A first inference card of a first inference server in the computing cluster receives an inference task. If a parameter for executing the inference task is not hit in the first inference card, the first inference card obtains the parameter from a local memory of the first server. If the parameter is not hit in the local memory of the first server, the first inference card obtains the parameter from the memory pool. The first inference card can execute the inference task based on all obtained parameters. A speed of obtaining the parameter by the first inference card can be improved based on a high-speed read/write capability of the local memory of the first inference server, to reduce a delay of obtaining the parameter by the first inference card, and meet a requirement for a low delay of executing the inference task. In addition, this application further provides a corresponding apparatus, computing cluster, and storage medium.

DATA-DRIVEN STATE ESTIMATION AND SYSTEM CONTROL UNDER UNCERTAINTY

NºPublicación:  WO2025146824A1 10/07/2025
Solicitante: 
MITSUBISHI ELECTRIC CORP [JP]
MITSUBISHI ELECTRIC CORPORATION
WO_2025146824_PA

Resumen de: WO2025146824A1

A control method for controlling an electro-mechanical system according to a task estimates the state of the system using an adaptive surrogate model of the system to produce an estimation of the state of the system. The adaptive surrogate model includes a neural network employing a weighted combination of neural ODEs of dynamics of the system in latent space, such that weights of the weighted combination of neural ODEs represent the uncertainty. The method controls the system according to the task based on the estimation of the state of the system and tunes the weights of the weighted combination of neural ODEs based on the controlling.

REMOTE DISTRIBUTION OF NEURAL NETWORKS

NºPublicación:  US2025225614A1 10/07/2025
Solicitante: 
SNAP INC [US]
Snap Inc
US_2025225614_PA

Resumen de: US2025225614A1

Remote distribution of multiple neural network models to various client devices over a network can be implemented by identifying a native neural network and remotely converting the native neural network to a target neural network based on a given client device operating environment. The native neural network can be configured for execution using efficient parameters, and the target neural network can use less efficient but more precise parameters.

INTEGRATION OF LEARNED DIFFERENTIABLE LOSS FUNCTIONS IN DEEP LEARNING MODELS

NºPublicación:  US2025225384A1 10/07/2025
Solicitante: 
MICROSOFT TECH LICENSING LLC [US]
Microsoft Technology Licensing, LLC
US_2025225384_PA

Resumen de: US2025225384A1

Systems and methods are disclosed herein for training a model with a learned loss function. In an example system, a first trained neural network is generated based on application of a first loss function, such as a predefined loss function. A set of values is extracted from one or more of the layers of the neural network model, such as the weights of one of the layers. A separate machine learning model is trained using the set of values and a set of labels (e.g., ground truth annotations for a set of data). The machine learning model outputs a symbolic equation based on the training. The symbolic equation is applied to the first trained neural network to generate a second trained neural network. In this manner, a learned loss function can be generated and used to train a neural network, resulting in improved performance of the neural network.

Inference Method and Apparatus for Neural Network Model, and Related Device

NºPublicación:  US2025225419A1 10/07/2025
Solicitante: 
HUAWEI TECH CO LTD [CN]
Huawei Technologies Co., Ltd
US_2025225419_PA

Resumen de: US2025225419A1

A method is applied to a computing cluster. The computing cluster includes a plurality of inference servers and a memory pool. Each inference server includes at least one inference card and a local memory. The method includes: a first inference card of a first inference server in the computing cluster receives an inference task. If a parameter for executing the inference task is not hit in the first inference card, the first inference card obtains the parameter from a local memory of the first server. If the parameter is not hit in the local memory of the first server, the first inference card obtains the parameter from the memory pool. The first inference card can execute the inference task based on all obtained parameters.

APPARATUS AND METHOD FOR MONITORING AND CONTROLLING OF A NEURAL NETWORK USING ANOTHER NEURAL NETWORK IMPLEMENTED ON ONE OR MORE SOLID-STATE CHIPS

NºPublicación:  US2025224724A1 10/07/2025
Solicitante: 
APEX AI IND LLC [US]
Apex AI Industries, LLC
US_2025224724_PA

Resumen de: US2025224724A1

A method of operating an apparatus using a control system that includes at least one neural network. The method includes receiving an input value captured by the apparatus, processing the input value using the at least one neural network of the control system implemented on first one or more solid-state chips, and obtaining an output from the at least one neural network resulting from processing the input value. The method may also include processing the output with another neural network implemented on solid-state chips to determine whether the output breaches a predetermined condition that is unchangeable after an initial installation onto the control system. The aforementioned another neural network is prevented from being retrained. The method may also include the step of using the output from the at least one neural network to control the apparatus unless the output breaches the predetermined condition. Similar corresponding apparatuses are described.

REFLEXIVE MODEL GRADIENT-BASED RULE LEARNING

NºPublicación:  US2025225386A1 10/07/2025
Solicitante: 
RAYTHEON COMPANY [US]
Raytheon Company
US_2025225386_PA

Resumen de: US2025225386A1

Systems, devices, methods, and computer-readable media for reflexive model generation and inference. A method includes receiving probabilistic rules of a reflexive model that correlate evidence with existence of an event of interest, training, based on ground truth examples of evidence and respective labels indicating whether the event of interest is/was/will be present or not, a neural network (NN) to encode the probabilistic rules and learn respective probabilities for the probabilistic rules, and providing, by the NN and responsive to new evidence, an output indicating a likelihood the event of interest exists.

METHODS AND SYSTEMS FOR CLASSIFICATION OF EGGS AND EMBRYOS USING MORPHOLOGICAL AND MORPHO-KINETIC SIGNATURE

NºPublicación:  US2025225798A1 10/07/2025
Solicitante: 
FAIRTILITY LTD [IL]
FAIRTILITY LTD
US_2025225798_PA

Resumen de: US2025225798A1

Methods and systems are described for classifying unfertilized eggs. For example, using control circuitry, first images of fertilized eggs can be received, and the first images can be labeled with known classifications. Using the control circuitry, an artificial neural network can be trained to detect the known classifications based on the first images of the fertilized eggs and a second image can be received of an unfertilized egg with an unknown classification. Using the control circuitry, the second image can be input into the trained artificial neural network and a prediction from the trained artificial neural network can be received that the second image corresponds to one or more of the known classifications.

Generation Of Graph-Based Dense Representations Of Events Of A Nodal Graph Through Deployment Of A Neural Network

NºPublicación:  US2025225183A1 10/07/2025
Solicitante: 
CISCO TECH INC [US]
Cisco Technology, Inc
US_2025225183_PA

Resumen de: US2025225183A1

A computerized method is disclosed that includes operations of receiving a plurality of alerts, generating a graph-based dense representation of each alert of the plurality of alerts including processing of each alert with a neural network, wherein a result of processing an individual alert by the neural network is a graph-based dense representation of the individual alert, computing relatedness scores between at least a subset of the plurality of alerts, and generating a graphical user interface illustrating a listing of at least a subset of the plurality of alerts, wherein the graphical user interface is configured to receive user input corresponding to selection of a first alert, wherein the graphical user interface is rendered on a display screen. Additionally, an additional operation may include training the neural network to produce graph-based dense representations, wherein the training is performed on a corpus of metapaths.

INTEGRATION OF LEARNED DIFFERENTIABLE LOSS FUNCTIONS IN DEEP LEARNING MODELS

NºPublicación:  EP4583004A1 09/07/2025
Solicitante: 
MICROSOFT TECHNOLOGY LICENSING LLC [US]
Microsoft Technology Licensing, LLC
EP_4583004_PA

Resumen de: EP4583004A1

Systems and methods are disclosed herein for training a model with a learned loss function. In an example system, a first trained neural network is generated based on application of a first loss function, such as a predefined loss function. A set of values is extracted from one or more of the layers of the neural network model, such as the weights of one of the layers. A separate machine learning model is trained using the set of values and a set of labels (e.g., ground truth annotations for a set of data). The machine learning model outputs a symbolic equation based on the training. The symbolic equation is applied to the first trained neural network to generate a second trained neural network. In this manner, a learned loss function can be generated and used to train a neural network, resulting in improved performance of the neural network.

METHOD AND SYSTEM FOR AUGMENTED SPEECH EMBEDDINGS BASED AUTOMATIC SPEECH RECOGNITION

NºPublicación:  EP4583099A1 09/07/2025
Solicitante: 
TATA CONSULTANCY SERVICES LTD [IN]
Tata Consultancy Services Limited
EP_4583099_PA

Resumen de: EP4583099A1

Though several data augmentation techniques have been explored in the signal or feature space, very few studies have explored augmentation in the embedding space for Automatic Speech Recognition (ASR). The outputs of the hidden layers of a neural network can be seen as different representations or projections of the features. The augmentations performed on the features may not necessarily translate into augmentation of the different projections of the features as obtained from the output of the hidden layers. To overcome the challenges of the conventional approaches, embodiments herein provide a method and system for augmented speech embeddings based automatic speech recognition. The present disclosure provides an augmentation scheme which works on the speech embeddings. The augmentation works by replacing a set of randomly selected embeddings by noise during training. It does not require additional data, works online during training, and adds very little to the overall computational cost.

CPU DEEP NEURAL NETWORK INFERENCE METHOD BASED ON IDLE CPU RESOURCES TO MAXIMIZE RESOURCE UTILIZATION IN LEARNING CLUSTERS

Nº publicación: KR20250102720A 07/07/2025

Solicitante:

성균관대학교산학협력단

WO_2025143765_PA

Resumen de: WO2025143765A1

A deep neural network inference method based on an idle CPU resource according to the present specification may comprise the steps of: when a training task is reserved and all GPUs are used, classifying a CPU core into a training task-specific group, and classifying an unallocated idle CPU core into an unallocated group (U group); executing a training task by using a CPU core of the training task-specific group and, when there is a request for an online inference task, executing the online inference task by using the idle CPU core of the U group; and when there is a request for a batch inference task, additionally executing the batch inference task by using at least one of the idle CPU core of the U group and an idle CPU core of the training task-specific group.

traducir