Absstract of: WO2025212906A1
A method of generating a set of predictions associated with a striking-based sporting event using an axial transformer neural network, the method including: receiving an input tuple, including a set of tensors representing game context, team strength, player strength, live team features, live player features, game events, and a super feature; inputting the input tuple into an axial transformer neural network by inputting each tensor from the set of tensors within a corresponding initial embedding layer; concatenating the initial embedding layers to form a single tensor; applying self-attention to the single tensor through axial transformer layers of the axial transformer neural network; mapping output embeddings from the axial transformer layers to target layers; and generating a set of target metric predictions for each of a set of players, one or more teams, and a match, based on the output embeddings from the target layers.
Absstract of: US2025315643A1
A method of generating a set of predictions associated with a possession-based sporting event using an axial transformer neural network, the method including: receiving an input tuple, including a set of tensors representing game context, team strength, player strength, live team features, live player features, game events, and a super feature; inputting the input tuple into an axial transformer neural network by inputting each tensor from the set of tensors within a corresponding initial embedding layer; concatenating the initial embedding layers to form a single tensor; applying self-attention to the single tensor; mapping output embeddings from the axial transformer layers to target layers, each of the output embeddings being of a dimension of a target metric; and generating a set of target metric predictions for each of a set of players, one or more teams, and a match, based on the output embeddings from the target layers.
Absstract of: US2025315695A1
A method of generating a set of predictions associated with position-based sporting events using an axial transformer neural network, the method including: receiving an input tuple, including a set of tensors representing game context, team strength, player strength, live team features, live player features, game events, and a super feature; inputting the input tuple into an axial transformer neural network by inputting each tensor from the set of tensors within a corresponding initial embedding layer; concatenating the initial embedding layers to form a single tensor; applying self-attention to the single tensor through axial transformer layers of the axial transformer neural network; mapping output embeddings from the axial transformer layers to target layers; and generating a set of target metric predictions for each racer, team, and overall for the position-based sporting events, based on the output embeddings from the target layers.
Absstract of: US2025307729A1
A system and method for variational annealing to solve financial optimization problems is provided. The financial optimization problem is encoded as objective function represented in terms of an energy function. An autoregressive neural network is trained to minimize the cost function via variational emulation of classical or quantum annealing. Optimal solutions to the financial optimization problem are obtained after a stopping criterion is set. An optimal solution may be selected according to user defined metrics, and optionally applied to a real-world system associated with the financial optimization problem.
Absstract of: US2025305834A1
The disclosed systems and techniques facilitate efficient detection and navigation of reduced drivability areas in driving environments. The disclosed techniques include, obtaining, using a sensing system of a vehicle, a set of camera images, a set of radar images, and/or a set of lidar images of an environment. The techniques further include generating, using a first neural network (NN), camera feature(s) characterizing the camera images, generating, using a second NN, radar features characterizing the radar images, and/or generating, using a third NN, lidar feature(s) characterizing the lidar images. The techniques further include processing the camera feature(s), the radar feature(s), and the lidar feature(s) to obtain an indication of a reduced drivability area in the environment.
Absstract of: US2025307604A1
Computer systems and computer-implemented methods train a neural network, by:(a) computing for each datum in a set of training data, activation values for nodes in the neural network and estimates of partial derivatives of an objective function for the neural network for the nodes in the neural network; (b) selecting a target node of the neural network and/or a target datum in the set of training data; (c) selecting a target-specific improvement model for the neural network, wherein the target-specific improvement model, when added to the neural network, improves performance of the neural network for the target node and/or the target datum, as the case may be; (d) training the target-specific improvement model; (e) merging the target-specific improvement model with the neural network to form an expanded neural network; and (f) training the expanded neural network.
Absstract of: AU2024204095A1
Typical classifiers must be trained on a large input sample to accurately classify inputs. In addition, if a new classification category needs to be added to a taxonomy after the classifier has already been trained to classify within the taxonomy, the classifier must be recreated and retrained to classify within the updated taxonomy. To address at least these technical problems with classifiers, a generative language model may be used to perform classification. A generative language model is a machine learning model that generates language, typically in the form of a textual response to a data input. A generative language model may utilize a large neural network to determine probabilities for a next token of a sequence of text conditional on previous or historical tokens in the sequence of text. An LLM is an example of a generative language model. Typical classifiers must be trained on a large input sample to accurately classify inputs. In addition, if a new classification category needs to be added to a taxonomy after the classifier has already been trained to classify within the taxonomy, the classifier must be recreated and retrained to classify within the updated taxonomy. To address at least these technical problems with classifiers, a generative language model may be used to perform classification. A generative language model is a machine learning model that generates language, typically in the form of a textual response to a data input. A generative language model may
Absstract of: US2025308121A1
In various examples, animations may be generated using audio-driven body animation synthesized with voice tempo. For example, full body animation may be driven from an audio input representative of recorded speech, where voice tempo (e.g., a number of phonemes per unit time) may be used to generate a 1D audio signal for comparing to datasets including data samples that each include an animation and a corresponding 1D audio signal. One or more loss functions may be used to compare the 1D audio signal from the input audio to the audio signals of the datasets, as well as to compare joint information of joints of an actor between animations of two or more data samples, in order to identify optimal transition points between the animations. The animations may then be stitched together—e.g., using interpolation and/or a neural network trained to seamlessly stitch sequences together—using the transition points.
Absstract of: US2025305836A1
In various examples, training sensor data generated by one or more sensors of autonomous machines may be localized to high definition (HD) map data to augment and/or generate ground truth data—e.g., automatically, in embodiments. The ground truth data may be associated with the training sensor data for training one or more deep neural networks (DNNs) to compute outputs corresponding to autonomous machine operations-such as object or feature detection, road feature detection and classification, wait condition identification and classification, etc. As a result, the HD map data may be leveraged during training such that the DNNs—in deployment—may aid autonomous machines in navigating environments safely without relying on HD map data to do so.
Absstract of: US2025308219A1
A user directed video generation method and system obtains a natural language-based communication from a user requesting that a computer-implemented system generate a virtual environment that is based on a description that is provided by the user. The description is interpreted by a trained neural network. Representations of pixel patterns are generated by a trained neural network in accordance with the interpretation. The representations of the pixel patterns are evaluated for consistency with context and then selected based on the evaluation. The selected pixel patterns are embodied in a video stream that is provided to the user. Natural language that may be in audio form may be generated to accompany the video stream.
Absstract of: US2025308512A1
A method of text-only and semi-supervised training for deliberation includes receiving training data including unspoken textual utterances that are each not paired with any corresponding spoken utterance of non-synthetic speech, and training a deliberation model that includes a text encoder and a deliberation decoder on the unspoken textual utterances. The method also includes receiving, at the trained deliberation model, first-pass hypotheses and non-causal acoustic embeddings. The first-pass hypotheses is generated by a recurrent neural network-transducer (RNN-T) decoder for the non-causal acoustic embeddings encoded by a non-causal encoder. The method also includes encoding, using the text encoder, the first-pass hypotheses generated by the RNN-T decoder, and generating, using the deliberation decoder attending to both the first-pass hypotheses and the non-causal acoustic embeddings, second-pass hypotheses.
Absstract of: WO2025207036A1
A method and apparatus for extracting commands between a controller and an operator, the method comprising: receiving a command and a response to the command; extracting an instruction from the command using a trained neural network model that processes text in a bidirectional manner; comparing the extracted instruction from the command and the response to the command to find text that is similar; and extracting text in the response to the command found to be most similar to the text in the extracted instruction of the command, wherein a fuzzy matching algorithm is used for comparing the extracted instruction of the command and the response to the command to find text that is similar, wherein the fuzzy matching algorithm compares a span of the extracted instruction of the command with a span of the response to the command and calculates a similarity score for the two compared spans.
Absstract of: EP4624980A1
The disclosed systems and techniques facilitate efficient detection and classification of traffic signs in driving environments. The disclosed techniques include, obtaining, using a sensing system of a vehicle, a set of camera images, a set of radar images, and a set of lidar images of an environment. The techniques further include generating, using a first neural network (NN), camera feature(s) characterizing the camera images, generating, using a second NN, radar features characterizing the radar images, and generating, using a third NN, lidar feature(s) characterizing the lidar images. The techniques further include processing the camera feature(s), the radar feature(s), and the lidar feature(s) to obtain an indication of a reduced drivability area in the environment.
Absstract of: EP4625249A1
A system and method for variational annealing to solve financial optimization problems is provided. The financial optimization problem is encoded as objective function represented in terms of an energy function. An autoregressive neural network is trained to minimize the cost function via variational emulation of classical or quantum annealing. Optimal solutions to the financial optimization problem are obtained after a stopping criterion is set. An optimal solution may be selected according to user defined metrics, and optionally applied to a real-world system associated with the financial optimization problem.
Absstract of: GB2639568A
Disclosed is a method of processing a query using a trained artificial neural network (ANN) implementing a large language model (LLM). A database of documents is maintained for use in processing queries, the database defining multiple repositories, each containing one or more documents. The documents in the database are processed to generate, for each repository, a vector store comprising vector embeddings encoding information obtained from documents of the repository. A user inputs a query string and a selection of one or more of the repositories to be used to process the query. A query embedding corresponding to the query string is generated and the vector stores corresponding to each selected repository are searched using the query embedding to identify one or more vectors that are relevant to the query. An LLM query is formulated to include the query string, a query context comprising information determined based on the identified relevant vectors and a predefined LLM prompt. The LLM is invoked with the LLM query as input whereby the LLM query is processed using the ANN to generate an LLM output. A query response is provided to the user based on query response data received from the LLM.
Absstract of: US2025299067A1
A system and method for automatically providing a bank agent with questions to ask a client of the bank based on known information about the client and answers to previous questions provided to the client, and then providing a financial solution or product that may help the client. The method includes asking the client an initial question, providing an answer by the client to the initial question, providing a follow-up question in response to the answer provided to the initial question that is generated by a machine learning model in a processor, accepting an answer to the follow-up question, and providing additional follow-up questions in response to previous questions and answers that are generated by the machine learning model, where the machine learning model uses at least one neural network having nodes that have been trained to provide the questions based on the previous questions and answers.
Absstract of: US2025299041A1
An adapter layer may be used to customize a machine learning component by transforming data flowing into, out of, and/or within the machine learning component. The adapter layer may include a number of neural network components, or “adapters,” configured to perform a transformation on input data. Neural network components may be configured into adapter groups. A router component can, based on the input data, select one or more neural network components for transforming the input data. The input layer may combine the results of any such transformations to yield adapted data. Different adapter groups can include adapters of different complexity (e.g., involving different amounts of computation and/or latency). Thus, the amount of computation or latency added by an adapter layer can be reduced for simpler transformations of the input data.
Absstract of: US2025299032A1
In an example, an apparatus comprises a compute engine comprising a high precision component and a low precision component; and logic, at least partially including hardware logic, to receive instructions in the compute engine; select at least one of the high precision component or the low precision component to execute the instructions; and apply a gate to at least one of the high precision component or the low precision component to execute the instructions. Other embodiments are also disclosed and claimed.
Absstract of: US2025299066A1
A method and related system for efficiently capturing relationships between event feature values in embeddings includes flattening an event sequence into a feature sequence including a first event prefix, a second event prefix, and a first set of feature values. The method includes generating an attention mask including first mask indicators to associate the first set of feature values with each other and second mask indicator to associate a first feature value of the first set of feature values with the second event prefix. The method includes providing the feature sequence and the attention mask to a self-attention neural network model to generate an embedding.
Absstract of: WO2025199173A1
A method and related system for efficiently capturing relationships between event feature values in embeddings includes flattening an event sequence into a feature sequence including a first event prefix, a second event prefix, and a first set of feature values. The method includes generating an attention mask including first mask indicators to associate the first set of feature values with each other and second mask indicator to associate a first feature value of the first set of feature values with the second event prefix. The method includes providing the feature sequence and the attention mask to a self-attention neural network model to generate an embedding.
Absstract of: US2025299051A1
An information processing apparatus configured to execute inference using a convolutional neural network, including: an obtainment unit configured to obtain target data from data for inference inputted in the information processing apparatus; and a computation unit configured to execute convolutional computation and output computation result data, the convolutional computation using computation data including the target data obtained by the obtainment unit and margin data different from the target data that is required to obtain the computation result data in a predetermined size, in which the obtainment unit obtains first data, which is a part of the margin data, from a data group existing around the target data separately from the target data in the data for inference and doses not obtain second data, which is the margin data except the first data, from the data group.
Absstract of: US2025299295A1
Apparatuses, systems, and techniques to enhance video are disclosed. In at least one embodiment, one or more neural networks are used to create a higher resolution video using upsampled frames from a lower resolution video.
Absstract of: EP4621769A2
A computing system is configured to generate a transformer-transducer-based deep neural network. The transformer-transducer-based deep neural network comprises a transformer encoder network and a transducer predictor network. The transformer encoder network has a plurality of layers, each of which includes a multi-head attention network sublayer and a feed-forward network sublayer. The computing system trains an end-to-end (E2E) automatic speech recognition (ASR) model, using the transformer-transducer-based deep neural network. The E2E ASR model has one or more adjustable hyperparameters that are configured to dynamically adjust an efficiency or a performance of E2E ASR model when the E2E ASR model is deployed onto a device or executed by the device.
Absstract of: WO2025191315A1
A computer-implemented method (200) for automated tuning of a stochastic spiking neural network (SSNN) for solving a combinatorial optimization problem (COP). The method includes (i) defining (205) a set of features of the COP. The method further includes (ii) building (210) the SSNN with an architecture based on the defined COP feature set. The method further includes (iii) selecting (215) a tunable set of parameters of the SSNN based on the architecture. The method further includes (iv) tuning (220) the selected tunable set of parameters using using a genetic algorithm - SSNN (GA-SSNN) model. The method further includes (v) implementing (225) the SSNN using the tuned set of parameters. The method further includes (vi) evaluating (230) the performance of the SSNN to determine whether a pre-determined criteria for solving the COP is met. The method further includes (vii) repeating (235) steps (iv) to (vi) until the performance meets the pre-determined criteria. The method further includes (viii) obtaining (240) the solution for solving the COP.
Nº publicación: US2025292362A1 18/09/2025
Applicant:
INTEL CORP [US]
Intel Corporation
Absstract of: US2025292362A1
Embodiments described herein provide techniques to facilitate hierarchical scaling when quantizing neural network data to a reduced-bit representation. The techniques includes operations to load a hierarchical scaling map for a tensor associated with a neural network, partition the tensor into a plurality of regions that respectively include one or more subregions based on the hierarchical scaling map, hierarchically scale numerical values of the tensor based on a first scale factor and second scale factor via the matrix accelerator circuitry, the first scale factor based on a statistical measure of a subregion of numerical values of within a region of the plurality of regions and the second scale factor based on a statistical measure of the region that includes the subregion, and generate a quantized representation of the tensor via quantization of hierarchically scaled numerical values.