Patent classifications
G06V10/424
COMPUTER ARCHITECTURE FOR MAPPING ANALOG DATA VALUES TO A STRING CORRELITHM OBJECT IN A CORRELITHM OBJECT PROCESSING SYSTEM
A string correlithm object generator is configured to output a string correlithm object comprising a plurality of sub-string correlithm objects. A node is configured to receive a plurality of data values. A memory is configured to store a node table that associates sub-string correlithm objects with the data values such that a first sub-string correlithm object is associated with a first data value and a second sub-string correlithm object is associated with a second data value. A processor is configured to receive a third data value that is between the first data value and the second data value, determine a third sub-string correlithm object that is interpolated between the first sub-string correlithm object and the second sub-string correlithm object, and associate the third sub-string correlithm object with the third data value.
Method and system for processing candidate strings generated by an optical character recognition process
A method and system of recognizing a string of characters in a target image. An acquired target image is analyzed using an optical character recognition process to identify a candidate string, the candidate string having an associated plurality of character positions, each character position being associated with a set of one or more candidate characters located at related positions in the target image. A minimum edit cost is determined between the candidate string and a template for an output string. Determining the minimum edit cost includes, for a given character position in the candidate string and a given output character position in the template, identifying, among the respective set of candidate characters of the candidate string, a subset of candidate characters that satisfy the respective character set of the template; and determining an edit cost based on the recognition score of one of the candidate characters belonging to the identified subset of candidate characters. An output string corresponding to the determined minimum edit cost is returned.
Training image-recognition systems using a joint embedding model on online social networks
In one embodiment, a method includes identifying a shared visual concept in visual-media items based on shared visual features in images of the visual-media items; extracting, for each of the visual-media items, n-grams from communications associated with the visual-media item; generating, in a d-dimensional space, an embedding for each of the visual-media items at a location based on the visual concepts included in the visual-media item; generating, in the d-dimensional space, an embedding for each of the extracted n-grams at a location based on a frequency of occurrence of the n-gram in the communications associated with the visual-media items; and associating, with the shared visual concept, the extracted n-grams that have embeddings within a threshold area of the embeddings for the identified visual-media items.
Pattern recognition device, pattern recognition method, and computer program product
According to an embodiment, a pattern recognition device is configured to divide an input signal into a plurality of elements, convert the divided elements into feature vectors having the same dimensionality to generate a set of feature vectors, and evaluate the set of feature vectors using a recognition dictionary including models corresponding to respective classes, to output a recognition result representing a class or a set of classes to which the input signal belongs. The models each include sub-models each corresponding to one of possible division patterns in which a signal to be classified into a class corresponding to the model can be divided into a plurality of elements. A label expressing a model including a sub-model conforming to the set of feature vectors, or a set of labels expressing a set of models including sub-models conforming to the set of feature vectors is output as the recognition result.
Textual representation of an image
At least a computer-implemented method and an apparatus for processing an image are described. In examples, numeric values for at least one property of the image are determined. These values are then converted into at least one corresponding text character, said conversion being independent of any text content within the image. This enables a text representation of the image to be generated that contains said plurality of text characters. This text representation may be used to index and search for the image.
Training Image-Recognition Systems Using a Joint Embedding Model on Online Social Networks
In one embodiment, a method includes identifying a shared visual concept in visual-media items based on shared visual features in images of the visual-media items; extracting, for each of the visual-media items, n-grams from communications associated with the visual-media item; generating, in a d-dimensional space, an embedding for each of the visual-media items at a location based on the visual concepts included in the visual-media item; generating, in the d-dimensional space, an embedding for each of the extracted n-grams at a location based on a frequency of occurrence of the n-gram in the communications associated with the visual-media items; and associating, with the shared visual concept, the extracted n-grams that have embeddings within a threshold area of the embeddings for the identified visual-media items.
EXPLORATION AND PRODUCTION DOCUMENT CONTENT AND METADATA SCANNER
A method involves extracting, from a file comprising an unstructured oilfield document, terms, calculating term frequency inverse document frequency (TF-IDF) of the terms to generate an input vector, execute a document content classification model on the input vector to generate a document content classification of unstructured oilfield document, and extract table information from a table in the unstructured oilfield document. The method further involves storing, with the file in storage, the document content classification and the table information.
Measuring semantic and syntactic similarity between grammars according to distance metrics for clustered data
The disclosure relates to various distance metrics that may quantify semantic and syntactic relationships between devices. More particularly, a first grammar associated with a first device and a second grammar associated with a second device may each comprise a symbol sequence that re-expresses one or more sequenced data items and one or more rules that represent a repeated pattern in the symbol sequence. Accordingly, one or more distance metrics that quantify a similarity between the first grammar and the second grammar may be calculated according to a comparison between the rules in the first grammar and the rules in the second grammar such that a relationship between the first device and the second device can be determined according to the one or more distance metrics.
Training image-recognition systems using a joint embedding model on online social networks
In one embodiment, a method includes identifying a shared visual concept in visual-media items based on shared visual features in images of the visual-media items; extracting, for each of the visual-media items, n-grams from communications associated with the visual-media item; generating, in a d-dimensional space, an embedding for each of the visual-media items at a location based on the visual concepts included in the visual-media item; generating, in the d-dimensional space, an embedding for each of the extracted n-grams at a location based on a frequency of occurrence of the n-gram in the communications associated with the visual-media items; and associating, with the shared visual concept, the extracted n-grams that have embeddings within a threshold area of the embeddings for the identified visual-media items.
Training Image-Recognition Systems Using a Joint Embedding Model on Online Social Networks
In one embodiment, a method includes identifying a shared visual concept in visual-media items based on shared visual features in images of the visual-media items; extracting, for each of the visual-media items, n-grams from communications associated with the visual-media item; generating, in a d-dimensional space, an embedding for each of the visual-media items at a location based on the visual concepts included in the visual-media item; generating, in the d-dimensional space, an embedding for each of the extracted n-grams at a location based on a frequency of occurrence of the n-gram in the communications associated with the visual-media items; and associating, with the shared visual concept, the extracted n-grams that have embeddings within a threshold area of the embeddings for the identified visual-media items.