Hash based rollup with passthrough
    142.
    发明授权

    公开(公告)号:US11675767B1

    公开(公告)日:2023-06-13

    申请号:US17099467

    申请日:2020-11-16

    Abstract: A system includes a plurality of computing units. A first computing unit of the plurality of computing units comprises: a communication interface configured to receive an indication to roll up data in a data table; and a processor coupled to the communication interface and configured to: build a preaggregation hash table based at least in part on a set of columns and the data table by aggregating input rows of the data table; for each preaggregated hash table entry of the preaggregated hash table: provide the preaggregated hash table entry to a second computing unit of the plurality of computing units based at least in part on a distribution hash value; receive a set of received entries from computing units of the plurality of computing units; and build an aggregation hash table based at least in part on the set of received entries by aggregating the set of received entries.

    Model ML Registry and Model Serving
    143.
    发明公开

    公开(公告)号:US20230177031A1

    公开(公告)日:2023-06-08

    申请号:US18162579

    申请日:2023-01-31

    CPC classification number: G06F16/219 G06F16/955 G06N5/022

    Abstract: A system includes an interface, a processor, and a memory. The interface is configured to receive a version of a model from a model registry. The processor is configured to store the version of the model, start a process running the version of the model, and update a proxy with version information associated with the version of the model, wherein the updated proxy indicates to redirect an indication to invoke the version of the model to the process. The memory is coupled to the processor and configured to provide the processor with instructions.

    FEATURE STORE WITH INTEGRATED TRACKING

    公开(公告)号:US20220374457A1

    公开(公告)日:2022-11-24

    申请号:US17514997

    申请日:2021-10-29

    Abstract: The present application discloses a method, system, and computer system for managing a plurality of features and storing lineage information pertaining to the features. The method includes obtaining one or more datasets, determining a first feature, wherein the first feature is determined based at least in part on the one or more datasets, and storing the first feature in a feature store. The first feature is stored in association with a dataset indication of the one or more datasets from which the first feature is determined. The feature store comprises a plurality of features.

    DATAFLOW GRAPH PROCESSING
    145.
    发明申请

    公开(公告)号:US20220309103A1

    公开(公告)日:2022-09-29

    申请号:US17362450

    申请日:2021-06-29

    Abstract: A system for dataflow graph processing comprises a communication interface and a processor. The communication interface is configured receive an indication to generate a dataflow graph, wherein the indication includes a set of queries and/or commands. The processor is coupled to the communication interface and configured to: determine dependencies of each query in the set of queries on another query; determine a DAG of nodes based at least in part on the dependencies; determine the dataflow graph by determining in-line expressions for tables of the dataflow graph aggregating calculations associated with a subset of dataflow graph nodes designated as view nodes; and provide the dataflow graph.

    Split front end for flexible back end cluster processing

    公开(公告)号:US11113043B2

    公开(公告)日:2021-09-07

    申请号:US16864074

    申请日:2020-04-30

    Abstract: A system for code development and execution includes a client interface and a client processor. The client interface is configured to receive user code for execution and receive an indication of a server that will perform the execution. The client processor is configured to parse the user code to identify one or more data items referred to during the execution. The client processor is also configured to provide the server with an inquiry for metadata regarding the one or more data items, receive the metadata regarding the one or more data items, determine a logical plan based at least in part on the metadata regarding the one or more data items; and provide the logical plan to the server for execution.

    Directory level atomic commit protocol

    公开(公告)号:US11068447B2

    公开(公告)日:2021-07-20

    申请号:US15487896

    申请日:2017-04-14

    Abstract: A system for directory level atomic commits includes an interface and a processor. The interface is configured to receive an indication to provide a set of files. The processor is configured to determine whether a file in a directory has been either 1) atomically committed or 2) written by a non-atomic process and not designated as deleted and provide the file as one file of the set of files in the event that the file in the directory has been either 1) atomically committed or 2) written by a non-atomic process and not designated as deleted.

    Autoscaling using file access or cache usage for cluster machines

    公开(公告)号:US10810051B1

    公开(公告)日:2020-10-20

    申请号:US16188989

    申请日:2018-11-13

    Abstract: The allocation system comprises an interface and a processor. The interface is configured to receive an indication to deactivate idle cluster machines of a set of cluster machines. The processor is configured to determine a set of tasks executing or pending on the set of cluster machines; determine a set of idle cluster machines of the set of cluster machines that are neither running one or more tasks of the set of tasks nor storing one or more intermediate data files of a set of intermediate data files, where the set of intermediate data files is associated with a set of tasks executing or pending on the cluster machines; and deactivate each cluster machine of the set of idle cluster machines.

    SPLIT FRONT END FOR FLEXIBLE BACK END CLUSTER PROCESSING

    公开(公告)号:US20200301684A1

    公开(公告)日:2020-09-24

    申请号:US16864074

    申请日:2020-04-30

    Abstract: A system for code development and execution includes a client interface and a client processor. The client interface is configured to receive user code for execution and receive an indication of a server that will perform the execution. The client processor is configured to parse the user code to identify one or more data items referred to during the execution. The client processor is also configured to provide the server with an inquiry for metadata regarding the one or more data items, receive the metadata regarding the one or more data items, determine a logical plan based at least in part on the metadata regarding the one or more data items; and provide the logical plan to the server for execution.

    QUERY WATCHDOG
    150.
    发明申请
    QUERY WATCHDOG 审中-公开

    公开(公告)号:US20200241950A1

    公开(公告)日:2020-07-30

    申请号:US16793921

    申请日:2020-02-18

    Abstract: A system for monitoring job execution includes an interface and a processor. The interface is configured to receive an indication to start a cluster processing job. The processor is configured to determine whether processing a data instance associated with the cluster processing job satisfies a watchdog criterion; and in the event that processing the data instance satisfies the watchdog criterion, cause the processing of the data instance to be killed.

Patent Agency Ranking