@gerrymcgovern I always hate to see these stats slurred across uses. The training of the model is a fixed cost. That output is used on inference servers that do the customer facing work. Those are still very powerful (hot), but the way the "each query costs... " really works is an economy of scale. Inference is one machine. Training is enormous clusters. More queries actually makes the cost of training go down per transaction.