Run pytesseract in a containerized transform and add the language packs in the Docker file, so that pytesseract can find and use them in the Python script.
Timestamp: February 13, 2024
Foundry ML Live does not currently support autoscaling for containers. However, autoscaling support is coming soon with scaling based on request throughput/queueing. The containers are managed in Kubernetes, not just as Docker containers in a Docker daemon, allowing for high availability with a fixed number of replicas/pods.
Timestamp: February 14, 2024
Tesseract library not found
error when calling a deployed model?The solution is to follow these steps:
Timestamp: February 14, 2024
foundry_ml
only supports up to Python 3.8 and they are concerned about the upcoming Python 3.8 end-of-life?The customer team should stick to Python 3.8 for now since foundry_ml
plans to increase support to Python 3.9 in near future.
Timestamp: February 13, 2024
It is possible to train language models on Foundry as Foundry enrollments have basic GPUs and deployment infrastructure. The feasibility depends on the specifics of the project such as data size, latency, and correctness required.
Timestamp: February 13, 2024
No, there is no cost per query for customers; the only cost is the compute cost while the deployment is running.
Timestamp: February 15, 2024
Foundry does not support AutoMLOps as AutoMLOps requires a different architecture and many AutoMLOps steps cannot be automated through code. However, you can perform many parts of AutoMLOps and can write specific code to reproduce the automated parts of AutoMLOps.
Timestamp: February 13, 2024
connection refused
error when deploying a model, and how can it be resolved?The connection refused
error is usually caused by the model adaptor script using the wrong port. The issue can be resolved by re-uploading the model with the correct port.
Timestamp: February 13, 2024
Currently, there is no way to set the Model API via the frontend similar to the Objective API in the Objectives UI. The solution is to retrain the model and publish a new model version or to copy the model adapter into a code repository to define the Model API.
Timestamp: February 13, 2024
The reason behind an input resolution error is likely due to a path and resource identifier (RIDs) issue within the repository, specifically if a model output path is reused in a different file. The issue can be resolved by ensuring paths are not reused in different files and, if needed, by re-creating all inputs and output RIDs.
Timestamp: February 13, 2024
The problem could be caused by a numpy type being outputted by the dictionary, which the Modeling Objective does not accept. Converting the numpy type to a float could resolve the issue.
Timestamp: February 13, 2024
No, there is no canonical built-in class for model segmentation in the new Model Assets. The new approach encourages users to write the adapter they need.
Timestamp: February 13, 2024
The metrics build failure with exit code 137 is likely caused by an Out Of Memory (OOM) issue. The resolution is to increase the Spark profile resources, specifically by using the MEMORY_LARGE Spark profiles.
Timestamp: February 14, 2024
No, the current API only supports LLM-like token stream output as it is implemented with SSE, which only supports text-based data.
Timestamp: February 14, 2024
It is expected to have different outputs during an upgrade because the load balancer will round-robin over all available replicas, which may include both replicas of the old release and the new release. Once the upgrade is complete only the new release should be answering for inference requests.
Timestamp: February 13, 2024
Yes, view permissions on the Objective are required to run inference.
Timestamp: March 12, 2024
This can happen if the evaluation dataset, input dataset, or model is not imported into the same project as the modeling objective. Ensure that the all of the inputs to the inference dataset are added as project references to the project where the modeling objective is located.
Timestamp: April 9, 2024
Direct deployments are not currently supported by Deployment Suite.
Timestamp: September 17, 2024