Question |
Answer |
---|---|
Can we use our on-premises machines with Valohai? |
Yes. Most of our on-premise workers are running Ubuntu 22.04 or higher, but we support other Linux distributions also. The machine will need Python 3.6 or higher, Docker, and the Valohai agent (“Peon”) installed, so it knows how to read jobs from Valohai and handle commands. Valohai doesn’t need direct access to the on-premises machine but the on-premises machine needs outbound access to your job queue virtual machine. |
What are Valohai workers? |
Valohai workers are virtual machines used to run machine learning jobs (e.g. data preprocessing, training, evaluation) that a user launches from Valohai. These machines are created and terminated according to the autoscaling rules your organization defines in the Valohai web app (e.g. min scale, max scale, shutdown grace period). |
What’s the purpose of the static (job queue) virtual machine? |
The Valohai queue machine keeps track of your jobs. Valohai will write a record about incoming jobs, and the workers will fetch their jobs that have been scheduled for their queue. Each worker will then write execution logs back to the queue machine, from where app.valohai.com will read them and show them in the user interface. |
Can we use private Docker images? |
Yes. Valohai supports standard docker login (username/password) and the main cloud providers. See the guide: Access a private Docker repository |
Does Valohai support SSO login? |
Yes. Valohai supports Okta, SSO, SAML, and AzureAD authentication. |
Network and security
Question |
Answer |
---|---|
Can we deploy to an existing VPC? |
Yes. You can either create a dedicated VPC for Valohai, or use an existing one. Valohai will just need to know which VPC and subnet(s) it’s allowed to use. We highly recommend creating a dedicated environment for running a Valohai trial, as it’s usually the fastest way to evaluate the platform. |
Can we access our existing databases and data warehouses? |
Yes. Valohai will spin all workers inside your VPC in the defined subnets. You’ll be able to access your data sources, as long as they are accessible from the workers. |
We have a self-hosted Git repository. How can we access it from Valohai? |
app.valohai.com will need to be able to access your self-hosted Git-repository. You should whitelist the IP If whitelisting is not possible, we suggest looking into the self-hosted option of Valohai. |
What ports do we need to open for Valohai? |
app.valohai.com (
|
What permissions do the Valohai managed EC2 instances have? |
By default each EC2 instance has a For example, you could define that a only a certain set of machines will have access to production RedShift databases. |
Can we limit user access to certain machine types? |
Yes. Each Valohai environment can be owned by an individual user, a team, or the whole organization. For example, you could limit access to a specific GPU type to a certain team, or limit machines and data stores in a specific AWS account and/or region to a specific set of users. |
How can we securely share credentials and keys with Valohai? |
You share credentials securely to Valohai through our self-hosted Vault. You’ll get your credentials from your Valohai contact. |
Data
Question |
Answer |
---|---|
How do we connect our existing S3 Buckets to Valohai? |
Valohai Data stores can be configured through the user interface. You can configure either project-specific data stores, or data stores that are shared with the whole organization. See the guide: Add AWS S3 |
Can we access RedShift data from Valohai? |
Yes. See the guide: Access Redshift from an execution |
Can we mount large files and storage? |
Yes. Valohai supports mounting storage over NFS and directly using EFS. See the guide Mounting a network file system (NFS) |
Comments
0 comments
Please sign in to leave a comment.