-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Not creating ECS models #237
Comments
Hey @alu-85! With the release of LISA 3.0, model management has become more dynamic, meaning models are no longer deployed automatically during deployment. From what you've done so far, it looks like you've successfully staged them into S3 for use by LISA. To proceed with model deployment, try using the Models API. For API authentication, refer to our guide on Programmatic API Tokens. If you're troubleshooting a Chat UI issue, check the following:
|
Hi @dustins, thanks for the quick reply. Was about to look into the api so thank for confirming that's the way to deploy models. Does that mean that any config below ecsModels section in the config is surplus and not required? I've looked at the API Tokens documents and cannot see the $DEPLOYMENT_NAME-LISAApiTokenTable mentioned. I only have these relating to API: $DEPLOYMENT_NAME-lisa-chat-prod-ConfigurationApiConfigurationTable4B2B7EE1-1BL0VS5FROVTC Are the docs out of sync or am I missing a table?... |
Looks like $DEPLOYMENT_NAME-LISAApiTokenTable is only created for internet facing deployments. Ours is currently internal - so the obvious question is how do you deploy models for internal only deployments if you need to use the model api and this table to generate and store a token?
|
Would you also confirm the version of LISA you have deployed? |
I was just about to link to the same spot! I’ve added it here for anyone else who might come across this issue in the future. While we've supported some partners running within a private VPC, this isn’t a typical deployment scenario for us. You’re correct that setting Given your setup, using the Chat UI for model management might be the most straightforward solution. However, I understand that this approach has caused separate issues for you. As an alternative, you can use the master model management token we generate during deployment. Although this is currently undocumented, you can find it in AWS Secrets Manager under a name like Thank you for reaching out and letting us know about this need. We’re planning to update our documentation to clarify this process. If it would be helpful, we’d be happy to hop on a quick call today to discuss this further. We’d also love to hear more about your use case, as our roadmap is customer-driven. The product manager enjoys connecting with everyone deploying LISA deployment. |
sure, it's the latest, 3.5.1 Will do some more digging into the Chat UI issues, looks like it may be a permissions issue as the calls to retrieve models, instance types, history are all timing out/failing. It may be an IdP/token thing?
That would be good to arrange I think - there are some questions about this deployment and other configs we have - would be good to discuss on a call. We're GMT timezone though... |
Getting forbidden errors when attempting this with the master token, for both listing models or attempting to create a model, e.g. results in |
I tested this and observed behavior very similar to when Lambda functions lack internet access. While you mentioned that your private subnets have internet connectivity, I recommend double-checking to confirm. Additionally, ensure that the route tables for your private subnets are correctly configured to route external traffic through the NAT Gateway or Internet Gateway. If this resolves the issue, the next steps will be to add rules to control access to the internal load balancer. If you'd like to set up another call, I'd be happy to walk you through that process. |
Since this issue has been resolved offline, I will be closing the ticket in 48 hours unless you require any further information. Please let me know if you need anything else. |
Hello
I have deployed LISA and connected it to IdP. Models are downloaded to the S3 bucket that specified in the LISA config and the infrastructure spins up without error. However, there are no containerised models present in ECS - the cluster instance is created but there are 0 containers. Similarly the associated model repo is empty. I can log into the Model Management page via the UI and attempted to create a model there, however the UI fails to load the instance types in the drop down and I can get no further there.
So the issue is the models detailed in the configuration as below do not seem to be deployed:
I can run
make checkModels
and get the following output:The CDK output gives no errors - can someone direct any areas that might help to track down this issue?
Additional info for the custom config:
The text was updated successfully, but these errors were encountered: