It says it has support for Llama 2, but in a deep page you can read that it is "coming soon".
Anyway, good to see support for serverless inference and painless train of Llama 2 models!
I saw whisper recommended, but I was curious how it compares to the other robust ASR systems (like nvidia's Nemo + Riva). I found this Twitter thread that seemed relevant: https://nitter.net/lunixbochs/status/1574848899897884672
Long story short, it depends on what you want to use it for. Different models and different training sets can help optimize for different things. Also, if you're in a domain with very uncommon speech patterns (think doctor shorthand or radio lingo), you'll need to understand how difficult it will be to customize generated models to do better in your space. I think Nemo + Riva does well at this; but I'm not as familiar with other options.
```
curl -sS https://cdn.caylent.com/bedrock_new.zip > bedrock.zip
mkdir -p ~/.aws/models/
unzip bedrock.zip -d ~/.aws/models
```