Artificial Intelligence
Parallelizing across multiple CPU/GPUs to speed up deep learning inference at the edge
AWS customers often choose to run machine learning (ML) inferences at the edge to minimize latency. In many of these situations, ML predictions must be run on a large number of inputs independently. For example, running an object detection model on each frame of a video. In these cases, parallelizing ML inferences across all available CPU/GPUs […]
Discovering and indexing podcast episodes using Amazon Transcribe and Amazon Comprehend
September 8, 2021: Amazon Elasticsearch Service has been renamed to Amazon OpenSearch Service. See details. As an avid podcast listener, I had always wished for an easy way to glimpse at the transcript of an episode to decide whether I should add it to my playlist (not all episode abstracts are equally helpful!). Another challenge […]

