Remove en-gb
article thumbnail

Streamline diarization using AI as an assistive technology: ZOO Digital’s story

AWS Machine Learning Blog

SageMaker asynchronous endpoints support upload sizes up to 1 GB and incorporate auto scaling features that efficiently mitigate traffic spikes and save costs during off-peak times. At the time of writing, using the faster-whisper Large V2 model, the resulting tarball representing the SageMaker model is 3 GB in size.

article thumbnail

Unleashing the Power of High Throughput OCR with Visual NLP

John Snow Labs

pretrained(checkpointName, "en", "clinical/ocr").setInputCols(["image","regionEmpty"]).setOutputCol("detected_text").setBorderWidth(10).setUseGPU(True).setUseCaching(True).setOutputFormat(OcrOutputFormat.TEXT) pretrained("image_text_detector_opt", "en", "clinical/ocr").setInputCol("image").setOutputCol("detected_regions").setWithRefiner(False)

NLP 52
professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Accelerate PyTorch with DeepSpeed to train large language models with Intel Habana Gaudi-based DL1 EC2 instances

AWS Machine Learning Blog

billion-parameter model using the wikicorpus-en dataset. Each dl1.24xlarge instance has eight Habana Gaudi accelerators, each with 32 GB of memory and a full mesh RoCE network between cards with a total bi-directional interconnect bandwidth of 700 Gbps each (see Amazon EC2 DL1 instances Deep Dive for more information).

article thumbnail

DICOM de-identification at scale in Visual NLP 3/3

John Snow Labs

2023-08-20 14:17:23|426776| + --+ + -+ --+ -+ + For handle big files (2 and more GB) need to use path as input instead of content. pretrained("image_text_detector_v2", "en", "clinical/ocr").setInputCol("image").setOutputCol("regions").setScoreThreshold(0.5).setTextThreshold(0.2).setSizeThreshold(10)

NLP 52
article thumbnail

Fine-tune Llama 2 for text generation on Amazon SageMaker JumpStart

AWS Machine Learning Blog

Fine-tuning technique Language models such as Llama are more than 10 GB or even 100 GB in size. nnn### Explanation:nWe answer the question with the input's date of birth and the date of death.nnn### Solution: 1102n Response from the fine-tuned model: Félix Luna died on November 5th, 2009.nn