Best Practices

Auto Added by WPeMatico

Build AWS architecture diagrams using Amazon Q CLI and MCP

Creating professional AWS architecture diagrams is a fundamental task for solutions architects, developers, and technical teams. These diagrams serve as essential communication tools for stakeholders, documentation of compliance requirements, and blueprints for implementation teams. However, traditional diagramming approaches present several challenges: Time-consuming process – Creating detailed architecture diagrams manually can take hours or even days […]

Build AWS architecture diagrams using Amazon Q CLI and MCP Read More »

AWS costs estimation using Amazon Q CLI and AWS Cost Analysis MCP

Managing and optimizing AWS infrastructure costs is a critical challenge for organizations of all sizes. Traditional cost analysis approaches often involve the following: Complex spreadsheets – Creating and maintaining detailed cost models, which requires significant effort Multiple tools – Switching between the AWS Pricing Calculator, AWS Cost Explorer, and third-party tools Specialized knowledge – Understanding

AWS costs estimation using Amazon Q CLI and AWS Cost Analysis MCP Read More »

Tailor responsible AI with new safeguard tiers in Amazon Bedrock Guardrails

Amazon Bedrock Guardrails provides configurable safeguards to help build trusted generative AI applications at scale. It provides organizations with integrated safety and privacy safeguards that work across multiple foundation models (FMs), including models available in Amazon Bedrock, as well as models hosted outside Amazon Bedrock from other model providers and cloud providers. With the standalone

Tailor responsible AI with new safeguard tiers in Amazon Bedrock Guardrails Read More »

Structured data response with Amazon Bedrock: Prompt Engineering and Tool Use

Generative AI is revolutionizing industries by streamlining operations and enabling innovation. While textual chat interactions with GenAI remain popular, real-world applications often depend on structured data for APIs, databases, data-driven workloads, and rich user interfaces. Structured data can also enhance conversational AI, enabling more reliable and actionable outputs. A key challenge is that LLMs (Large

Structured data response with Amazon Bedrock: Prompt Engineering and Tool Use Read More »

Effective cost optimization strategies for Amazon Bedrock

Customers are increasingly using generative AI to enhance efficiency, personalize experiences, and drive innovation across various industries. For instance, generative AI can be used to perform text summarization, facilitate personalized marketing strategies, create business-critical chat-based assistants, and so on. However, as generative AI adoption grows, associated costs can escalate in several areas including cost in

Effective cost optimization strategies for Amazon Bedrock Read More »

Build a serverless audio summarization solution with Amazon Bedrock and Whisper

Recordings of business meetings, interviews, and customer interactions have become essential for preserving important information. However, transcribing and summarizing these recordings manually is often time-consuming and labor-intensive. With the progress in generative AI and automatic speech recognition (ASR), automated solutions have emerged to make this process faster and more efficient. Protecting personally identifiable information (PII)

Build a serverless audio summarization solution with Amazon Bedrock and Whisper Read More »

Implement semantic video search using open source large vision models on Amazon SageMaker and Amazon OpenSearch Serverless

As companies and individual users deal with constantly growing amounts of video content, the ability to perform low-effort search to retrieve videos or video segments using natural language becomes increasingly valuable. Semantic video search offers a powerful solution to this problem, so users can search for relevant video content based on textual queries or descriptions.

Implement semantic video search using open source large vision models on Amazon SageMaker and Amazon OpenSearch Serverless Read More »

Multi-account support for Amazon SageMaker HyperPod task governance

GPUs are a precious resource; they are both short in supply and much more costly than traditional CPUs. They are also highly adaptable to many different use cases. Organizations building or adopting generative AI use GPUs to run simulations, run inference (both for internal or external usage), build agentic workloads, and run data scientists’ experiments.

Multi-account support for Amazon SageMaker HyperPod task governance Read More »

How climate tech startups are building foundation models with Amazon SageMaker HyperPod

Climate tech startups are companies that use technology and innovation to address the climate crisis, with a primary focus on either reducing greenhouse gas emissions or helping society adapt to climate change impacts. Their unifying mission is to create scalable solutions that accelerate the transition to a sustainable, low-carbon future. Solutions to the climate crisis

How climate tech startups are building foundation models with Amazon SageMaker HyperPod Read More »

Architect a mature generative AI foundation on AWS

Generative AI applications seem simple—invoke a foundation model (FM) with the right context to generate a response. In reality, it’s a much more complex system involving workflows that invoke FMs, tools, and APIs and that use domain-specific data to ground responses with patterns such as Retrieval Augmented Generation (RAG) and workflows involving agents. Safety controls

Architect a mature generative AI foundation on AWS Read More »