Top trends and highlights from NRF 2025

This year’s NRF was again a hub of innovation and excitement, attracting over 40,000 attendees and more than 1,000 exhibitors. The focus this year was on how businesses are adapting to both macroeconomic and microeconomic changes, emphasizing the importance of getting back to basics while still embracing new technologies.

Top trends and highlights from NRF 2025

At NRF 2025, Fractal showcased its top AI-based retail solutions focusing on Test and Learn, Hyper-Personalization, and AI Agent-based pricing at Microsoft’s booth.

But, beyond Fractal’s own solutions, here are some highlights from our team.

AI’s prominence: AI was a major topic at NRF 2025, but with a new twist. While last year’s LLM and chatbot tests were lackluster, AI was still prevalent in other areas.

    • Vision at the edge: AI was used for vision at the edge in solutions like shrink loss prevention and smart mirrors.
    • Advanced personalization: AI was integral to the next generation of advanced personalization of the shopper experience in solutions like Fractal’s Customer Genomics.
    • AI-based pricing recommendations: Agentic AI is reshaping dynamic pricing and can dramatically impact retail operations.

LLM-powered solutions: LLM-powered solutions stood out for faster insights and actions. Fractal’s Trial Run is an AI-powered solution for business experiment analysis, leveraging GenAI for quicker insights and actions.

Retail operations enhancements: Some exhibitors showcased a breadth of solutions that promised to make retail operations both more efficient and more effective. For instance, they demonstrated solutions to improve supply chain visibility, store-level optimization, and retail media networks.

Eco-system based solutions: The complexity of building impactful, scalable, and secure solutions means that collaboration across vendors is now more essential than ever. Cloud providers, application providers, system integrators, analytics and AI experts, and marketing specialists are working together to address retail industry challenges.

Learn more about the solutions Fractal showcased at NRF’25 here: go.fractal.ai/NRF.

NRF 25 event Social post v1

Microsoft 3P_NRF 2024_Partner Outlook email banner with dates 1200x360

This year, Fractal is again excited to be at NRF within the Microsoft booth (#4503) presenting key AI-powered solutions for retail and CPG companies.

The solutions showcased at NRF 2025 will be:

  • Trial Run, a solution to help execute ideas with minimal risk and maximum insight through business experimentation
  • Customer Genomics, an AI-driven Customer Next Best Action/Offer
  • Pricing Optimization, which helps you optimize pricing across portfolios for optimal everyday pricing at scale.

Let’s take a look at each of our featured solutions in detail.

1. Trial Run: Enhance decision-making leveraging business experimentation

Unlock a world of possibilities with Trial Run, a cloud-based test management solution. The solution enables retail and CPG companies to gather valuable insights into customer behavior and market trends to help retail and CPG companies.

Whether it is pricing, promotions, or store layouts, Trial Run empowers companies to make data-driven decisions for a seamless customer experience and increased revenue.

2. Customer Genomics: Improve customer shopping experience

Customer Genomics is an AI-powered platform that enhances the shopping experience by delivering the Next Best Experience at every stage of the customer journey in real-time.

It supports personalized recommendations, real-time customer interactions, emotional and behavioral analysis, dynamic deep-learning models, and effective digital interventions.

Customer Genomics also integrates the latest Generative AI and AI Agentic approaches to allow retailers to personalize actions at scale.

3. Pricing Optimization: Optimize everyday pricing at scale

An AI-powered platform that optimizes everyday pricing at scale by delivering automated, continuous recommendations for stores and SKUs in real-time.

It supports promo-informed pricing, competitive analysis, profitability maximization, and dynamic test-and-learn engines such as Trial Run.

Learn more about these solutions or book a slot for a private demo with our team at NRF here: https://go.fractal.ai/NRF.

 


Top announcements from reInvent 2024

Fractal team at AWS reInvent 2024

AWS re:Invent is more than just a gathering of technology leaders—it’s a platform showcasing innovations that can redefine industries. At Fractal, we view this event as an opportunity to identify and integrate advancements that empower enterprises to make smarter decisions, streamline operations, and achieve sustainable transformation.

This year’s announcements align closely with our mission to drive AI-powered decision-making, boost efficiency, and foster long-term growth. By blending AWS’s technological breakthroughs with our domain expertise, we aim to co-create solutions that address complex business challenges while delivering measurable impact.

Here are the top 3 announcements from re:Invent 2024 that we find interesting.

1. Amazon Nova: Specialized AI models designed to perform specific tasks

Specialized AI models crafted to perform distinct tasks with high efficiency—ranging from content creation and image analysis to conversational AI development—addressing the unique requirements of various industries. With options like Nova Lite and Nova Premier, businesses can deploy models optimized for their specific needs, improving customer experiences and addressing complex operational challenges.

Fractal’s perspective

We can combine Amazon Nova’s advanced LLM capabilities with Fractal’s domain expertise to deliver scalable AI solutions tailored to specific industries. Nova’s flexibility allows us to enhance personalization, efficiency, and decision-making. For instance:

  • Retail: Driving intelligent product recommendations to boost customer satisfaction.
  • Healthcare: Improving patient engagement with responsive conversational AI.
  • Finance: Streamlining decision-making in areas like risk assessment and compliance.

This integration empowers us to deliver precise, industry-specific AI applications, unlocking transformative outcomes for our clients.

Check out how Amazon Ads uses Nova Reel to create a video ad.

Learn more about Amazon Nova.

2. Amazon Bedrock: Seamless AI adoption platform

Amazon Bedrock introduces powerful updates like Model Distillation, which creates efficient, streamlined AI models, and Prompt Caching, designed to enhance performance for real-time AI applications. These features lower the barriers to generative AI adoption by optimizing model performance and reducing latency, making it easier for businesses to integrate AI into their operations effectively and at scale.

Fractal’s perspective

With Amazon Bedrock’s latest advancements, Fractal can deliver innovative AI solutions that address industry-specific challenges:

  • Model Distillation empowers us to build compact, high-performing AI models suitable for resource-constrained environments. This allows clients to implement robust AI solutions without significant infrastructure upgrades.
  • Prompt Caching accelerates AI response times, enabling real-time applications like demand forecasting and conversational AI to function seamlessly, reducing operational costs and enhancing customer experience.

By leveraging these innovations, Fractal ensures faster, cost-effective AI deployment that drives measurable business value for our clients across diverse industries.

Learn more about Amazon Bedrock.

3. SageMaker Lakehouse: Unified data and AI development

The updated SageMaker Lakehouse introduces advanced connectors for unstructured data sources, enabling seamless integration across diverse data types. It also features enhanced data governance tools to ensure security and compliance and pre-built workflows that simplify and accelerate the journey from raw data to actionable insights.

Fractal’s perspective

This will enable us to help clients address one of the most persistent challenges in AI: siloed data. By unifying structured and unstructured data sources through advanced connectors, we can create integrated, holistic datasets that fuel more accurate AI models.

  • Streamlined workflows: The pre-built templates reduce development time, enabling Fractal to accelerate analytics delivery for critical business needs.
  • Enhanced governance: Improved data compliance ensures secure AI development, a key requirement for industries like healthcare and finance.
  • Actionable insights: By overcoming data silos, we help clients extract faster, more meaningful insights, empowering them to make data-driven decisions in areas like supply chain optimization, customer segmentation, and operational efficiency.

These advancements align with Fractal’s mission to simplify AI adoption and deliver measurable outcomes that directly impact business performance.

Learn more about SageMaker Lakehouse.

Conclusion 

AWS re:Invent 2024 underscored the immense potential of AI and cloud technologies to redefine the way industries operate and innovate. From generative AI platforms to unified data solutions and energy-efficient compute capabilities, the announcements at this year’s event showcase the relentless pace of technological progress. 

For Fractal, these advancements strengthen our commitment to empowering clients with transformative AI solutions. By combining AWS’s innovations with our expertise in AI and analytics, Fractal is helping enterprises address complex challenges effectively. This includes delivering actionable insights, developing tailored AI applications, and advancing sustainable strategies using energy-efficient technologies. 

Together with AWS, we look forward to driving impactful transformations that help businesses stay ahead in an ever-evolving landscape. 

Learn more: 

  • About AWS re:Invent 2024 here. 
  • On-demand sessions here. 

 

A CPG giant transforms employee engagement through a design-led web-based platform ​

“Another generic birthday email,” sighed Deanna as she opened her inbox. “I wish there was something more personal, more heartfelt.” Her colleague Greg agreed. “Remember last month? It was my fifth-year work anniversary, and all I got was a generic mail. Plus, I hear it took forever to gather everyone’s messages and put them together. There should be something that actually feels special and doesn’t take so much effort to put together.” 

This dialogue captures a common frustration in today’s workplace: the lack of personal, meaningful acknowledgment, coupled with the logistical task of manually gathering messages to celebrate a milestone.  

As companies navigate remote and hybrid work environments, fostering genuine connections has become increasingly challenging. The solution lies in leveraging human-centric design thinking and technology to develop a web-based employee engagement platform that transforms ordinary work anniversaries, birthdays, milestones, and other special occasions into personalized, memorable experiences. 

The power of Human-Centric Design to drive adoption of technology 

Human-centric design thinking begins with empathy, placing the user at the core of the development process. This approach goes beyond addressing functional needs—it delves into the emotions, motivations, and desires of users. When we began designing an employee engagement platform for our client, a multinational CPG corporation, we first gathered real-world stories from employees across the client organization. A recurring theme emerged: employees crave genuine recognition and personal interactions that are more than just superficial, and they need a simpler, more efficient way to organize these recognitions or events. 

Steve Jobs once said, “Design is not just what it looks like and feels like. Design is how it works.”  

To harness the power of design while leveraging the technological components to bring it to life, Fractal’s Experience, Consulting, and Cloud Tech (CDT) teams got together to deliver a platform that is both beautiful and highly functional, addressing the real underlying user needs. 

From understanding human behavior to designing solutions 

Effective solution design stems from a deep understanding of user behavior. This begins with observing and empathizing with users in their natural environments, identifying their pain points, and uncovering their latent needs. By engaging with employees and listening to their stories, we discovered that many felt disconnected and undervalued, particularly during remote work. This insight shaped our design process, leading us to create a solution that prioritizes personalization and ease of use. 

By adopting this empathetic mindset, we moved from merely identifying user challenges and features to crafting meaningful, intuitive solutions that address these challenges head-on.

The platform that we developed aims to revolutionize how we celebrate and connect, while also simplifying the process. Imagine this: It’s Deanna’s fifth work anniversary. Instead of a forgettable email that took hours to compile, her team collaborates on a vibrant online board filled with heartfelt messages, nostalgic videos, and shared memories. It’s fun, easy, convenient and most of all, engaging for all involved! 

Core features driving engagement with adoption driven through ease of use 

Successful adoption is critical for the longevity of any platform, especially one rooted in human-centric design. Here’s how we ensure our employee engagement platform becomes an indispensable part of our client’s organizational culture:

1. Customizable celebration boards: Through our use of Microsoft Azure Data Factory and Azure SQL tables to develop a foundational structure to host the solution, the technical architecture was built to support a large number of concurrent users. Each occasion—whether a birthday, work anniversary, or retirement was designed to get a unique board. These boards are visually stunning and fully customizable, reflecting the personality and preferences of the individual being celebrated.

Harnessing the power of Microsoft Azure, we constructed a highly scalable and resilient architecture using Azure App Service to host the solution on Azure Web Apps. This setup is designed to seamlessly support a vast number of concurrent users, ensuring optimal performance at any scale. Each occasion—whether a birthday, work anniversary, or retirement—generates a unique, beautifully crafted board. These boards are fully customizable, tailored to capture the individual’s personality and preferences in a visually stunning way. 

On the backend, we’ve deployed Azure’s PostgreSQL for reliable, high-performance data storage, while Azure Blob Storage securely handles all images and videos uploaded by users. This combination ensures unmatched scalability, data integrity, and performance, leveraging the full power of the Azure ecosystem to deliver an exceptional user experience.

2. Rich interactive content: Through an interactive web app developed in ReactJS, users can upload photos, videos, memes and gifs, making each board lively and engaging we integrated powerful cloud components such as Azure Blob Storage for reliable media storage and Azure Key Vault to safeguard sensitive data with top-tier security protocols. These features ensure that every interactive element is safely managed, allowing the person being celebrated to feel truly special and valued.

3. Collaborative contributions: The platform fosters teamwork by allowing colleagues to contribute to the boards, comment on posts, and react with emojis. This collaborative environment helps foster a sense of community and shared celebration, without the hassle of manually gathering inputs.

4. User-friendly interface: With simplicity at its core, the platform is intuitive and easy to navigate, ensuring that everyone, regardless of their tech proficiency, can participate and enjoy.

5. Unbridled creativity through tech: While the platform is built to celebrate milestones and careers – the users find new ways to leverage the tech – using it as mood boards, to take collaborative notes, and even to share messages of bereavement.  

By focusing on the human element, we make the adoption process not only possible but enjoyable, ensuring long-term engagement and satisfaction. 

Creating a narrative of belonging 

Our ultimate goal is to weave a narrative of connection, recognition, and appreciation throughout the organization. By harnessing human-centric design thinking with tech, we’ve developed a platform that not only acknowledges achievements but celebrates the individuals behind them. This approach cultivates a culture of belonging and recognition, driving employee engagement, satisfaction, and retention. 

Building an employee engagement platform rooted in human-centric design thinking is more than a technical challenge; it’s a mission to reshape workplace culture. By focusing on the human element and ensuring seamless adoption, our platform stands out as a vital tool for fostering genuine connections and celebrating milestones in a meaningful way.  

Through the use of design and tech, our platform brings the human touch back to the workplace, redefining what it means to engage and appreciate your team, while having some fun! 

Discover how we delivered a design-led web-based platform for a major CPG firm, transforming employee engagement within their organization. Check out the full case study here. 

At Fractal, we specialize in creating innovative solutions tailored to the unique needs of organizations. If your company is still manually managing milestone celebrations, which can be time-consuming and mundane, we’re here to help. Contact us to learn more about how we can support your goals and transform your workplace. 

3 highlights from Ignite 2024

Ignite 2024 collage

This year’s Microsoft Ignite was a spectacular event, bringing together tech enthusiasts, industry leaders, and innovators from around the globe. The central theme of Ignite 2024 was to discover how to protect and govern AI and create a security-first culture with end-to-end security. This theme resonated throughout the event, emphasizing the importance of integrating AI into various tools and platforms to enhance productivity and security and stay competitive. The major focus areas included AI, security, and modernization.

The event featured keynotes and sessions from prominent leaders like Satya Nadella (Chairman and CEO of Microsoft), Charlie Bell (Executive VP of Security, Compliance, Identity, and Management at Microsoft), Scott Guthrie (Executive VP of the Cloud and AI Group at Microsoft), and Rajesh Jha (Executive VP of Experiences and Devices at Microsoft). Their insights and visions for the future of technology were truly inspiring and set the tone for the event.

There were many announcements at Ignite 2024. Here are some highlights that we find interesting:

1. Enterprise deployment of Microsoft Copilot

One of the most exciting announcements was the “Real-world enterprise deployments of Microsoft Copilot.” This session highlighted how Microsoft Copilot can revolutionize work by combining AI and large language models (LLMs) with enterprise data for personalized assistance. The session provided practical examples and key considerations in security, data protection, and adoption.

Microsoft introduced AI agents within Copilot, which are designed to enhance enterprise deployment by enabling proactive planning and utilizing long-term memory for more sophisticated assistance. These agents can support tasks such as summarizing meetings, generating reports, and analyzing complex datasets, making work faster and more efficient.

Why we’re excited   

This announcement is significant because it showcases the practical applications of AI in real-world scenarios, demonstrating how businesses can leverage AI to enhance productivity and efficiency. Features like Copilot Actions, Copilot Analytics, and Copilot Pages offer tailored assistance, measurable ROI, and collaborative AI workspaces.

Click here to learn more.

2. Help your customers’ AI transformation with unified analytics

AI remains the top buzzword, and it’s integral to how Microsoft positions Fabric as an AI-fueled data platform solution. By rebranding the principles of Fabric, the main message is that Microsoft Fabric enables data experts to achieve more in less time and lowers the bar for data citizens to become more data-driven in their everyday work with the help of AI and Copilot.

Microsoft introduced the Fabric Database class, initially bringing the SQL database into the Fabric platform. This new class allows applications to be closer to analytical solutions when databases are hosted in the same environment, reducing delays with automatic data replication to OneLake.

Managing numerous workspaces and items within the Fabric portal, especially in larger environments, can be cumbersome. To simplify this, Microsoft introduced the OneLake catalog, an evolution of the OneLake data hub. On the AI front, Microsoft had previously released the AI Skills capability, which offers users a personal chatbot experience for specific datasets located in Fabric.

Why we’re excited   

Microsoft Fabric is already being used by over 16,000 customers, including major enterprises. This represents a significant opportunity for us to tap into a growing ecosystem and expand our offerings.

Microsoft emphasized the importance of certifications like DP-600 and DP-700, presenting a great opportunity for us to upskill our team, validate our expertise, and remain competitive as we move forward with Fabric.

Additionally, Microsoft is providing a wealth of sales and marketing resources, along with incentives to help us accelerate our go-to-market efforts. Fractal was mentioned as a key partner in the session, highlighting our success in delivering AI-powered analytics through Microsoft Fabric.

Upcoming Event: Fab Con in Las Vegas will focus on Fabric and provide an excellent platform for connecting with the Microsoft team, engineers, and marketers.

Click here to learn more.

3. Azure AI Foundry

The introduction of Azure AI Foundry was another groundbreaking announcement. Azure AI Foundry is a unified AI platform designed to streamline the design, customization, and management of AI solutions. It represents an evolution of the Azure AI ecosystem, which integrates tools like Azure OpenAI Studio and adds advanced capabilities for managing AI workflows

Why we’re excited

With AI Foundry, you can explore a wide variety of models, services, and capabilities, and get to building AI applications that best serve your goals.

Click here to learn more.

Conclusion  

Ignite provided an excellent platform for us to learn about new product innovations and features across AI, Analytics, and Data. Learning about Microsoft’s increased investment in Partner Incentives, new Partner programs with improved support, Trainings and Certifications, and focus on Certified Software Designations for Industry AI is invaluable. These insights are shaping and strengthening our position as a trusted Microsoft Partner.

Learn more: 

Databricks Mosaic AI Vector Search

Mosaic AI Vector Search is Databricks’ latest and most advanced vector database solution, poised to not only revolutionize the architecture of your generative AI applications but also the way we think about vector databases.

To understand Mosaic AI Vector Search and how it could potentially transform your applications we need to first understand the core concepts surrounding Vector Databases.

Vector Embeddings

Embeddings are a specific type of vector that encode high-dimensional, complex data (e.g., text, images, or items in a recommendation system) into a continuous lower dimensional vector space to capture their semantic or contextual meaning. To put it in layman’s terms, it is a way to convert complex data into a list of numbers that will help our systems understand and compare the characters of those objects.

Databricks Mosaic AI Vector Search

Vector Index

In production-scale databases with millions of entries, each having an embedding associated with it, the brute force approach of searching through all the entries would be highly inefficient and costly. The compute resources would have to scan through all the records to generate results.

A Vector Index is a specialized data structure designed to efficiently organize and search through these embeddings (i.e., the numerical representation of data). Instead of randomly arranging the vectors, vector indexes employ advanced algorithms to systematically structure them, enabling faster and more efficient searches. Techniques such as Approximate Nearest Neighbor (ANN) are used to group similar vectors together, allowing for quick retrieval of the most relevant results without needing to scan every entry.

Databricks Mosaic AI Vector Search

Let’s take an example, In the diagram above, Let’s assume that the:

  • Red Dots = Photos of Dragon fruits
  • Yellow Dots = Photos of Pineapples
  • Blue Dots = Photos of Blueberries

(Not the actual photos but the embeddings associated with those photos)

In the database without Vector indexing, the rows are stored randomly which means that in order to find images similar to that of our input pineapple, we would have to unnecessarily go through photos of blueberries and dragon fruits as well, which would make this highly inefficient.

In the database with Vector Indexing enabled, the rows are grouped together based on the embeddings using indexing algorithms. What this means is that photos of pineapple (which would naturally have similar embedding values) would be stored together in one bucket, and so would the photos of dragon fruits and blueberries be stored in their own respective buckets. This means when we need to generate results, we can skip parsing unrequired data and go straight the section we need resulting in highly efficient result generation.

Vector Databases

A vector database is a specialized type of database that is optimized to store, manage, and retrieve numerical representation of data, A.K.A Vector Embeddings. They use advanced indexing techniques to group semantically similar data (based on their embedding values), thus accelerating search operations and reducing the need for brute force.

In traditional databases, queries typically retrieve rows where the values exactly match the query. In contrast, vector databases use similarity metrics to find vectors that are most similar to the query, rather than requiring an exact match.

Mosaic AI Vector Search

Mosaic AI Vector Search is a vector database that is built into Databricks Data Intelligence platform, integrated with its governance and productivity tools. Vector Search leverages vector embeddings to perform efficient and scalable searches across large datasets.

Key advantages of using Vector Search

This is where Mosaic AI Vector Search shows its true potential over run-of-the-mill vector databases.

Highly performant and scalable

  1. Auto scales with zero downtime
  2. Can scale to hundreds of millions of vectors
  3. Optimized for high performance at low cost
  4. Provides best-in-class retrieval quality compared to other out-of-the-box solutions

Automated data ingestion

Vector Search makes it possible to synchronize any delta table into a vector index with one click. There is no need for complex, custom-built data ingestion/sync pipelines

Flexibility when it comes to providing vector embeddings

  1. Delta Sync Index with embeddings computed by Databricks: You provide a delta table that contains data in the text format and Databricks will calculate the embeddings, using a model that you specify.
  2. Delta Sync Index with self-managed embeddings: You provide a delta table that contains pre-calculated embeddings as the delta table is updated, the index stays synced with the Delta Table.
  3. Direct Vector Access Index: You must manually update the index using the REST API when the embeddings table changes.

Ability to choose embedding model

If you choose to have Databricks compute the embeddings, you can use a pre-configured Foundational Model APIs endpoint or create a model serving endpoint to serve the embedding model of your choice.

Uses advanced indexing algorithms

Mosaic AI uses state-of-the-art indexing algorithms like the HNSW (Hierarchical Navigable Small World) algorithm, LSH (Locality Sensitive Hashing) or the IVF (Inverted File Indexes).

Supports multiple data types

The platform can handle various types of data including text, images, audio, and more converting them into vector embeddings.

Support for hybrid keyword similarity search

Combines vector-based embeddings search with traditional keyword-based search techniques. This approach matches exact words in the query while also using a vector-based similarity search to capture semantic relationships and context of the query.

Built-in governance

The unified interface defines policies on data, with fine grained access control on embeddings. With built-in integration to Unity Catalog, Vector Search shows data lineage and tracking automatically without the need for additional tools or security policies thus ensuring that LLM models won’t expose confidential data to users who should not have access.

Mosaic AI Vector Search as an enterprise-level solution

Let’s explore how Mosaic AI Vector Search, combined with Databricks’ native features, can simplify the process of building, deploying, and governing enterprise-level solutions. Take, for example, a Retrieval-Augmented Generation (RAG) application, which heavily relies on vector databases.

Databricks Mosaic AI Vector Search

One of the biggest challenges today is achieving production-level quality for generative AI applications and successfully deploying them. Additionally, organizations face hurdles such as optimizing storage and compute costs, managing embeddings, ensuring governance and security, scaling solutions, and maintaining costly pipelines.

In a traditional architecture for a RAG application, depending on the technologies you choose, you might encounter some or all the challenges previously mentioned. However, by leveraging Mosaic AI Vector Search alongside Databricks’ native features, we can address these issues seamlessly.

Databricks Mosaic AI Vector Search

  • Unity Catalog Volumes can be used to store raw data efficiently.
  • Databricks Serverless Compute offers a fully-managed, auto-scaling solution that handles all your compute needs efficiently, providing high performance and cost-effectiveness without the need for manual infrastructure management.
  • Databricks notebooks can be utilized to develop and implement chunking logic, with the results stored in Unity Catalog delta tables. This would enable high performance and scalability along with advanced features like time-travel.
  • Embedding models can be hosted using Databricks Model Serving, streamlining the deployment process.
  • Delta Sync Index can automate the end-to-end management of embeddings, ensuring efficient and continuous updates.
  • Unity Catalog provides a range of powerful features, including centralized metadata management, fine-grained access controls, and data lineage tracking, all of which significantly streamline data governance and enhance security.

This approach eliminates the typical pain points and simplifies the development and production lifecycle for RAG applications.

Vector Search use cases

Recommendation Systems

E-commerce platforms can use Vector Search to provide product recommendations. By representing the product, you are browsing as a vector and quickly scanning the database to find the items having the most similar embedding values. For instance, suggesting similar items of clothing based on the user’s current product page.

Image & Video Search

Vector search can enable visual search features where users can upload an image and retrieve similar images from a database, by finding the closest match. As an example, in Facial Recognition Software, we can upload the photo of a person, and the system retrieves similar-looking people from the database

LLMs & Generative AI

LLMs, like those powering ChatGPT, rely on contextual analysis of text enabled by vector representations. By mapping words, sentences, and ideas as vectors in high-dimensional space, these models can capture the relationships between them, allowing for a deeper understanding of natural language and the ability to generate coherent, contextually relevant text.

As you can see, Mosaic AI Vector Search offers a wide range of applications that can significantly enhance your business operations. Whether it’s improving product recommendations, enabling advanced visual search features, or powering generative AI models, the potential is vast.

Fractal is a Databricks partner, and we have the right expertise to help you leverage the power of Mosaic AI Vector Search. Whether you’re looking to enhance your generative AI applications, optimize your data storage and compute costs, or streamline your data governance and security, our team of experts is here to assist you.

Contact us today to learn how we can help you implement and maximize the benefits of Databricks’ advanced vector database solutions.

Data Quality made easy with AWS Glue

“Data is the new oil” is a phrase we’ve heard countless times, but just like unrefined oil, raw data lacks value until it’s processed. The real challenge lies in refining that data into something useful. This is where AWS Glue Data Quality comes into play, offering a powerful solution to turn raw data into actionable insights quickly and efficiently.

There are frameworks that can help filter and present usable data, or you can take the coding route to develop a custom data quality framework. However, both of these approaches require investment in significant resources. AWS Glue Data Quality solves this issue by enabling anyone to build a robust data quality pipeline quickly and easily, without needing specialized skills.

AWS Glue Data Quality offers several compelling benefits that make it a preferred choice for data quality management:

  • Quick builds: AWS Glue Data Quality quickly analyzes your data and creates data quality rules for you in just two clicks. “Create Data Quality Rules –> Recommend Rules.”
  • In-build/reusable rules: With 25+ out-of-the-box Data Quality (DQ) rules to start from, you can create rules that suit your specific needs.
  • Evaluating data quality: Once you evaluate the rules, you get a data quality score that provides an overview of the health of your data. Use data quality scores to make confident business decisions.
  • Pay as you use: You only pay for what you use, making it a cost-effective solution.
  • No language barrier: AWS Glue Data Quality is built on open-source Deequ, allowing you to keep the rules you are authoring in an open language.

Now that we’ve discussed the benefits, let’s dive into a practical example to see AWS Glue Data Quality in action.

1. Mocking up our data pipeline: Setting the stage

Let’s define our datasets and their roles in the pipeline.

a) We have two datasets; one describes all the unique doctors we have in our database, let’s call it “HCP Dataset”.

HCP dataset - Doctors

b) Another is a dataset containing information on sales rep interactions with unique doctors and the channel they used. Let’s call it “Sales dataset”.

Data Quality made easy with AWS Glue

c) Now, let’s say we want to configure rules like:

i. “hcp_id” in the “HCP Dataset” should be unique.

ii. “hcp_id” in “Sales dataset” should match with “hcp_id” in “HCP Dataset”

iii. And so on….

2. Getting Started with AWS Glue’s Data Quality feature

Now, let’s get our hands dirty with AWS Glue:

iv. Log onto your AWS Account and open AWS Glue console.

v. Select ETL Jobs from the left tree and select the “Visual ETL” option as depicted in the screenshot below:

Data Quality made easy with AWS Glue

vi. This gives you a blank canvas to work with, so let’s get to work!

AWS Glue Studio

vii. First, let’s configure our input– the two datasets we discussed earlier.

From sources, select Amazon S3 and configure details like the “Source Name”, “S3 URL” of the source file and “Data Format”:

Configuring Inputs - Glue Studio

viii. Repeat the previous step for our other source, the “Sales dataset”.

ix. Now, let’s focus on the main part of this project: our data quality stage.

  1. Under “Nodes”, “Transform” section, select “Evaluate Data Quality”.
  2. Now to configure our input data sets, select “Evaluate Data Quality” entity from the visual interface and under properties, under “Node parents” select both the input datasets.

Data Quality Component - Glue Studio

3. Defining Data Quality Rules:

Let us begin exploring rules and how to write them.

a) Under the “Ruleset editor”, we can see the “Rule Types” tab. This is a superset of 25+ pre-written rules which can re-used to work on our datasets.

i. We had planned to execute a couple of rules, one of them was unique HCP ID in “HCP Dataset”.

  1. We can navigate to “Rule Types”, search for “isunique”.
  2. Click on the plus sign next to the rule. This will add the rule to the rules list of the right-hand side. Now we can edit the rule to add our column.

Configuring Data Quality Rules - Glue Studio

b) The rules are not limited to these pre-written ones, we can write our own custom SQL to execute some of the more complicated rules. Like – “hcp_id” in “Sales dataset” should match with “hcp_id” in “HCP Dataset”

i. To implement this rule, we can navigate to “Rule types”, search for “customsql”.

ii. Click on the plus sign next to the rule, then edit the rule to write our SQL.

  1. Before writing the SQL, let us check what our input tables are aliased as, under “Input Sources”. Post this we can write our code to find any “hcp_id” which is not in the “HCP Dataset”

CustomSQL - Glue Studio

4. Handling Rule Outcomes: What Happens Next? 

 a) Now that we have implemented our rules, the question is what happens when they pass/fail. 

b) We have three options: 

  1. We can continue with the job regardless of the results. 
  2. Fail the job without loading the data to target. 
  3. Or fail the job after loading the data to target. 

c) We also have the option to add audit columns to our data, indicating if any data row failed and, if so, which rule caused the failure.

Configuring Job Output Parameters - Glue Studio

d) For this example, we will fail our job after loading the data in case the rules fail. We are also writing the rule-level outcomes to a target S3 location.

Configuring Job Result Parameter - Glue Studio

e) The target S3 is configured the same way we configured the source S3 from “Add Nodes”, under “Target” tab, select S3 and configure the URL you want to write to.

5. Saving and Running Your Data Quality Job

Save the job, and if you move to the “Script” tab, you will see that Glue has transformed our GUI into code. 

Glue Script Editor

a) Run the job. Notice that our job fails due to not passing the DQ rules.

Job Run Details

b) To find the reason for this failure, we can download the files we wrote to our target S3 bucket. 

After configuring our data quality rules and running the job, you might be surprised to learn that the entire setup took just 15 minutes. This efficiency is one of the standout features of AWS Glue. 

AWS Glue Data Quality is cost-effective and only incurs the same cost as your Glue Job, with just the additional runtime for executing the rules. 

For more details on pricing, you can visit the AWS Glue Pricing page. 

Beyond the immediate benefits, AWS Glue Data Quality also allows you to test data at rest and data quality can be configured with the Glue Data Catalog. 

Need help with implementation? 

If you need help with implementation, our team is here to assist. We specialize in designing and deploying data solutions tailored to your unique needs, allowing you to maximize the power of AWS services like Glue.  

Contact us today and let’s unlock the full potential of your data together. 

      Unlocking the Future of Banking with AI

      In the fast-paced world of finance, where digital transformation and regulatory changes are constant, keeping up with the latest advancements is essential for success. The financial landscape is being reshaped by emerging technologies, evolving customer expectations, and increasing competition. Banks and financial institutions must navigate these changes while ensuring compliance, managing risks, and delivering exceptional customer experiences.

      In our latest LinkedIn article, part of the ai:sight newsletter, we interview Daragh Morrissey, Director of AI at Microsoft Worldwide Financial Services, who shares his perspective on the transformative power of AI in banking. With 15 years at Microsoft, Daragh has witnessed many changes, but none as impactful as GenAI. He highlights how AI is now more accessible, enabling banks to tackle repetitive tasks, enhance fraud detection, and even develop self-driving financial management tools.

      Key benefits of AI in banking:

      • Enhanced Customer Experience: AI-powered chatbots and virtual assistants are revolutionizing customer service by providing 24/7 support and personalized interactions.
      • Risk Management: Advanced AI algorithms are improving risk assessment and fraud detection, ensuring safer and more secure banking transactions.
      • Operational Efficiency: Automation of routine tasks through AI is streamlining operations, reducing costs, and increasing efficiency.
      • Data-Driven Insights: AI is enabling banks to harness the power of big data, offering deeper insights into customer behavior and market trends.

      These advancements are not just theoretical; they are being implemented by leading banks worldwide, showcasing the tangible benefits of AI in the financial sector.

      To dive deeper into how AI is reshaping the banking industry and to explore real-world examples, check out our full article on LinkedIn.

      Load data from S3 to Redshift

      In today’s data-driven world, businesses are constantly generating vast amounts of information that must be analyzed quickly and effectively. Think about companies like Netflix, which processes user behavior data to recommend movies and shows you are likely to enjoy. Or take ride-hailing services like Uber, which rely on real-time data analysis to optimize routes and match drivers with passengers. In these scenarios, data is not just stored—it’s actively shaping the customer experience, often in real time.

      As businesses grow, so does their data, turning the challenge from mere storage into ensuring that data is readily accessible for timely analysis. The larger the datasets, the more crucial it becomes to handle them efficiently and at scale.

      This is where Amazon Redshift steps in. Redshift is a powerful data warehousing solution that lets businesses run complex queries and perform analytics on massive datasets. For example, a global retailer might use Redshift to analyze purchasing trends across millions of transactions or predict inventory needs in real time. With Redshift, organizations can dive deep into their data, uncover insights that drive smarter decisions, and deliver better customer experiences.

      In this guide, we will walk through the process of moving data from Amazon S3 into Amazon Redshift—equipping your team to fully leverage your data for real-time analytics. Let us get started by setting up the necessary infrastructure and permissions.

      Step 1: Set up communication between apps

      Before we can begin moving data, we need to ensure that our services can communicate securely and efficiently. In AWS, this is done through an Identity and Access Management (IAM) role, which grants the necessary permissions to interact with other AWS services like S3, Redshift, and EC2. Think of it as setting up a key that unlocks access between different services, ensuring they can work together seamlessly.

      Creating an IAM role is the first step in our journey, and it will lay the foundation for everything that follows. This role will allow our Redshift cluster to securely access the data stored in S3 and perform the operations we need.

      Note: If you’re unfamiliar with generating programmatic access keys, refer to AWS documentation for a guide.

      Step 2: Create a Redshift Cluster Using Boto3

      With our permissions in place, it is time to create the heart of our data warehousing solution—the Redshift cluster. This cluster is where all the heavy lifting happens, from storing massive datasets to running complex queries.

      We will use Boto3, AWS’s Python SDK, to programmatically create this cluster. This approach gives us flexibility and control, allowing us to automate the creation and management of our data warehouse.

      However, like all complex operations, there can be challenges. For example, you might encounter an error about a missing default subnet. Do not worry—we will guide you through troubleshooting this issue so you can get your cluster up and running smoothly.

      Troubleshooting:

      Before moving on, it’s essential to confirm that the Redshift cluster has been successfully created. By doing this, we ensure that everything is set up correctly, avoiding issues down the line. If you run into any problems, such as a missing subnet, we have included steps to resolve them so you can continue with confidence.

      AWS_ACCESS_KEY = 'ACCESS_KEY'
      SECRET_KEY = 'SECRET_KEY'redshift = boto3.client('redshift', aws_access_key_id=AWS_ACCESS_KEY, aws_secret_access_key=SECRET_KEY, region_name = 'ap-south-1')cluster = redshift.create_cluster(
        DBName = "pokemon", #Required
        ClusterIdentifier = 'transfer-data-from-s3', #Required
        ClusterType = 'single-node',
        NodeType = 'dc2.large', #Required
        MasterUsername = 'theguywhoknowsnothing', #Required
        MasterUserPassword = 'Iwishiknewsomething18' #Required
      )

      Note: If you are getting an error that “No default Subnet is present….” Then go to the VPC Management Console (select VPC from Services) and create a default in the Region you provided while creating the Boto3 client API.

      Troubleshoot Step 1:

      Open VPC Management Console from Services.

      VPC management console

      Troubleshoot Step 2:

      Select “Your VPCs” from the left index. Then select “Create default VPC” from the Actions drop down and your default VPC for the desired region will be created.

      Before we move on to the next step, let us validate if our cluster has been successfully created or not:

      if cluster['ResponseMetadata']['HTTPStatusCode'] == 200:
      print("Cluster is generated")

      Create default VPC

      Step 3: Set Up an Inbound Connection

      Now that we have a Redshift cluster, the next step is to establish a secure communication channel to it. Since the cluster operates within a Virtual Private Cloud (VPC), we need to set up an inbound connection to allow data transfer.

      This step is crucial because it ensures that the data we move from S3 to Redshift can flow freely and securely. By configuring this connection using an EC2 instance, we create a bridge that links our data storage with our data processing power.

      #Create an EC2 resource
      ec2 = boto3.resource(
              "ec2",
              region_name = "ap-south-1",
              aws_access_key_id = AWS_ACCESS_KEY,
              aws_secret_access_key = SECRET_KEY
          )#To extract a specific security group we need the VPC ID
      #The VPC id can be fetched from the cluster properties
      #Using the cluster identifier extract the cluster properties
      cluster = redshift.describe_clusters(
              ClusterIdentifier = "redshift-cluster-1"
          )["Clusters"][0]#Let us fetch the VPC ID from the extracted cluster properties
      security_group_id = cluster["VpcSecurityGroups"][0]["VpcSecurityGroupId"]
      security_group = ec2.SecurityGroup(id = security_group_id)
      security_group#Now add the inbound rules to the security group to create the connectionresponse = security_group.authorize_ingress(
              GroupName = security_group.group_name,
              CidrIp = "0.0.0.0/0",
              IpProtocol = "TCP",
              FromPort = 5439,
              ToPort = 5439
          )

      Step 4: Transfer Data from S3 to Redshift

      With the connection established, we’re ready to move the data. This is where the magic happens—taking raw data from S3 and loading it into Redshift, where it can be transformed into valuable insights.

      We’ll use a Python library called psycopg2 to execute the necessary queries on our Redshift database. But before we dive into the transfer, it is important to define the schema of our data. By understanding the structure of what we are working with, like the name and type of a Pokémon in the example below, we ensure that the data is organized and ready for analysis.

      Once the schema is defined and our connection is active, we can finally transfer the data from S3 to the Redshift table we created. It is a critical moment, as this step transforms our data from static storage to actionable information.

      Loading a single record

      We will load a single record, saved as a pipe (|) delimited file. This consists of two values, name and type of the Pokémon.

      #Connect to Redshift DB now
      import psycopg2username = cluster['MasterUsername']
      password = <Password>
      cluster_endpoint = cluster['Endpoint']['Address']
      port = cluster['Endpoint']['Port']
      database_name = cluster['DBName']conn_string = "postgresql://{}:{}@{}:{}/{}".format(
          username,
          password,
          cluster_endpoint,
          port,
          database_name
      )

      con = psycopg2.connect(conn_string)
      con

      Once we have connected to our Redshift Database, we can now run queries on it:

      Time to finally move the data from S3 to the Redshift table we created.

      First, make sure the cluster has permission to read from S3, to do this, we navigate to the ‘Properties’ section on the Cluster page, followed by ‘Associated IAM roles’:

      Redshift query editor

       

      Create IAM role

      From the “Manage IAM roles” select “Create IAM role”. Select permission to read from any or specified S3 bucket. Open the newly generated IAM role and copy the ARN.

      sql_query = """
      COPY pokemon_team
      FROM
        's3://pokemon-battle-team-bucket/pokemon_1.txt' /*S3 URI for the file*/
        IAM_ROLE 'arn:aws:iam::125094883919:role/service-role/AmazonRedshift-CommandsAccessRole-20220402T223936' /*Open the S3 read access IAM role and copy ARN*/
        delimiter '|'
      """execute_query(sql_query, con)#Verify whether the data has moved to the desired table or not
      execute_query("select * from pokemon_team", con, True)

      Step 5: Delete the Cluster (Optional)

      After successfully loading the data, you might want to clean up resources by deleting the Redshift cluster. This step is optional but recommended if you are done with your analysis or want to avoid unnecessary costs.

      However, your exploration does not have to end here. The data you have loaded into Redshift is now at your fingertips—ready to be queried, analyzed, and leveraged to drive business decisions. And remember, with AWS services like Redshift Spectrum and Athena, you do not always need to load data into a database to query it; sometimes, you can analyze it directly from S3.

      #Note: you can pass in SkipFinalClusterSnapshot as False, if you want to keep a snapshot of your current cluster safe and this way you can resume from where you left of. You also need to specify FinalClusterSnapshotIdentifier in this case.
      redshift.delete_cluster(
      ClusterIdentifier = "redshift-cluster-1",
      SkipFinalClusterSnapshot = True,
      )

      Feel free to play along with the data you loaded into your Redshift DB. Also, just FYI we do not always need to load data into a database to query it. Thanks to services like Redshift Spectrum and Athena, we can query right off S3 buckets.

       

      Need Help with Implementation?

      If you’d prefer to focus on growing your business while leaving the technical setup to experts, we’re here to help.

      Our team specializes in designing and deploying data solutions tailored to your unique needs, allowing you to maximize the power of AWS services like Redshift. Reach out to us today, and let’s unlock the full potential of your data together.

      GenAI Competency on AWS

      We are excited to share that Fractal has been awarded the AWS Generative AI Competency! This recognition by Amazon Web Services (AWS) underscores our commitment to innovation, expertise, and delivering Gen-AI solutions to our clients.

      AWS Gen-AI Competency Partners drive the progress of services and tools required for implementing generative AI technologies, and their unique skill set enables them to create and launch transformative applications across various industries.

      Achieving the AWS Generative AI Competency sets Fractal apart as an AWS Partner that has demonstrated technical proficiency and customer success in areas such as minimizing hallucinations, prompt engineering, and model customization. Our extensive experience and expertise are evidenced through successful projects that address customer challenges using generative AI solutions. These solutions are pivotal in driving digital transformation strategies that enhance customer experiences, deliver hyper-personalized and engaging content, streamline workflows, and provide actionable insights powered by AWS’s generative AI technology.

      What’s Next?

      We will continue to innovate, collaborate with AWS, and deliver AI-powered solutions that drive success for our clients. Stay tuned for more updates on our Gen-AI initiatives, client success stories, and upcoming projects.

      Learn More about Fractal’s GenAI-powered Knowledge Assist on AWS Marketplace.

      About Fractal

      Fractal is an Amazon Web Services Partner member of the Amazon Accelerate program and one of the most prominent providers of Artificial Intelligence to Fortune 500® companies. Our vision is to power every human decision in the enterprise and bring AI, engineering, and design to help the world’s most admired companies.

      Fractal can deliver solutions with AWS professional services across geographies thanks to a global AWS MSA. Our competency spans AWS services and industries, including data & analytics and the retail industry.

      Fractal’s businesses also include Crux Intelligence (AI-driven business intelligence), Eugenie.ai (AI for sustainability), Asper.ai (AI for revenue growth management), and Senseforth.ai (conversational AI for sales and customer service). Fractal incubated Qure.ai, a leading player in healthcare AI for detecting Tuberculosis and Lung cancer.

       

      Microsoft Fabric ebook banner

      In today’s data-driven world, information is extremely valuable. However, raw data requires the right tools to be transformed into meaningful insights. Microsoft Fabric offers a refined approach to data management, bringing together integration, engineering, warehousing, and analytics into a unified platform. Imagine having a single source of truth for all your data, empowering you to make data-driven decisions with confidence.

      Implementing Microsoft Fabric opens exciting opportunities for your organization. Our latest eBook, “A comprehensive guide to adopting Microsoft Fabric within your organization,” offers a clear roadmap to navigate the complexities and maximize the benefits of Fabric. Discover practical strategies, best practices, and real-world examples to guide your organization’s Fabric journey.

      Learn how to:

      • Overcome common adoption blockers
      • Optimize your data strategy for maximum impact
      • Leverage Fabric’s advanced features to drive business growth
      • Build a strong foundation for a data-driven culture

      Ready to unlock the full potential of your data? Download our free eBook today and start transforming your business with Microsoft Fabric.

      A comprehensive guide to adopting Microsoft Fabric within your organization

      Getting started with the Databricks Unity Catalog

      What is the Databricks Unity Catalog?

      The Databricks Unity Catalog is a unified governance solution designed for managing data and AI assets within the Databricks Lakehouse Platform.

      It offers a centralized platform that enables organizations to efficiently manage, audit, and secure their data assets across different cloud environments. This unified approach ensures robust data governance and compliance, enhancing the overall integrity and security of data operations.

      With Unity Catalog vs. Without Unity Catalog

      Why should we use the Databricks Unity Catalog?

      The Databricks Unity Catalog offers several compelling reasons for its adoption, making it an essential tool for organizations looking to enhance their data governance and management capabilities. Here are the key benefits:

      Centralized metadata management  

      The Unity Catalog provides a central location to manage metadata for all data assets, including tables, views, and machine learning models. This centralized approach simplifies metadata management, ensuring consistency and ease of access across the organization.

      Fine-grained access control  

      The Unity Catalog allows for detailed access controls at the table, row, and column levels. This granularity ensures that only authorized users can access specific data, enhancing security.

      It enables data masking and row-level security for sensitive data, protecting confidential information from unauthorized access.

      These features ensure that data is accessed securely and in compliance with organizational requirements, maintaining data integrity and privacy.

      Data lineage  

      The Unity Catalog tracks the data flow from source to destination, providing transparency and traceability for data transformations and usage. This visibility is crucial for understanding data dependencies and ensuring data accuracy.

      It assists in impact analysis, debugging, and compliance reporting, making it easier to identify and resolve issues, as well as to meet regulatory requirements.

      Integration with BI tools 

      The Unity Catalog seamlessly integrates with popular BI tools like Power BI and Tableau, facilitating easy data consumption and analysis. This integration enables users to leverage their preferred BI tools while benefiting from the robust data governance provided by the Unity Catalog.

      By leveraging the Databricks Unity Catalog, organizations can achieve a higher level of data governance, security, and compliance, ultimately enhancing their data management capabilities and driving better business outcomes.

      What is an example of a use case?  

      Scenario 

      A financial services company, FinCorp, manages large volumes of sensitive data, including customer information, transaction records, and financial reports. Ensuring data privacy, security, and compliance with regulations like GDPR and CCPA is crucial. FinCorp also needs to streamline data access for analysts, data scientists, and business users while maintaining strict access controls. 

      FinCorp objectives 

      • Centralize Data Governance: Manage data governance policies across multiple Databricks workspaces. 
      • Enhance Security: Implement fine-grained access controls and data masking for sensitive information. 
      • Ensure Compliance: Track data lineage and audit data access for regulatory compliance. 
      • Improve Data Discovery: Provide a searchable catalog of data assets for users. 
      • Facilitate Collaboration: Enable seamless data sharing and collaboration across teams. 

      FinCorp’s implementation of Unity Catalog 

      Step 1: Setup the metastore and attach the workspaces  

      FinCorp currently operates multiple Databricks workspaces, each with its own Hive Metastore. To streamline this setup, we create a Unity Catalog metastore to centralize metadata storage. Then, we attach all of FinCorp’s relevant Databricks workspaces to the centralized metastore.  

      Step 2:  Define data governance policies 

      To enhance data organization and security, we create schemas and tables for different data domains, such as customer data and transaction data. We implement fine-grained access controls to restrict access to sensitive data, ensuring that only authorized users can view or modify it. Additionally, we apply data masking to sensitive columns, further protecting confidential information from unauthorized access.  

      Step 3: Enable data lineage and auditing 

      To ensure comprehensive data governance, we track data lineage to monitor data transformations and usage. Additionally, we use built-in auditing features to log data access and modifications.  

      Step 4: Enhance data discovery  

      To enhance data discovery, we tag and describe data assets to improve their discoverability. Additionally, we utilize the Databricks UI to search and explore data assets.  

      Step 5: Facilitate collaboration 

      To facilitate collaboration, we enable data sharing across teams using Unity Catalog’s access controls. This allows data scientists and analysts to collaborate on data projects while ensuring data security.  

      Benefits 

      • Centralized governance: FinCorp has established a unified platform for managing data governance policies, reducing complexity and ensuring consistency across the organization. 
      • Enhanced security: The implementation of fine-grained access controls and data masking techniques ensures the protection of sensitive information. 
      • Regulatory compliance: Detailed data lineage and comprehensive audit logs enable FinCorp to adhere to regulatory requirements and facilitate efficient audits. 
      • Improved data discovery: Users can effortlessly locate and comprehend available data assets, boosting productivity. 
      • Seamless collaboration: Teams can collaborate efficiently without compromising data security. 

      Use case conclusion 

      By implementing the Unity Catalog, FinCorp can effectively manage and govern its data assets, ensuring security, compliance, and efficient data usage across the organization. 

      In summary  

      Databrick’s Unity Catalog offers a powerful, centralized solution for data governance in the Databricks Lakehouse Platform. It simplifies the management of data assets, enhances security through fine-grained access controls, and ensures compliance with regulatory requirements.

      With features like data lineage, auditing, and a searchable catalog, Unity Catalog empowers organizations to streamline their data governance processes, improve data discovery, and facilitate collaboration across teams. By adopting Unity Catalog, organizations can effectively manage their data assets, ensuring they are used efficiently and securely. 

      Ready to get started? 

      Are you ready to take the next steps with Databrick’s Unity Catalog? Fractal can help. Contact us to get started.

      Read more about the Fractal – Databricks partnership here.

      Fractal’s Trial Run: Elevating Retail Excellence at Microsoft Innovation Hubs

      TR at the MTC LI post

       

      We are thrilled to announce a significant milestone in our partnership with Microsoft. Fractal’s retail business experiments solution, Trial Run, is now featured as a self-standing demo at all Microsoft Innovation Hubs globally. We are proud to stand alongside Microsoft’s and other leading partners’ technologies, showcased in over 40 Hubs across the Americas, Europe, Asia, and Australia. 

      Trial Run is an Azure-based SaaS solution for in-store, in-market, and shopper-focused business idea testing. Powered by AI and advanced analytics, it delivers unique, actionable, and statistically significant insights that can empower retailers’ various business functions.

      Trial Run enables retailers to build, develop, and scale their experimentation capabilities efficiently and affordably. It also allows businesses to make informed, data-backed decisions on pricing, promotions, and merchandising. 

      We invite retailers to experience firsthand Trial Run capabilities under the guidance of Microsoft’s Technology Architects during their next Hub visit.  

      Together, let’s shape the future of retail, one experiment at a time. 

      Learn more about Trial Run on Azure Marketplace. 

      Azure Notification Hubs: Key features and use cases 

      In the digital ecosystem, the efficiency of communication directly correlates with user engagement and satisfaction. Azure Notification Hubs serve as a pivotal element in this communication chain, offering developers a scalable and reliable platform to send notifications across various devices and platforms.

      Azure Notification Hubs is a highly scalable mobile push notification engine by Microsoft that is compatible with all significant platforms like

      • APNs (Apple Push Notification Service)
      • GCM (Google Cloud Messaging)
      • WNS (Windows Push Notification Service)
      • MPNS (Microsoft Push Notification Service)

      The Notification Hub can filter push notifications based on a customer, location, or any custom-defined tag.

      This capability enables quick delivery of millions of notifications to iOS, Android, Windows, or Kindle devices.

      With the Notification Hubs tagging functionality, you can send push alerts to everyone at once or narrow it to specific customer devices. Tags allow you to segment users by activity, interest, geography, or preference, ensuring that the right content gets delivered to the correct person at the right time. A tag could be up to 120 characters long string that includes alphanumeric and non-alphanumeric characters.

      Azure Notification Hubs: Key features and use cases The notification tagged with Data_Scientist only reaches the device registered with the tag Data_Scientist, as shown in the diagram.

      While sending a push notification with Notification Hubs, tag expressions allow you to target specified groups of devices or, more precisely, registrants. We can target individual device registrations only by associating tags with specific device registrations and then targeting those tags.

      Both commercial and consumer scenarios benefit from Notification Hubs. Here are a few things you can do:

      • Send minimal latency breaking news alerts to millions of people.
      • Send coupons to user segments who are interested based on their location.
      • For media, sports, social, gaming, or e-commerce apps, send event or deals-related notifications to users or broadcast to groups.
      • To attract and advertise to customers, push promotional information to apps.
      • To notify a user about the status of an activity performed.

      Push notifications play a crucial role in boosting user engagement in consumer apps and delivering real-time business updates in enterprise applications. They are efficient on mobile devices, customizable for senders, and remain accessible even when apps are not active, making them the preferred communication service from apps to users.

      Below in this article, we will cover:

      1. Configuring the Notification Hub for APNs
      2. Prerequisite for Registering / Viewing / Deleting the devices with Notification Hub,
      3. Registering the devices,
      4. Viewing the registered devices,
      5. Deleting the registered devices.

       

      1. Configuring the Notification Hub for APNs

      Prerequisites:

      • A Notification Hub deployed in the Azure portal. In case you need to create a Notification Hub refer to this Documentation.
      • In case of Certificate-based authentication, you must have a .p12 certificate file available with the password.
      • In the case of Token-based authentication, you will require Team ID, Bundle ID, Key ID, and a Token value from a .p8 certificate file.
      • You can generate the above prerequisite details for authentication modes by following this Documentation.

      To configure the Notification Hub with Apple Push Notification Services, we have two Authentication modes available as below:

      • Certificate-based
      • Token-based

      You can select either Certificate or Token,

      For Certificate-based mode:

      • Select the file icon and select the .p12 file from your local system that you want to upload.
      • Enter a password associated with the certificate file.
      • Select Sandbox mode while in development, select Production mode to send push notifications to users who purchased your app from the store.

      Azure Notification Hubs: Key features and use cases 

      For Token-based mode:

      • Enter the values for Key ID, Bundle ID, Team ID, and Token (obtained from a .p8 certificate file).
      • Like Certificate-based mode, Select Sandbox mode while in development and select Production mode to send push notifications to users who purchased your app from the store.

      Azure Notification Hubs: Key features and use cases 

       

      2. Prerequisite for Registering / Viewing / Deleting the devices with Notification Hub

      To Register, View, or Delete the devices with the Notification Hub, we have HTTP’s PUT, GET and DELETE REST API methods. The first and foremost requirement to access these APIs is to generate the SAS Token, used as Authorization Header.

      The Authorization Header should consist of Token generated as specified in SAS Authentication and authorization with Microsoft Azure Active Directory Access Control.

      Microsoft provides a guide to generate SAS tokens using multiple programming languages for using Azure Service Bus REST APIs in this Document.

      For the Python approach to generate SAS tokens for using Azure Service Bus REST APIs, we would require a few details as mentioned below:

      • Namespace Name: The Namespace resource name associated with Notification Hub (namespace could contain one or multiple hubs).
      • Notification Hub Name: The Notification Hub resource name.
      • SAS Name: The Shared Access Key Name.
      • SAS Value: The Shared Access Key Value.

      You can find the Namespace name and Notification Hub name from the portal when you navigate to Notification Hub in the overview section, as shown below:

      Azure Notification Hubs: Key features and use cases 

      You can find those details for the SAS name and value parameters when navigating to Access policies under the Manage section in the left panel.

      You can find two policies with different permissions. We require the details from the “DefaultFullSharedAccessSignature” policy since it has Full permissions (Listen, Manage, Send).

      Azure Notification Hubs: Key features and use cases 

      We require copying the connection string from DefaultFullSharedAccessSignature policy, which looks like something shown below and contains the Shared Access Key name and its value.

      Azure Notification Hubs: Key features and use cases 

      These details would be required to generate the SAS Token as Authorization Header for accessing APIs.

      3. Registering the devices with Notification Hub

      Using a Registration or Installation approach, we can accomplish the registration of devices with a Notification Hub.

      An installation/registration will associate the PNS handle with tags that we maintain to send a notification to the appropriate user rather than broadcasting them. In the case of APNs, the PNS handle is a unique device token for each device, and to segregate device handles; we can then use Tags.

      Note: Notification Hubs allow a maximum of 60 tags for each device. 

      Compared to the Registration approach, an Installation approach is a much-enhanced approach to registering a device that provides multiple push-related properties. It is the recently developed and enhanced approach to registering your devices with Notification Hub.

      Microsoft provides REST APIs for registration and installation, which are user-friendly and require minimal details. Just a Device Token for APNs is enough to register the device with Notification Hub. At the same time, the Tags are an optional element.

      Below are a few critical points discussed to choose Installation over Registration for registering the device with Notification Hub:

      • We could use installation REST APIs to repeatedly make that same call while producing the same result, without any second thought of creating Duplicate registration. It would overwrite the existing registration while reflecting the latest change, if made any. 
      • The Installation approach provides an exceptional Tag format $InstallationId:{<INSTALLATION_ID_GOES_HERE>}, which enables the Notification Hub to classify the unique devices and send the notifications to particular devices rather than broadcasting them. 

      Suppose a user provides Installation id as testinstallation123 for a particular device while registering with Notification Hub. He could then direct the message to the specific device only by applying $InstallationId:{testinstallation123} Tag, without even doing any extra programming stuff to do so. 

      This article would move forward with the Installation Approach and a popular tool like Postman to make API calls to register, view, and delete the devices with Notification Hub. 

      Microsoft provides multiple REST API Methods to perform these tasks, and we would be using Installation REST API for registering the devices with our Notification Hub: 

      Prerequisites:  

      • Namespace Name 
      • Notification Hub Name 
      • SAS Token for Authorization Header 
      • PNS Handle from iOS device 

      As discussed earlier in Prerequisite for Registering / Viewing / Deleting the devices with Notification Hub section, we will require the SAS Token generated for Authorization Header, which will look similar to as shown below: 

      Azure Notification Hubs: Key features and use cases 

      And the request body would contain parameters and their value as below: 

      • installationid: Any uniquely identifiable string value. 
      • tags(optional): Any string value that can uniquely identify or group the devices (#,:, _ and – are only accepted characters).  
      • platform: APNS. 
      • pushchannel: The PNS handle or device token from the device (device token must be a hexadecimal digit). 

      The Request Body would look something similar to this: 

      Azure Notification Hubs: Key features and use cases 

      Once we send the request, the generated response with status code is 200 with an empty body, which means we have successfully created or overwritten and installation. 

      Azure Notification Hubs: Key features and use cases 

      4. Viewing the registered devices with Notification Hub

      There are multiple possible ways to retrieve the registrations, and you can retrieve all registrations with a Notification Hub or even retrieve selected registrations using the installationid.

      MSFT provides multiple REST APIs to deal according to the scenario if we have installationid to retrieve particular registration using Read installation or even retrieve all registrations made with the Notification Hub using Read registrations.

      We would retrieve the earlier created registration where we used installationid as 21-2121. To verify if the update gets reflected, we will update its tags and then re-retrieve the registration.

      We would make a GET call by providing only the installationid. The only prerequisites here would be the installationid and the SAS Token for Authorization:

      Azure Notification Hubs: Key features and use cases 

      We would now add a new tag to the same registration and see if it reflects the update. To add the tag, we would be using the earlier API that we used to register the device, as it will then overwrite an installation.

      Azure Notification Hubs: Key features and use cases 

      To verify the latest change reflects, let’s now retrieve the installation. 

      Azure Notification Hubs: Key features and use cases 

      We were able to retrieve the registered device using the installationid with the latest reflected change. 

      We would use the similar API below that would return all registered devices with Notification Hub: 

      Azure Notification Hubs: Key features and use cases 

      We can even see that the installation has generated an exceptional Tag format, which we discussed earlier $InstallationId:{21-2121}, which would help Notification Hub to classify the unique devices and send the notifications to particular devices rather than broadcasting them.

      5. Deleting a registered device

      To delete multiple registrations in a single go, we currently don’t have any REST API. Possibly, there are two ways to delete a registered device using either the installationid or registration id. 

      We would demonstrate by attempting to delete a registered device using the installationid as 21-2121, using the DELETE Method as discussed in this Documentation. 

      Azure Notification Hubs: Key features and use cases 

      The installation was successfully removed, and we could see the DELETE Method returned the 204 status code. Even if the installation was not present / not found, the service will return the status code 204. 

      Later, when we try to retrieve the same registered device after it gets deleted, the GET method used to retrieve devices would return the status code as 404, stating it did not retrieve the installation as below: 

      Azure Notification Hubs: Key features and use cases 

      Now that we have completed this demonstration, you can configure your Notification Hub to send alerts and notify your customers about the deals and offers via a quick, simple notification sent using Azure.  

      In one of our projects for a client, we leveraged Notification Hubs to notify the mobile device users whenever an image successfully uploads to Azure Data Lake Storage (ADLS) containers. Based on specific parameters set to validate the face in the image, a Machine Learning model would accept/reject the uploaded image and notify the device using Azure Notification Hubs. 

      So why wait? Empower your app with Azure Notification Hub’s powerful capabilities. Contact us to seamlessly integrate and leverage this solution for targeted customer engagement and real-time business updates.  

      Enhance user interaction across iOS, Android, Windows, and Kindle devices with Fractal’s expertise in deploying Azure Notification Hub to drive app-to-user communication like never before. 

      Tenant-to-tenant Power BI data migration

      In today’s dynamic landscape, Power BI stands out as a pivotal and indispensable tool. It has swiftly become the go-to choice for organizations seeking robust reporting and comprehensive dashboards. Within its data-rich framework lie the essential insights that drive informed business decisions.

      Moreover, there are many scenarios where an organization may need to migrate its Power BI tenant. For example, if Company A acquires Company B or Company B merges with Company A, it is essential to transfer all the Power BI data of Company B to Company A’s tenant. Power BI migration can be tricky as there are limited support options for Power BI migration compared to other tools such as Power Apps or Power Automate.

      This is where Fractal comes into the picture with its in-house solution that companies can leverage to migrate their Power BI data from one tenant to another.

      What is a tenant?

      A tenant represents an organization. It’s a dedicated instance of Azure Active Directory AD that an organization or app developer receives at the beginning of a relationship with Microsoft. For example, this relationship could start with signing up for Azure, Microsoft Intune, or Microsoft 365.

      Migrating using a third-party migration tool

      Image source: Microsoft 

      Common challenges of tenant-to-tenant migrations

      Tenant-to-tenant migration involves multiple parties and teams within each organization, and these groups will be interdependent on the migration tasks. A roadmap that outlines the task owners and their dependencies must be followed.

      When Power BI assets, such as reports, datasets, dataflows, etc., are published in a new tenant, their GUIDs (globally unique IDs) change. Thus, these GUIDs need to be carefully modified to make the dependent assets work.

      In addition, when company mergers take place, it is often that their SharePoint sites and SharePoint lists also get migrated. Power BI reports using these SharePoint sites would not work as migration of these SharePoint sites would change the Object IDs on the SharePoint side. So, new Object IDs must be placed on the Power BI side to make it work.

      There are many such corner cases in tenant-to-tenant Power BI data migrations. For brevity, we will not delve deeply into them. It is essential that these migrations are handled with careful attention and coordinated effort among stakeholders to ensure their success.

      How to perform a tenant-to-tenant Power BI data migration

      Pre-requisites

      • Preferably a global admin in source tenant
      • Power BI Pro or Premium Per User Licenses as per requirement in both source and target tenant
      • Apart from personal accounts in both tenants, a common service account to setup refreshes in the target tenant

       

      Instructions 

      Step 1: Analyses of Power BI data in source tenant 

      All the information related to the Power BI assets is extracted first. This includes information on workspaces, reports, gateways, workbooks, datasets, etc. This would provide us with the intuition on what all thing needs to be migrated. 

      Step 2: Download all the reports 

      Download all the reports via automated scripts and keep them on local system. The reports having incremental refresh implemented can’t be downloaded. For this, we need to ask developers to provide us with the source .pbix files. Make sure that all developers stop their developments before you start downloading the reports to avoid version mismatch.  

      Step 3: Setting up target tenant 

      Ask your clients to setup all the infrastructure such as gateways, data sources, and VM in target tenant as they have it in their source tenant. Moreover, they also need to provide access on these gateways and data sources to the person who is migrating. 

      Step 4: Workspace creation 

      Create all the required workspaces in target tenant with the same configuration as source tenant. 

      Step 5: Modify reports 

      We need to modify reports where source applications are also being migrated such as SharePoint, etc.  

      For datasets, which are ingesting data from Power BI Dataflows or any other Power BI services asset, we need to modify their asset IDs in reports.  

      Step 6: Publish reports 

      Upload/Publish all the reports and datasets in their respective workspaces via automated scripts. 

      Step 7: Attach the data sources to the datasets 

      Using the common service account, attach the respective data sources to their datasets using automated scripts. Using the same common account, setup the scheduled refreshes using the automated scripts. 

      Step 8: Create all related Power BI assets 

      Create all dashboards, apps, workbooks, scorecards, and others which is created in only Power BI services from scratch manually. 

      Step 9: Provide access 

      Provide same users access on workspaces and reports as they had in their source tenants. 

      Step 10: Power BI capacities 

      If applicable, attach the workspaces to their capacities in target tenant. But first new capacities need to be setup in target tenant. 

      Step 11: User Acceptance Testing (UAT) 

      Ask users to perform UAT and inform in case of any discrepancies. 

       

      Limitations 

      • Assets that are created in Power BI services cannot be migrated. They need to be created from scratch in target tenant. 
      • Source tenant capacities cannot be reused in target tenant. New capacities need to be bought and setup. 

       

      Why Fractal?

      There are a few advantages of working with a Microsoft Solutions Partner like Fractal to help you with your tenant-to-tenant Power BI data migration.

      • Microsoft does not provide services for Power BI data migration from one tenant ID to another tenant ID. Fractal has the right tools and expertise to support clients looking for such requirements.
      • There are also many dependencies in the whole process. Fractal has the proper roadmap on the execution of whole migration which includes steps for all the parties and has helped many clients in migrating their data.

      Fractal demonstrated its expertise by assisting a client during a company acquisition. The client needed to migrate over 1500 Power BI reports to the parent tenant. Fractal not only devised a tailored execution plan for the migration but also completed it ahead of schedule. Remarkably, after the migration, end users continued to enjoy a seamless experience with the reports in the parent tenant, just as they did in the old tenant.

      If you want to learn more about tenant-to-tenant Power BI migration? Contact us to speak to one of our experts.

       

      References:  

      Governing LLMs through enhancing techniques

      Generative AI has been transformative across industries, helping businesses to accomplish tasks quicker and more efficiently than ever before. It has been used to create text and visual content, get answers to questions, generate code, and more. Though generic large language models (LLMs), such as the models available through Amazon Bedrock, are incredibly robust and valuable for accomplishing a variety of tasks, there are techniques that organizations can use to improve the quality of their computer-generated content. 

      Starting with the end-user, the first method of improving the output of generative AI is to craft the perfect prompt through so-called prompt engineering. Organizations can also help their users get more precise output by governing the LLMs themselves. There are two ways an organization could do this by combining approaches across two axes: 

      • Context optimization: Providing more specific and detailed information for the LLM to draw from. 
      • LLM optimization: Adapting the LLM to accomplish specific tasks or work within a given domain. 

      Let’s go into more detail about how organizations can maximize the effectiveness of their generative AI applications below. 

      Governing LLMs through enhancing techniques

      Prompt engineering 

      Prompt engineering refers to crafting specialized and effective prompts to guide language models and optimizing the context for the desired outputs. Put simply, that means writing the prompt in a way to get exactly what is expected. Here are five guidelines to help end users with their prompt writing. 

      • Be specific and contextualize: Clearly define the topic and provide context. Specify the audience, tone, and timeframe. The more precise the prompt, the better the AI’s response will be. 
      • Iterate for refinement: Use an iterative approach. Refine the prompt based on initial outputs. Adjust and experiment to get more useful information. 
      • Include an action or task: Add an action or task for the AI to complete. For instance, ask it to summarize, outline, list, plan, compare, or predict. This helps guide the AI toward a specific outcome. 
      • Set parameters: Specify parameters such as word count, exclusions, or formatting preferences (e.g., subheadings, paragraphs). This ensures the AI generates content that aligns with the end-user’s requirements. 
      • Avoid bias: Be mindful of bias. Biased prompts can lead to biased outputs. For example, asking a leading question may produce a skewed answer. Aim for balanced prompts that explore both sides of an issue. 

      Prompt engineering is a practice that is accessible to end users in any generative AI product they use, whether those products use general LLMs or those managed by their organization via Amazon Bedrock. 

      From an organizational perspective, companies can help end-users maximize the use of their GenAI apps by providing instruction and guidance on how to properly craft prompts for a given task. Generative AI is a new technology that many end-users are still getting used to. Giving them resources to learn more quickly is one of the best ways to help them get the most out of the app. 

      Context Optimization: Retrieval Augmented Generation (RAG) 

      To further increase the quality and usefulness of AI-generated content, organizations can also provide LLMs with additional information before the prompt so that the LLM executes prompts in a certain way. Called Retrieval Augmented Generation (RAG), this method allows organizations to pre-define the scope or context of the prompt. Here is an example of how it might be used in real life. 

      Scenario: 

      Imagine you are an executive at an electronics company that sells devices like smartphones and laptops. Your company wants to provide detailed answers about its products to customers, journalists, and investors. You have a vast knowledge base containing information about various product specifications, features, and use cases. 

      Customer Query: “What makes our latest smartphone stand out?” 

      Without RAG: 

      In this scenario, without RAG the LLM can produce a result on the prompt, however, the LLM lacks specific details about your products beyond what it learned during training. 

      Response (without RAG): “Our latest smartphone is a great device. It has a powerful processor, a nice display, and a good camera. The battery life is decent too.” 

      Using RAG: 

      By incorporating the knowledge base into the LLM’s prior knowledge, when you ask the LLM a question about a specific product, it can retrieve relevant information from the knowledge base and integrate it into the generated response. The result is a much more targeted and specific response. 

      RAG-Enhanced response: “Our latest smartphone boasts a powerful Snapdragon processor, a stunning AMOLED display, and a triple-camera setup for exceptional photography. Additionally, its long-lasting battery ensures uninterrupted usage throughout the day.” 

      RAG can be enabled through various approaches, and Amazon Bedrock offers multiple tools that can be used depending on the selected approach. 

      LLM optimization: Fine-Tuning 

      To further improve the results of generative AI responses, organizations can fine-tune pre-trained LLM models to specific tasks and domains. There are three ways they can approach this customization: 

      Supervised Learning Fine-Tuning:  

      • In supervised learning, organizations provide a predefined correct answer (ground truth) for a specific task. 
      • The LLM is then fine-tuned using labeled data to generate more accurate and contextually relevant responses. 
      • For example, if an organization wants to create an LLM for code completion, they can fine-tune it using a dataset of code snippets and their corresponding completions. 

      Reinforcement Learning from Human Feedback (RLHF):  

      • RLHF involves training the LLM based on feedback from human evaluators. 
      • Human evaluators rank different model responses based on quality, relevance, and correctness. 
      • The LLM learns from this feedback to improve its performance over iterations. 
      • RLHF is particularly useful when there is no predefined correct answer, as it allows the model to learn from human preferences and adapt accordingly. 

      Domain-Specific Fine-Tuning:  

      • Organizations can fine-tune LLMs on domain-specific data. 
      • For example, a healthcare organization can LLM fine-tuned on medical texts to provide accurate diagnoses or treatment recommendations. 
      • Similarly, in finance or technology, fine-tuning on relevant data enhances the LLM’s utility for specific tasks. 

      This fine-tuning strikes a balance between generalization and specialization, allowing LLMs to excel in specific contexts while retaining their overall capabilities. 

      Organizations can fine-tune LLMs for their users by using Amazon Bedrock. 

      Putting it all together 

      Of course, these techniques aren’t mutually exclusive. You could use one or both of them depending on your needs and the resources available in your organization. 

      If you are looking for help in putting the pieces together, Fractal experts can help you plan and implement a solution that will work for you. 

      Contact us to learn more. 

      Advancing Streaming Piracy Detection and Prevention: A Comprehensive Approach

      In the dynamic world of Media & Entertainment, the fight against content piracy is a constant battle. Despite the industry’s efforts to implement legal and technological safeguards, piracy continues to be a significant issue, affecting revenue and exposing consumers to potential risks such as malware and identity theft. The rise in digital piracy is fueled by factors such as the demand for immediate content access, economic challenges, and the uncontrolled use of generative AI technology.

      A comprehensive strategy is essential to combat piracy effectively. This strategy should include educating consumers, deploying advanced technology, and establishing strong legal frameworks. It is also vital for industry stakeholders to work together to develop efficient anti-piracy measures and explore new business models that improve content accessibility and user experience.

      A cutting-edge anti-piracy solution utilizes AI, scalable data, and cloud engineering to provide real-time insights and protect content. Such a solution, hosted on Microsoft Azure, will need to feature capabilities such as near real-time IP and device monitoring, automated risk scoring, and investigation of suspicious devices to identify and address piracy threats.

      Fractal's Anti-piracy framework

      To achieve this, a solution needs to be based on three fundamental principles: capturing dynamic user behavior, using AI and ML for anomaly detection and network analysis, and applying design thinking to create user-centric solutions. The outcome will then be a robust, transparent, and actionable anti-piracy system that empowers analysts and streamlines the decision-making process.

      By implementing this intelligent anti-piracy framework, a content provider can understand pirate behavior, adapt to changing piracy patterns, and reduce reliance on manual decision-making.

      The Media & Entertainment industry must address piracy with a holistic approach. It is crucial for the Media & Entertainment industry to protect content creators and consumers from the negative impacts of content piracy and promote a sustainable digital environment.

      Download our eBook to learn more.

      Governing LLMs through enhancing techniques

      Generative AI has been transformative across industries, helping businesses to accomplish tasks quicker and more efficiently than ever before. It has been used to create text and visual content, get answers to questions, generate code, and more. Though generic large language models (LLMs), such as the models available through Google Cloud Vertex AI, are incredibly robust and valuable for accomplishing a variety of tasks, there are techniques that organizations can use to improve the quality of their computer-generated content.

      Starting with the end-user, the first method of improving the output of generative AI is to craft the perfect prompt through so-called prompt engineering. Organizations can also help their users get more precise output by governing the LLMs themselves. There are two ways an organization could do this by combining approaches across two axes:

      • Context optimization: Providing more specific and detailed information for the LLM to draw from.
      • LLM optimization: Adapting the LLM to accomplish specific tasks or work within a given domain.

      Let’s go into more detail about how organizations can maximize the effectiveness of their generative AI applications below.

      Governing LLMs through enhancing techniques

      Prompt engineering

      Prompt engineering refers to crafting specialized and effective prompts to guide language models and optimizing the context for the desired outputs. Put simply, that means writing the prompt in a way to get exactly what is expected. Here are five guidelines to help end users with their prompt writing.

      • Be specific and contextualize: Clearly define the topic and provide context. Specify the audience, tone, and timeframe. The more precise the prompt, the better the AI’s response will be.
      • Iterate for refinement: Use an iterative approach. Refine the prompt based on initial outputs. Adjust and experiment to get more useful information.
      • Include an action or task: Add an action or task for the AI to complete. For instance, ask it to summarize, outline, list, plan, compare, or predict. This helps guide the AI toward a specific outcome.
      • Set parameters: Specify parameters such as word count, exclusions, or formatting preferences (e.g., subheadings, paragraphs). This ensures the AI generates content that aligns with the end-user’s requirements.
      • Avoid bias: Be mindful of bias. Biased prompts can lead to biased outputs. For example, asking a leading question may produce a skewed answer. Aim for balanced prompts that explore both sides of an issue.

      Prompt engineering is a practice that is accessible to end users in any generative AI product they use, whether those products use general LLMs or those managed by their organization via Vertex AI.

      From an organizational perspective, companies can help end-users maximize the use of their GenAI apps by providing instruction and guidance on how to properly craft prompts for a given task. Generative AI is a new technology that many end-users are still getting used to. Giving them resources to learn more quickly is one of the best ways to help them get the most out of the app.

      Context Optimization: Retrieval Augmented Generation (RAG)

      To further increase the quality and usefulness of AI-generated content, organizations can also provide LLMs with additional information before the prompt so that the LLM executes prompts in a certain way. Called Retrieval Augmented Generation (RAG), this method allows organizations to pre-define the scope or context of the prompt. Here is an example of how it might be used in real life.

      Scenario:

      Imagine you are an executive at an electronics company that sells devices like smartphones and laptops. Your company wants to provide detailed answers about its products to customers, journalists, and investors. You have a vast knowledge base containing information about various product specifications, features, and use cases.

      Customer Query: “What makes our latest smartphone stand out?”

      Without RAG:

      In this scenario, without RAG the LLM can produce a result on the prompt, however, the LLM lacks specific details about your products beyond what it learned during training.

      Response (without RAG): “Our latest smartphone is a great device. It has a powerful processor, a nice display, and a good camera. The battery life is decent too.”

      Using RAG:

      By incorporating the knowledge base into the LLM’s prior knowledge, when you ask the LLM a question about a specific product, it can retrieve relevant information from the knowledge base and integrate it into the generated response. The result is a much more targeted and specific response.

      RAG-Enhanced response: “Our latest smartphone boasts a powerful Snapdragon processor, a stunning AMOLED display, and a triple-camera setup for exceptional photography. Additionally, its long-lasting battery ensures uninterrupted usage throughout the day.”

      RAG can be enabled through various approaches, and Google’s Vertex AI offers multiple tools that can be used depending on the selected approach.

      LLM optimization: Fine-Tuning

      To further improve the results of generative AI responses, organizations can fine-tune pre-trained LLM models to specific tasks and domains. There are three ways they can approach this customization:

      Supervised Learning Fine-Tuning: 

      • In supervised learning, organizations provide a predefined correct answer (ground truth) for a specific task.
      • The LLM is then fine-tuned using labeled data to generate more accurate and contextually relevant responses.
      • For example, if an organization wants to create an LLM for code completion, they can fine-tune it using a dataset of code snippets and their corresponding completions.

      Reinforcement Learning from Human Feedback (RLHF): 

      • RLHF involves training the LLM based on feedback from human evaluators.
      • Human evaluators rank different model responses based on quality, relevance, and correctness.
      • The LLM learns from this feedback to improve its performance over iterations.
      • RLHF is particularly useful when there is no predefined correct answer, as it allows the model to learn from human preferences and adapt accordingly.

      Domain-Specific Fine-Tuning: 

      • Organizations can fine-tune LLMs on domain-specific data.
      • For example, a healthcare organization can LLM fine-tuned on medical texts to provide accurate diagnoses or treatment recommendations.
      • Similarly, in finance or technology, fine-tuning on relevant data enhances the LLM’s utility for specific tasks.

        This fine-tuning strikes a balance between generalization and specialization, allowing LLMs to excel in specific contexts while retaining their overall capabilities.

        Organizations can fine-tune LLMs for their users by using Vertex AI.

        Putting it all together

        Of course, these techniques aren’t mutually exclusive. You could use one or both of them depending on your needs and the resources available in your organization.

        If you are looking for help in putting the pieces together, Fractal experts can help you plan and implement a solution that will work for you.

        Contact us to learn more.

        Fractal’s PYO Autonomous AI solution presentation on Microsoft booth at Hannover Messe ‘24

        Fractal is proud to announce that its Production Yield Optimization (PYO) solution will be presented Wednesday, April 24th at 10 am on Microsoft’s stand (Hall 17, Stand G06) at the 2024 Hannover Messe.

        PYO is a proven Autonomous AI solution that enables manufacturers to reduce waste and optimize manufacturing output with AI agents that dynamically fine-tune existing control systems parameters.

        PYO is built on Azure Machine Learning and leverages industry-standard deep reinforcement learning (DRL) techniques and libraries to translate subject matter experts’ expertise into so-called “reward functions” that allow the AI agent to self-train through a custom-built AI simulator. This expertise transfer is commonly referred to as “Machine Teaching”.

        Fractal’s end-to-end DRL experience, accelerators, and best practices help manufacturers customize PYO to their unique needs. This is one of the reasons why companies such as PepsiCo have trusted Fractal for their PYO project on their Cheetos product line.

        How does PYO work?

        The PYO AI agent design, training, and deployment is a multi-step and iterative process.

        PYO technical architecture

        Once the initial Machine Teaching process is completed, the manufacturing SMEs expertise will help select the appropriate process training data and define the best reinforcement learning reward function. The AI agent is then trained using the new simulation.

        The trained agent is then validated virtually with the simulation. However, in most cases, multiple cycles of simulation fine-tuning, reward function adaptation, agent training, and virtual validation will be required to converge to a satisfactory PYO AI agent. The agent is then trialed and fully deployed in production.

        The AI agent sends control signals to the simulation or the production system. The reward function will measure the difference between the expected simulation or system state versus its actual state and will modify the agent’s deep neural network weights accordingly. Depending on the system controlled, this training loop will run between hundreds of thousands to millions of times.

        Why should you consider PYO?

        • Optimize production with AI: PYO agents learn, through SME expertise transfer, to optimize production for complex and changing environments. Those agents will help with both manufacturing line-level and human-level challenges.
        • Real-world AI solution: PYO uses DRL and simulations to train the AI agents without the need for pre-existing labeled datasets.
        • Fractal’s end-to-end expertise: Bringing an AI agent from design to deployment requires a large set of data and AI skills. Fractal is a recognized Microsoft Solution Partner with the expertise to support you throughout your data and AI transformation journey.

        If you want to learn more about PYO and how it can help you achieve your production goals, come and attend our presentation at the Microsoft booth on Wednesday, April 24th, at 10 am.

        You can also check out our PYO solution page here.

        Digital twin vs. simulations: the quick cheat sheet

        What is a digital twin?

        Wikipedia defines a digital twin as a “virtual representation that serves as the real-time digital counterpart of a physical object or process.”

        In theory, a digital twin will gather input from connected sensors, machinery, and people to store and display them in a cloud-hosted application. However, besides being able to look back in time what happened when and perform some post-mortem analysis, a digital twin limited to a backward-looking view won’t have much business interest.

        Therefore, often digital twins will also integrate simulations of what is represents. Those simulations can be at the device, process, or even plant level. It will allow users to leverage this combination of real-time data and system-level behavior modeling through the simulator for multiple use cases.

        Digital twin vs. simulations: the quick cheat sheet

        For instance, a digital twin can be used to:  

        • Replay a system behavior based on historical data  
        • Do advanced “what-if” analysis before deciding which path to choose  
        • Train new operators on virtual processes before letting them work on the actual real-life process  
        • Simulate a process to train an AI Agent using Deep Reinforcement Learning (DRL)
        • And more depending on the industry that digital twins are used in 

        Digital twins represent a great business improvement opportunity for customers across industries. However, what they are, how they work, how they can positively impact operations, and what technologies are involved with digital twins are often questions customers struggle to answer” – Manish Amin, Data & AI and IoT Principal and Advisor, Microsoft

        Digital twin vs. simulator  

        A simulator’s scope is often limited to a particular piece of equipment or process, although not always. Once programmed or trained, the simulator will run separately from the real-life process. 

        Conversely, a digital twin will often encompass a broader process comprised of multiple pieces of equipment, and it will remain connected to the live system to represent it faithfully. 

        Therefore, a simplified way to think about the difference between a digital twin and a simulator is to consider that a digital twin is a simulation whose states (inputs, outputs) are updated to accurately reflect their real-life value. A simulator could end up drifting from real life or even provide wrong data; a digital twin won’t if it remains connected. 

        Digital twin vs. simulations: the quick cheat sheet

        Conversely, simulators operate separately from a real-life process and can even be developed without an existing process to test hypothesis. 

        Enabling technologies for digital twins and simulations 

        To build and run a digital twin, several technology blocks are required. 

        1. A simulation engine 
        2. Real-time process data collection technologies 
        3. Cloud and data services to collect, store and analyze the process and simulation data 

        How do you build a simulation? 

        Digital twin vs. simulations: the quick cheat sheet

        There are multiple ways to build a simulator, and the three most used are: 

        • Physics-based simulators
        • Software package-based simulators using products such as AnyLogic or Simulink 
        • AI or data-based simulators that train AI models, most often deep neural networks-based ones  

        For the latter approach, data-logging only digital twins can be used to create the dataset necessary to train this AI simulator. The historical process data (both inputs, states, and outputs) that the digital twin recorded can provide the breadth and quantity of labeled data required for those types of AI simulators’ supervised learning. This is one of the areas where a partner with extensive data science experience can significantly help with the speed and quality of the simulation development.

        Additional enabling technologies  

        Digital twin vs. simulations: the quick cheat sheet

        To collect real-time process data, smart sensors using technologies such as Azure IoT are going to be required. Adding intelligence at the edge to existing sensors or deploying new smart sensors such as vision AI ones, we will be able to instrument all the relevant process inputs and outputs. 

        This real-time data and the simulator(s) will be hosted on an appropriate cloud platform to enable the above-mentioned use cases. Solutions such as Azure Digital Twin will enable easy integration of those elements and access to device, process, line (or building), or plant (or campus) digital twins. 

        “IoT is inextricably linked with digital twins. To create a comprehensive digital twin of a manufacturing environment, one must connect every major process on a manufacturing floor to IoT for process digitization, modeling, and simulation”Mohammad Ahmed, Azure Infrastructure Specialist, Microsoft 

        Although this article somewhat oversimplifies both what digital twins are and what is required to build and run them, it provides a base for more in-depth research if a more comprehensive understanding of the subject is required. To help with this additional research, we listed a few links below to get you started.   

        Also, as Fractal possesses the end-to-end technological capabilities required to instrument, build, train, deploy, and maintain digital twins, feel free to contact us, if you are interested in learning more about this topic.  

         

        Additional resources on digital twins: 

        Governing LLMs through enhancing techniques

        Generative AI has been transformative across industries, helping businesses to accomplish tasks quicker and more efficiently than ever before. It has been used to create text and visual content, get answers to questions, generate code, and more. Though generic large language models (LLMs), such as the OpenAI models available through Azure AI, are incredibly robust and valuable for accomplishing a variety of tasks, there are techniques that organizations can use to improve the quality of their computer-generated content.

        Starting with the end-user, the first method of improving the output of generative AI is to craft the perfect prompt through so-called prompt engineering. Organizations can also help their users get more precise output by governing the LLMs themselves. There are two ways an organization could do this by combining approaches across two axes:

        • Context optimization: Providing more specific and detailed information for the LLM to draw from.
        • LLM optimization: Adapting the LLM to accomplish specific tasks or work within a given domain.

        Let’s go into more detail about how organizations can maximize the effectiveness of their generative AI applications below.

        Governing LLMs through enhancing techniques

        Prompt engineering

        Prompt engineering refers to crafting specialized and effective prompts to guide language models and optimizing the context for the desired outputs. Put simply, that means writing the prompt in a way to get exactly what is expected. Here are five guidelines to help end users with their prompt writing.

        • Be specific and contextualize: Clearly define the topic and provide context. Specify the audience, tone, and timeframe. The more precise the prompt, the better the AI’s response will be.
        • Iterate for refinement: Use an iterative approach. Refine the prompt based on initial outputs. Adjust and experiment to get more useful information.
        • Include an action or task: Add an action or task for the AI to complete. For instance, ask it to summarize, outline, list, plan, compare, or predict. This helps guide the AI toward a specific outcome.
        • Set parameters: Specify parameters such as word count, exclusions, or formatting preferences (e.g., subheadings, paragraphs). This ensures the AI generates content that aligns with the end-user’s requirements.
        • Avoid bias: Be mindful of bias. Biased prompts can lead to biased outputs. For example, asking a leading question may produce a skewed answer. Aim for balanced prompts that explore both sides of an issue.

        Prompt engineering is a practice that is accessible to end users in any generative AI product they use, whether those products use general LLMs or those managed by their organization via Azure. In the Microsoft technology stack, users can access these LLMs through Copilot which is available as a standalone app and is integrated into Edge and Microsoft 365.

        From an organizational perspective, companies can help end-users maximize the use of their GenAI apps by providing instruction and guidance on how to properly craft prompts for a given task. Generative AI is a new technology that many end-users are still getting used to. Giving them resources to learn more quickly is one of the best ways to help them get the most out of the app. Microsoft also offers the Copilot Prompt Lab, which can guide users in making more effective prompts.

        Context Optimization: Retrieval Augmented Generation (RAG)

        To further increase the quality and usefulness of AI-generated content, organizations can also provide LLMs with additional information before the prompt so that the LLM executes prompts in a certain way. Called Retrieval Augmented Generation (RAG), this method allows organizations to pre-define the scope or context of the prompt. Here is an example of how it might be used in real life.

        Scenario:

        Imagine you are an executive at an electronics company that sells devices like smartphones and laptops. Your company wants to provide detailed answers about its products to customers, journalists, and investors. You have a vast knowledge base containing information about various product specifications, features, and use cases.

        Customer Query: “What makes our latest smartphone stand out?”

        Without RAG:

        In this scenario, without RAG the LLM can produce a result on the prompt, however, the LLM lacks specific details about your products beyond what it learned during training.

        Response (without RAG): “Our latest smartphone is a great device. It has a powerful processor, a nice display, and a good camera. The battery life is decent too.”

        Using RAG:

        By incorporating the knowledge base into the LLM’s prior knowledge, when you ask the LLM a question about a specific product, it can retrieve relevant information from the knowledge base and integrate it into the generated response. The result is a much more targeted and specific response.

        RAG-Enhanced response: “Our latest smartphone boasts a powerful Snapdragon processor, a stunning AMOLED display, and a triple-camera setup for exceptional photography. Additionally, its long-lasting battery ensures uninterrupted usage throughout the day.”

        RAG can be enabled through various approaches, and Azure offers multiple tools that can be used depending on the selected approach: Azure AI Search , Azure AI Studio, Azure OpenAI Studio, and Azure Machine Learning.

        LLM optimization: Fine-Tuning

        To further improve the results of generative AI responses, organizations can fine-tune pre-trained LLM models to specific tasks and domains. There are three ways they can approach this customization:

        Supervised Learning Fine-Tuning: 

        • In supervised learning, organizations provide a predefined correct answer (ground truth) for a specific task.
        • The LLM is then fine-tuned using labeled data to generate more accurate and contextually relevant responses.
        • For example, if an organization wants to create an LLM for code completion, they can fine-tune it using a dataset of code snippets and their corresponding completions.

        Reinforcement Learning from Human Feedback (RLHF): 

        • RLHF involves training the LLM based on feedback from human evaluators.
        • Human evaluators rank different model responses based on quality, relevance, and correctness.
        • The LLM learns from this feedback to improve its performance over iterations.
        • RLHF is particularly useful when there is no predefined correct answer, as it allows the model to learn from human preferences and adapt accordingly.

        Domain-Specific Fine-Tuning: 

        • Organizations can fine-tune LLMs on domain-specific data.
        • For example, a healthcare organization can LLM fine-tuned on medical texts to provide accurate diagnoses or treatment recommendations.
        • Similarly, in finance or technology, fine-tuning on relevant data enhances the LLM’s utility for specific tasks.

        This fine-tuning strikes a balance between generalization and specialization, allowing LLMs to excel in specific contexts while retaining their overall capabilities.

        Organizations can fine-tune LLMs for their users by using Azure AI Search and Azure Machine Learning.

        Putting it all together

        Of course, these techniques aren’t mutually exclusive. You could use one or both of them depending on your needs and the resources available in your organization.

        If you are looking for help in putting the pieces together, Fractal experts can help you plan and implement a solution that will work for you.

        Contact us to learn more.

         

        References:

        Three pillars of the retail industry: Replenishment, allocation, and transportation 

        The retail industry is one of the most dynamic and fast-paced sectors, comprised of a variety of stakeholders engaged in selling finished products to end-user consumers. In 2022, the U.S. retail sector was estimated at more than seven trillion dollars. The sector is projected to continue to grow, and by 2026 U.S. retail sales are expected to reach approximately USD 7.9 trillion. With the increased demand for consumer goods in different sectors and the ever-increasing choices of products at low costs, investments in the retail sector have also grown over the past few years.

        As there is always an element of change in this industry, new challenges are encountered every day. Take product stockouts, for example. Suppose a customer walks into a grocery store to purchase items of their favorite brand but discovers that the product is not available. Frustrated by this, the customer chooses to either buy another brand or postpone the purchase; both scenarios are unfavorable to the business. The brand image and sales of the product are damaged because of this out-of-stock issue. The out-of-stock situation occurs when the inventory of a particular product is exhausted; this causes a problem for both suppliers and retailers.

        There can be multiple reasons that would cause the product stockout, such as inaccurate inventory data, lack of demand forecasting, or an unseasonal spike in purchasing. Many of these underlying causes of stockouts can be avoided if the business implements adequate processes to be carried out every month.

        To avoid situations like the stockout example above, retail companies need to develop a methodology for streamlining the following operations:

        1. Replenishment
        2. Allocation
        3. Transportation

        3 pillars of retail industry

          These three operations create the three pillars of the retail industry that help monitor real-time insight into customer behavior and understand their buying patterns hence strengthening the retail foundation.

          Replenishment

          Replenishment refers to a situation where the amount of stock left in the store is counted so that the right products are available at an optimal quantity. It is considered an essential aspect of inventory management as it ensures that the right products are being reordered to meet the customer demand.

          In operational terms, the efficiency of store replenishment has a significant impact on profitability. The effectiveness and accuracy of store ordering affect sales through shelf availability and storage, handling, and wastage costs in stores and other parts of the supply chain. By optimizing demand forecasting, inventory management, and setting of order cycles and order quantities by making them more systematic, the gains obtained are significant, often amounting to savings of several percent of total turnover.

          For companies that must manage a large number of SKUs, one of the most effective ways of making store replenishment more accurate, efficient, and cost-effective is by using a replenishment system specifically tailored to business operations. When many different products need to be managed, manual ordering is highly labor-intensive and expensive; this results in companies using the replenishment system.

          An efficient replenishment system reduces process costs, improves inventory turnover, and provides higher service levels. The system constantly monitors the stock, sales, and demand while considering the forecast changes in demand and adjusting the replenishment orders. Recognizing the sales frequency, sales value, or profit margin, the company can control its inventory in such a way that ensures long-term profitability. The replenishment system calculates the safety stock level for each SKU separately and sets them to meet the service level targets with efficiency, considering the predictability of demand.

          Allocation

          In allocation, the new stock of products is distributed to individual store units, such that they maximize the sale of the product and prevent any stock out situation in the future. This process enables the assigning of supplies so that they support the organization’s strategic goals. Having sufficient stock levels is an essential component for any retail business; with the changing consumer habits, it becomes crucial for the stock to be available in the right place at the right time.

          To meet new and increasing demands, retailers need an efficient process to gather, interpret, and analyze data from customer behaviors and habits, which would help get a more localized and specific idea of what is sold at a larger quantity in different locations. Items that are high sellers in one particular area may not sell well in others, so recognizing and monitoring this can ensure that the stock is allocated to the most needed location. Due to this, an opportunity is provided for the retailers to encourage sales by pushing stock of a similar type that a customer may favor at a particular location.

          Transportation

          Transportation plays a significant role in delivering the right stock of products at the right point of delivery. It connects the business to its supply chain partners and influences the customers’ satisfaction with the organization. With the ever-changing customer preferences and as their expectations continue to evolve, the transportation industry is undergoing a dramatic transformation to meet these demands.

          Today, data plays a vital role in shaping how the industry will progress amidst tough competition. Due to the maturation of automation technologies, AI will help the transportation industry to manage drivers and fleet managers. By employing the techniques of AI, fleet and truck adjustments will offer data in real-time, eventually improving the industry’s standard. The safety and retention of the drivers will also increase from these newly acquired standards, and with enhanced access to data, there will be transparent communication between drivers and carriers.

          The average time between goods purchasing and delivery decreases by using real-time findings, making retailers focus on transportation to improve their business performance. The ability to automate insights, alerts, and data exchange more quickly will be the game-changer for this industry.

          These three pillars of retail can be strengthened by adopting in-house solutions and capabilities like Trial Run, customer analytics, supply chain analytics, and store operation analytics.

          How could these solutions help the retail industry?

          Trial Run is a data-driven, cloud-based test management product used to test business ideas for sites and markets using Google Cloud capabilities combined with advanced technologies to improve customer experience, enhance product recommendations, streamline operations, optimize inventory, and enhance the supply chain.

          Trial Run helps in scientific and systematic testing, which can unlock insights and provide direction with the following tests:

          • Marketing and Merchandizing tests
          • In-store experience tests
          • Store operations tests

          Customer Analytics is an AI-driven suite that helps a retailer to know their customers in a better way by understanding the customer needs and preferences from every angle, like acquisition, engagement, retention, and growth, gaining insights that can fuel growth in marketing initiatives, loyalty programs, or eCommerce platforms.

          Supply-chain Analytics is an advanced analytics and intelligent automation solution that helps in end-to-end supply chain visibility to stay competitive, keeping the distribution networks customer-oriented and efficient while reducing environmental impact. It helps in streamlined operations, which results in better cost savings, ultimately delivering more value in every step of the supply chain process.

          Store Operation Analytics helps boost sales productivity and reduce costs across every facet of store operations – from labor, facilities, and inventory management to enhanced customer service and satisfaction.

          All these solutions and capabilities help understand the customer motivations, preferences, and desires to meet their demands and increase sales effectively, hence strengthening the pillars of the retail industry.

          Conclusion

          To meet these growing customer expectations, retailers should give priority to collecting the customer data and analyzing it to support business decisions throughout their value chain. The inventory stocking patterns and shipping routes will shift in relation to patterns informed by this data. Retailers should make a concentrated effort to leverage the data while making critical business decisions, and to remain efficient; they must remain flexible and transform their operations as they capture more insights from their data.

          Over the past 20+ years, Fractal has helped many retail companies make their replenishment, allocation, and transportation operations more efficient by leveraging AI, engineering, and design. If you would like to learn more about optimizing these aspects of your retail business, please contact us to speak with one of our experts.

          Find how Trial Run enabled decision-making and helped clients with increased accuracy in test analysis and better ROI measurement resulting in an annual financial gain of $25Mn.

          Trail Run illustration

          Nowadays, companies want to be able to test business decisions and ideas at a scale large enough to believe the results but also at a scale small enough to reduce the large investments and risks that come with full-scale execution.

          Trial Run helps conduct tests such as altering store layouts and remodeling, loyalty campaigns, and pricing to recommend the best possible tailored rollout to maximize gains. You can now implement new ideas with minimal risk and maximum insight with the power of business experimentation. Trial run helps you:

          • Test each business idea at scale to generate customer insights without excessive spending.
          • Find out why your customers behave the way they do.
          • Learn how your customers will react to your new big idea.

           

          What is Trial Run?

          Trial Run is a data-driven, cloud-based test management platform used to test business ideas for sites, customers, and markets. Trial run is built using Amazon EKS, Amazon Redshift, Amazon EC2, Amazon ElastiCache, and Amazon Beanstalk. It is intuitive for beginners and experts alike and helps companies scale experimentation efficiently and affordably.

          Trial Run supports the entire experimentation lifecycle, which includes:

          Trail Run illustration

           

           

          1. Design: Build a cost-effective and efficient experiment that gives you the data you need to proceed with confidence.
          2. Analyze: Work with variables that provide you with targeted and actionable insights.
          3. Act: Use the generated insights to ensure your new rollout provides your stakeholders with the precise ROI.

          Trial Run offers valuable support across various operational and administrative departments, including Retail, Consumer Packaged Goods (CPG), and Telecommunications.

          Through its scientific and methodical testing approach, Trial Run can uncover fresh perspectives and guide decision-making through a range of tests, including:

          • Marketing and merchandising strategies.
          • Enhancing the in-store experience.
          • Examining store operations and processes.

          These tests are carried out at the store operations and process, product, or consumer levels.

          Trial Run offers a dynamic, affordable, and modern way of experimentation so you can stay relevant in a rapidly changing business environment. Trial Run also helps you to drive experiments through:

          • Driver Analysis: Identify key factors that are significant in driving the business outcomes
          • Rollout simulator: Maximize the ROI of a campaign
          • Synthetic Control Algorithm: Determine the right number of control stores with appropriate weights to create the replica of the test store
          • Experiment calendar: Avoid overlaps in experiments
          • Clean search: Let Trial Run parse the experiment repository and find entities that are available for a test

             

            What you can expect from Trial Run

            • Graphical design elements make it easy to use the program as an expert or a beginner
            • Automated workflows can guide you through the process from start to finish
            • Highly accurate synthetic control results with automated matching processes that only require minimal human intervention
            • Experiments at speed and scale without the hassle of expert teams or expensive bespoke solutions
            • Training, troubleshooting, and best practices from the best in the business
            • Easy pilots to help your new idea go live in as little as 6 to 8 weeks

            Trial Run stands out from other solutions by offering a transparent methodology and easily explainable recommendations. Trial Run utilizes a cutting-edge technique called “synthetic control” for matching, ensuring precise results. Trial Run can be used as a SaaS offering that is easily scalable based on demand and can be hosted on the cloud of customer’s choice. With Trial Run software, customers have unlimited test capabilities, enabling them to design and measure numerous initiatives without any restrictions. Finally, Trial Run success is proven in enterprises, with over 1,000 use cases deployed on our platform.

            How do I get started?

            Are you ready to implement cutting-edge technology to help you build cost-effective and efficient experiments that provide you with the data you need to make decisions?

            If you want to achieve successful Trial Run implementation, get started on the AWS Marketplace.

            Interested in learning more about how Fractal can help you implement Trial Run, contact us to get in touch with one of our experts.

            Trail Run illustration

            Nowadays, companies want to be able to test business decisions and ideas at a scale large enough to believe the results but also at a scale small enough to reduce the large investments and risks that come with full-scale execution.

            Trial Run helps conduct tests such as altering store layouts and remodeling, loyalty campaigns, and pricing to recommend the best possible tailored rollout to maximize gains. You can now implement new ideas with minimal risk and maximum insight with the power of business experimentation. Trial run helps you:

            • Test each business idea at scale to generate customer insights without excessive spending.
            • Find out why your customers behave the way they do.
            • Learn how your customers will react to your new big idea.

             

            What is Trial Run?

            Trial Run is a data-driven, cloud-based test management platform used to test business ideas for sites, customers, and markets. Trial run is built using Azure Kubernetes Services, Azure Synapse Analytics, and Azure Virtual Machines. It is intuitive for beginners and experts alike and helps companies scale experimentation efficiently and affordably.

            Trial Run supports the entire experimentation lifecycle, which includes:

            Trail Run illustration

             

             

            1. Design: Build a cost-effective and efficient experiment that gives you the data you need to proceed with confidence.
            2. Analyze: Work with variables that provide you with targeted and actionable insights.
            3. Act: Use the generated insights to ensure your new rollout provides your stakeholders with the precise ROI.

            Trial Run offers valuable support across various operational and administrative departments, including Retail, Consumer Packaged Goods (CPG), and Telecommunications.

            Through its scientific and methodical testing approach, Trial Run can uncover fresh perspectives and guide decision-making through a range of tests, including:

            • Marketing and merchandising strategies.
            • Enhancing the in-store experience.
            • Examining store operations and processes.

            These tests are carried out at the store operations and process, product, or consumer levels.

            Trial Run offers a dynamic, affordable, and modern way of experimentation so you can stay relevant in a rapidly changing business environment. Trial Run also helps you to drive experiments through:

            • Driver Analysis: Identify key factors that are significant in driving the business outcomes
            • Rollout simulator: Maximize the ROI of a campaign
            • Synthetic Control Algorithm: Determine the right number of control stores with appropriate weights to create the replica of the test store
            • Experiment calendar: Avoid overlaps in experiments
            • Clean search: Let Trial Run parse the experiment repository and find entities that are available for a test

               

              What you can expect from Trial Run

              • Graphical design elements make it easy to use the program as an expert or a beginner
              • Automated workflows can guide you through the process from start to finish
              • Highly accurate synthetic control results with automated matching processes that only require minimal human intervention
              • Experiments at speed and scale without the hassle of expert teams or expensive bespoke solutions
              • Training, troubleshooting, and best practices from the best in the business
              • Easy pilots to help your new idea go live in as little as 6 to 8 weeks

              Trial Run stands out from other solutions by offering a transparent methodology and easily explainable recommendations. Trial Run utilizes a cutting-edge technique called “synthetic control” for matching, ensuring precise results. Trial Run can be used as a SaaS offering that is easily scalable based on demand and can be hosted on the cloud of customer’s choice. With Trial Run software, customers have unlimited test capabilities, enabling them to design and measure numerous initiatives without any restrictions. Finally, Trial Run success is proven in enterprises, with over 1,000 use cases deployed on our platform.

              How do I get started?

              Are you ready to implement cutting-edge technology to help you build cost-effective and efficient experiments that provide you with the data you need to make decisions?

              If you want to achieve successful Trial Run implementation, get started on Azure Marketplace.

              Interested in learning more about how Fractal can help you implement Trial Run, contact us to get in touch with one of our experts.

              Metaverse-and-digital-twins-ft.-image

              What is the metaverse?

              The metaverse has been a trending topic recently. Although there may be some variations in its definition, there are a few key elements that are the foundation of what the (or “a”) metaverse is. Whether we are referring to a metaverse for consumers or for B2B, the metaverse refers to an interconnected network of connected, persistent, and virtual worlds (or spaces) where digital avatars of human users can interact with each other and with elements of the metaverse itself. With the development of virtual reality headsets and projecting developments, the metaverse market is poised to significantly expand.

              Though the recent hype about the consumer metaverse has focused on creating a social virtual space and gaming environment, the “industrial metaverse” has already taken its leap.

              The Metaverse represents a global 3D network of virtual world. The transformation of the web into Web3.0 and the development of blockchain and decentralization help in creating digital assets like NFTs as digital currency using which one can purchase or sell virtual goods.

              The 3 key pillars of the metaverse are:

                • Virtual reality (VR)
                • Augmented reality (AR)
                • 3D visualization of data

              Here virtual reality is computer computer-generated 3D environment that surrounds the user and responds to an individual’s action in a natural way through a head-mounted display such as Oculus Quest. On the other hand, augmented reality is the real-time use of information such as text, graphics, audio, video with real-world objects. “Pokémon Go” is a perfect example of augmented reality where it enables the user to experience the presence of a virtual world. The new Apple Vision Pro headset combines these pillars by allowing users to experience augmented reality and 3D visualization of data in an immersive environment.

              Emerging metaverse technologies

              Looking forward to the development of metaverse technology, many organizations are extending their capabilities. Microsoft hopes to acquire Activision Blizzard, a video game developer and publisher. Also, Unity software has acquired Weta Digital (a digital VFX company) to develop RT3D (real-time 3D). Similarly, NVIDIA is developing the Omniverse to provide the tools metaverse developers need.

              To help accelerate both its first-party and third-party metaverse ecosystem, Microsoft shared what it believes are the key technology stack elements required. More than a brand-new platform, this approach builds on existing and expanding platforms and tools to provide a fairly comprehensive solution to build future B2B metaverses.

              Microsoft-metaverse-technology-stack

              Source: Microsoft Metaverse technology stack

              Metaverse platforms

              In addition to Microsoft, most technology companies, and some more traditional ones, have started to build metaverse platforms or integrations:

              Trending metaverse platforms

              The industrial metaverse

              During the Industrial Metaverse event hosted by RealWear, Brian Vogelsang, Senior Director of Product Management at Qualcomm said: “Industrial metaverse is not a new technology. It is an evolution of technology, a convergence.”

              Digital twins are representations of an intended or actual real-world physical product, system, or process. With digital twin tools such as Azure Digital Twin and 3D visualization tools such as “AltSpaceVR” and “Azure Digital Twins 3D Scenes Studio”, it is possible to visualize real-life process data in a 3D space. It helps link together 3D content, digital twin, and business-specific logic.

              A great example of such use of the industrial metaverse with digital twin is the following Contoso Refinery demo from Microsoft.

              Microsoft-3D-Scenes-Studio-preview-for-Azure-Digital-Twins

              Source: Microsoft 3D Scenes Studio preview for Azure Digital Twins

              Contoso Refinery is an end-to-end demonstration implemented with Azure Digital Twins 3D Scenes Studio. In this demo, users can monitor and diagnose real-time operational data with the visual context of 3D assets. Visualized 3D models can be easily accessed through any web browser. The demo offers a low code builder to add 3D elements to digital twins and define both user interface and business logic visualization.

              Real-world use cases for the industrial metaverse

              A growing number of companies have already developed industrial metaverses. Here are a few examples across different industries and use cases.

              Kawasaki: Robot care and maintenance

              Kawasaki unveiled the “Kawasaki DX.” It uses Microsoft’s industrial metaverse technology to implement a real-time virtual approach to robot care and maintenance. It enables remote robot operation by using a digital twin with Microsoft Azure cloud computing platform, and Microsoft’s HoloLens mixed reality.

              Bosch: Integrated asset performance management

              Bosch has developed a solution for asset performance management with digital twin on Azure which empowers rotating machines.

              This platform enables propellers, turbines, and electric motors to indicate when they need maintenance to run with optimal costs and maximum efficiency.

              Novo Nordisk: Efficiency and quality assurance

              Denmark-based Novo Nordisk, a leader in the global production of medicines, is using HoloLens 2 and Dynamics 365, among others, to increase production process efficiency while maintaining the highest quality standards and acting in accordance with required requirements and regulations.

              Here are a few other industrial metaverse use cases

              • Using AR/VR at workstation to effectively monitor and repair faults using better insights
              • Creating and deploying simulations suggests improvements and QA testing of workflow before physical deployments
              • Fast and more effective skill training. For example, in aviation, by simulating practice and using scenario-based exercises
              • Creating an interactive 3D space for customer engagement in retail stores

               

              Conclusion

              Use cases help in understanding the metaverse. Along with this, the available technologies help in visualizing the opportunities in the improvement of existing solutions or attempting a new venture. Industries are already testing 3D visualization of physical entities and real-life data virtual representation.

              With more sophisticated hardware, the Metaverse will open new significant improvement opportunities such as testing and building products virtually before doing it in real life, it can save costs and time, test products with different conditions, and provide risk-free environment for humans.

              Fractal offers end-to-end technological capabilities required for planning, building, and operating an effective industrial metaverse. Fractal has provided solutions to integrate IoT devices with the Azure cloud.

              We can efficiently process and store data for effective monitoring, analysis, and visualization of the result to enhance the manufacturing value chain and effective management of resources. Feel free to contact us, if you are interested in learning more about this topic.

              Databricks Spark jobs optimization techniques: Pandas UDF

              Pandas UDF was introduced in Spark 2.3 and continues to be a useful technique for optimizing Spark jobs in Databricks. The idea of Pandas UDF is to narrow the gap between processing big data using Spark and developing in Python. When Spark engineers develop in Databricks, they use Spark DataFrame API to process or transform big data which are native Spark functions.

              However, Python has become the default language for data scientists to build ML models, where a huge number of toolkits and libraries can be very useful. For example, while developing ML models, developers may depend on certain libraries available in Python that are not supported by Spark natively (like the basic Scikit learn library, which cannot be applied to Spark DataFrame). However, if developers develop in pure Python on Databricks, they barely take advantage of features (especially parallel processing for big data) from Spark.

              In that case, Pandas UDF is there to apply Python functions directly on Spark DataFrame which allows engineers or scientists to develop in pure Python and still take advantage of Spark’s parallel processing features at the same time.

              UDF vs Pandas UDF

              UDF is an abbreviation of “user defined function” in Spark. Generally, all Spark-native functions applied on Spark DataFrame are vectorized, which takes advantage of Spark’s parallel processing. Although Spark already supports plenty of mainstream functions that cover most of the use cases, we might still want to build customized functions to transform data for migration existing scripts or for developers who are not familiar with Spark.

              For example, let’s say we need a function to hash columns. Spark supports sha or md5 function natively, but UDF allows us to reuse the same hash and salt method on multiple columns. In addition, UDF allows the user to develop more complicated hash functions in pure Python or reuse the same function they have already developed. By converting UDF in Pandas UDF, the Pandas UDF will also process the column parallelly, which provides better performance than a UDF.

              Native Spark Function

              Databricks-Spark-tutorial-3-Native-Spark-Function

              Spark Native Function: 

              • 11.11 seconds 
              • Always the fastest if functions are supported 

               

              UDF

              Databricks-Spark-tutorial-3-UDF

              UDF: 

              • 31.84 seconds 
              • Easy to migrate. Much slower. 

               

              Pandas UDF

              Databricks-Spark-tutorial-3-PandasUDF

              Pandas UDF: 

              • 24.39 seconds 
              • Faster than UDF 

               

              Spark native functions will always have the best performance overall. However, when we have to develop some transformation function that is not supported by Spark, or it’s easier for developers to develop in pure Python, using Pandas UDF can optimize Spark jobs performance.

              Grouped Map UDFs 

              Another useful feature of Pandas UDF is the grouped map. The grouped map feature will split a Spark DataFrame into groups based on the groupby condition, and applies user-defined function to each group, which could transform each group of data parallelly like a native Spark function. 

              One useful scenario for grouped map is to train multiple models based on groups when we have a training function. In pure Python, without additional parallel or groupby settings, developers will prepare a training dataset and a testing dataset for each group, then train the model one by one. By using Grouped Map UDFs, developers can apply the function on each group simultaneously, which works like parallel processing. 

              Sequential train

              Databricks-Spark-tutorial-3-Sequential-train

              Sequential train: 

              • 27.4 minutes 
              • Apply function on each group sequentially 

               

              Spark Grouped Map Pandas UDF

              Databricks-Spark-tutorial-3-Spark-Grouped-Map-PandasUDF

              Spark Grouped Map Pandas UDF: 

              • 3.84 minutes 
              • Apply Pandas UDF on each group simultaneously 

               

              There are 8 Spark executors in the cluster. After applying Pandas UDF, the performance is almost optimized 8x, which means the 8 groups are trained at the same time. The largest benefit for Grouped Map Pandas UDF is that it can be easily converted from a normal Python function. In addition, it can be applied directly to Spark DataFrame without converting into Pandas DataFrame. 

              Additional: Koalas 

              In addition to Pandas UDF, Spark org released a new package called Koalas which is also targeted to optimize Python in Spark environments. Besides using Spark DataFrame API, users can also develop functions in pure Python using Pandas API but also take advantage of Spark parallel processing. 

              To put it in context of Pandas UDF: Koalas can apply functions on Pandas DataFrame while Pandas UDF applies functions on Spark DataFrame. 

              In summary, we have three options 

                1. Spark DataFrame API 
                2. Pandas UDF on Spark DataFrame 
                3. Koalas API (currently Spark Pandas API) on Pandas DataFrame 

              All three will take advantage of parallel processing. 

              Looking for more Databricks Spark job optimization tutorials? 

              Check out some of the other techniques we’ve covered below: 

              Programmer coding on multiple screens

              Spark is known for its parallel processing, which means a data frame or a resilient distributed dataset (RDD) is being distributed across the worker nodes to gain maximum performance while processing. However, one problem we could face while running Spark jobs in Databricks is this: How do we process multiple data frames or notebooks at the same time (multi-threading)?

              The benefits of parallel running are obvious: We can run the end-to-end pipeline faster, reduce the code deployed, and maximize cluster utilization to save costs. Let’s see what this looks like with an example comparing sequential loading and multi-threading.

              Sequential loading vs. Multi-threading

              The following scenario shows an example when we have multiple sources to read from, coalesce into one parquet file, and then write in the destination location for each part. In this scenario, coalescing into one partition can only work on one CPU core in Spark, so all the other cores will become idle. By using a multi-threading pool, each CPU will have jobs to work on, which not only saves time but also creates a better load balance.

              Our test cluster has one 4 cores/8 GB master node with two 4 cores/8 GB worker nodes.

              Sequential loading

              Sequential-loading-Databricks-tutorial

              Without multi-threading, under the sequential method, we read each part from the source, filter the data frame and write the result as one parquet file in the destination, which took about 20 seconds to load 8 tables.

              Multi-threading

              Multi-threading-pool-Databricks-tutorial

              Under the same functions, after applying ThreadPool (8 threads at the same time), 8 tables can be loaded within 5 seconds which is 4x faster than the sequential loading method.

              Conclusion 

              Multi-threading is relatively quick to set up compared with other optimization methods. The improvement could be unlimited if we have a large enough cluster and plenty of jobs to run parallelly (under suitable scenarios). We can also use the multi-threading pool to parallel run multiple notebooks which do not have dependencies on each other even if we do not have the same scenario as shown above.

              The purpose of using multi-threading is not only to save time, but also to fully utilize the clusters’ compute power to save cost by finishing the same amount of jobs within less time, or within the same amount of time on a smaller cluster, which gives us more options to manage the end-to-end pipeline.

              Possible scenarios for the multi-threading technique

              • Optimize bottlenecks in a pipeline to save end-to-end running time
              • Parallel run independent notebooks to optimize load balance, saving both time and cost
              • Read/write data from/to multiple tables

              Extras

              A multi-threading pool can also be developed by the “concurrent.futures.ThreadPoolExecutor” library in Python or the “scala.concurrent.ExecutionContext” library in Scala.

              Want to learn more about Databricks Spark job optimization? Check out our previous blog on the topic to learn about the shuffle partition technique.

              Programmer writing a code on his desktop

              Generally speaking, partitions are subsets of a file in memory or storage. However, Spark partitions have more usages than a subset compared to the SQL database or HIVE system. Spark will use the partitions to parallel run the jobs to gain maximum performance. While we operate Spark DataFrame, there are majorly three places Spark uses partitions which are input, output, and shuffle.

              Input and output partitions could be easier to control by setting the maxPartitionBytes, coalesce to shrink, repartition to increasing partitions, or even set maxRecordsPerFile, but shuffle partition whose default number is 200 does not fit the usage scenarios most of the time. This blog will introduce general ideas about how to set up the right shuffle partition number and the impact of shuffle partitions on Spark jobs.

              Key points for optimizing performance with the shuffle partition technique

              1. Each partition size should be smaller than 200 MB to gain optimized performance.
              2. Usually, the number of partitions should be 1x to 4x of the number of cores you have to gain optimized performance (which means creating a cluster that matches your data scale is also important).

                Best practices for common scenarios

                • The limited size of cluster working with small DataFrame: set the number of shuffle partitions to 1x or 2x the number of cores you have. (each partition should be less than 200 MB to gain better performance)

                e.g. input size: 2 GB with 20 cores, set shuffle partitions to 20 or 40

                  • The limited size of clusters, but working with huge DataFrame: set the number of shuffle partitions to Input Data Size / Partition Size (<= 200mb per partition), even better to be the multiple of the number of cores you have

                  e.g. input size: 20 GB with 40 cores, set shuffle partitions to 120 or 160 (3x to 4x of the cores & make each partition less than 200 MB)

                    • Powerful clusters which have more number of cores than the number calculated above: set the number of shuffle partitions to 1x or 2x the number of cores

                    e.g. input size: 80 GB with 400 cores, set shuffle partitions to 400 or 800.

                    Here is an example of how to improve the performance by simply changing the number of partitions on a small DataFrame working with a limited size of cluster (8 cores total).

                    Default 200 shuffle partitions

                    default-200-shuffle-partitionsdetails-for-default-results

                    200 is way too much for this size of data and size of a cluster. It takes longer to allocate the jobs to finish all 200 jobs. 

                    8 shuffle partitions to match the number of cores

                    8-shuffle-partitions-to-match-number-of-coresdetails-for-shuffle-partition-results

                    By simply changing the # of shuffle partitions without changing anything else, the process is running about 40% faster than the default. 

                    Conclusion

                    The first and most important thing you need to check while optimizing Spark jobs is to set up the correct number of shuffle partitions. The number of shuffle partitions will not only solve most of the problems but also it is the fastest way to optimize your pipeline without changing any logic. 

                    Note: 

                    The example was using a small DataFrame with a limited cluster, which does not need to consider the size of each partition and has no skew keys. While optimizing a larger DataFrame, the solution will also include checking the size for each partition and making sure each partition is well distributed. 

                    The ideal size of each partition is around 100-200 MB. The smaller size of partitions will increase the parallel running jobs, which can improve performance, but too small of a partition will cause overhead and increase the GC time. Larger partitions will decrease the number of jobs running parallel and leave some cores ideal by having no jobs to do. If you also have a skew key issue, try to add a dummy column and force Spark to partition on the well-distributed dummy column while partitioning then drop the dummy column while writing. 

                    Check out this latest tutorial: Databricks Spark jobs optimization techniques: Multi-threading. 

                    Top 7 AI predictions for 2024

                    AI is transforming the world in ways we never imagined. From creating art and entertainment to enhancing productivity and security, AI is reshaping every aspect of our lives and society.

                    In this blog post, we will highlight seven predictions for AI in 2024, based on the insightful LinkedIn post by our Co-founder, Group Chief Executive & Vice-Chairman, Srikanth Velamakanni.

                    2023 was the year when the world woke up to the potential of AI, 2024 will be a year when we see a very tangible role of AI in making our lives better,” Srikanth said.

                    Now, let’s jump into some of the most exciting and impactful developments in AI and how they will affect us soon.

                    1. The foundation models race will heat up: Gemini will launch before February, followed quickly by GPT-5 by March.
                    2. Foundation models vs domain-specific models: Foundation models will continue to challenge domain-specific ones, yet hybrid approaches will remain valuable.
                    3. Productivity boost across sectors: AI assistants, apps, and agents will enhance productivity in repetitive, cognitive, and creative tasks.
                    4. Enterprise adoption of Gen AI: Over 80% of Fortune 100 companies will implement Gen AI infused use cases, as risks related to privacy and compliance get mitigated.
                    5. Enhanced enterprise and website search: Powered by Gen AI, enterprise search and website search experience will improve substantially.
                    6. AI media creation: Text-to-video, image upscaling, and image-to-video technologies will become mainstream, revolutionizing content creation.
                    7. AI assistants in coding: Near-universal adoption of these tools will boost productivity by approximately 30% in coding tasks.

                    If you want to know more about the other predictions, click here, where you can find the original post and join the discussion.

                    Top 7 AI predictions for 2024

                    AI is transforming the world in ways we never imagined. From creating art and entertainment to enhancing productivity and security, AI is reshaping every aspect of our lives and society.

                    In this blog post, we will highlight seven predictions for AI in 2024, based on the insightful LinkedIn post by our Co-founder, Group Chief Executive & Vice-Chairman, Srikanth Velamakanni.

                    2023 was the year when the world woke up to the potential of AI, 2024 will be a year when we see a very tangible role of AI in making our lives better,” Srikanth said.

                    Now, let’s jump into some of the most exciting and impactful developments in AI and how they will affect us soon.

                    1. The foundation models race will heat up: Gemini will launch before February, followed quickly by GPT-5 by March.
                    2. Foundation models vs domain-specific models: Foundation models will continue to challenge domain-specific ones, yet hybrid approaches will remain valuable.
                    3. Productivity boost across sectors: AI assistants, apps, and agents will enhance productivity in repetitive, cognitive, and creative tasks.
                    4. Enterprise adoption of Gen AI: Over 80% of Fortune 100 companies will implement Gen AI infused use cases, as risks related to privacy and compliance get mitigated.
                    5. Enhanced enterprise and website search: Powered by Gen AI, enterprise search and website search experience will improve substantially.
                    6. AI media creation: Text-to-video, image upscaling, and image-to-video technologies will become mainstream, revolutionizing content creation.
                    7. AI assistants in coding: Near-universal adoption of these tools will boost productivity by approximately 30% in coding tasks.

                    If you want to know more about the other predictions, click here, where you can find the original post and join the discussion.

                    Top 7 AI predictions for 2024

                    AI is transforming the world in ways we never imagined. From creating art and entertainment to enhancing productivity and security, AI is reshaping every aspect of our lives and society.

                    In this blog post, we will highlight seven predictions for AI in 2024, based on the insightful LinkedIn post by our Co-founder, Group Chief Executive & Vice-Chairman, Srikanth Velamakanni.

                    2023 was the year when the world woke up to the potential of AI, 2024 will be a year when we see a very tangible role of AI in making our lives better,” Srikanth said.

                    Now, let’s jump into some of the most exciting and impactful developments in AI and how they will affect us soon.

                    1. The foundation models race will heat up: Gemini will launch before February, followed quickly by GPT-5 by March.
                    2. Foundation models vs domain-specific models: Foundation models will continue to challenge domain-specific ones, yet hybrid approaches will remain valuable.
                    3. Productivity boost across sectors: AI assistants, apps, and agents will enhance productivity in repetitive, cognitive, and creative tasks.
                    4. Enterprise adoption of Gen AI: Over 80% of Fortune 100 companies will implement Gen AI infused use cases, as risks related to privacy and compliance get mitigated.
                    5. Enhanced enterprise and website search: Powered by Gen AI, enterprise search and website search experience will improve substantially.
                    6. AI media creation: Text-to-video, image upscaling, and image-to-video technologies will become mainstream, revolutionizing content creation.
                    7. AI assistants in coding: Near-universal adoption of these tools will boost productivity by approximately 30% in coding tasks.

                    If you want to know more about the other predictions, click here, where you can find the original post and join the discussion.

                    4 steps to improve the data quality of your organization

                    The world today is swimming with data, and organizations need to handle a large amount of it to derive valuable insights for decision-making. But data is only helpful if it is of good quality. According to SAP, bad data costs the US $3 trillion annually. These costs include employee’s time and effort to correct bad data and errors.

                    Why is data quality important?

                    Data quality is a base for data analytics and data science. It measures how well-suited your data is to accomplish a particular task accurately and consistently. Good data helps an organization make key spending decisions, improve operations, and develop growth tactics. Even though technologies like AI and machine learning have enormous potential to handle large volumes of data, they need good quality data to produce reliable results quickly.

                    As data is an integral part of an organization, data quality impacts many aspects, from marketing to sales to content creation.

                    Good quality data helps you make informed decisions, target the right audience, drive effective marketing campaigns, strengthen customer relationships, gain competitive advantage, and so on.

                    Benefits of improving data quality

                    In this competitive era, organizations try to understand their customers better and make better financial, marketing, and development decisions based on accurate data for a better ROI. Bad data is unstructured data that may show quality issues like inconsistency, inaccuracy, insufficiency, or even duplicate information. It could be misleading and even more harmful for a business than a lack of data.

                    Improving the data quality of your company can result in the following advantages:

                    Benefits of improving data quality

                    • Data-driven decision-making: Decision-making is based on solid reasoning, and the correct data can only help make wise business decisions and provide the best outcomes.
                    • Customer intimacy: Drive marketing and customer experience by analyzing entire consumer views of transactions, sentiments, and interactions by using data from the system of record.
                    • Innovation leadership: Learn more about your products, services, usage trends, industry trends, and competition outcomes to help you make better decisions about new products, services, and pricing. ​
                    • Operational excellence: Make sure the correct solution is provided quickly and dependably to the right people at a fair price.

                          Challenges faced while maintaining data quality

                          Poor data quality can lead to financial losses, increased complexity, and limited usefulness of real-world evidence. Understanding data quality challenges is crucial for informed decision-making, reducing business risks, and achieving data-driven goals. Explaining a few challenges below.

                          • Data debt reduces ROI: Data debt refers to the negative consequences of accumulating poorly managed data. This includes inaccurate, inconsistent, incomplete, or inaccessible data. If your organization has a lot of unresolved data-related problems, such as issues with data quality, categorization, and security, it can hinder you from achieving the desired Return on Investment (ROI).
                          • Lack of trust leads to lack of usage: A lack of data confidence in your organization leads to a lack of data consumption, which has a detrimental impact on strategic planning, KPIs, and business outcomes.
                          • Strategic assets become liabilities: Poor data puts your company in danger of failing to meet compliance standards, resulting in millions of dollars in fines.
                          • Increased expenses and inefficiency: Time spent correcting inaccurate data equals less workload capacity for essential efforts and an inability to make data-driven decisions.
                          • Adoption of data-driven technologies: Predictive analytics and artificial intelligence, for example, rely on high-quality data. Delays or a lack of ROI will come from inaccurate, incomplete, or irrelevant data.
                          • Customer experience: Using bad data to run your business can hinder your ability to deliver to your customers, increasing their frustration and reducing your capacity to retain them.

                          Improving data quality with Fractal’s methodology 

                          Maintaining high levels of data quality allows organizations to lower the expense of finding and resolving incorrect data in their systems. Companies can also avoid operational errors and business process failures, raising operating costs and diminishing revenues. 

                          Good data quality enhances the accuracy of analytics applications, leading to improved business decisions that increase sales, improve internal processes, and provide firms with a competitive advantage over competitors. High-quality data can also help increase the use of BI dashboards and analytics tools. If analytics data is perceived as trustworthy, business users are more inclined to depend on it instead of making judgments based on gut feelings or their spreadsheets. 

                          Fractal has developed a process that significantly improves data quality across enterprises. Here’s how we do it. 

                          Data Preparation and Rule Calculations 

                          Fractal helps handle large volumes of data, preparing it for further processing. It also performs calculations based on data rules, identifying defective records in the data. 

                          Data extraction and preparation 

                          Fractal leverages a back-end engine for data extraction, data preparation, and data rules configuration on various data sets. 

                          Optimized process 

                          The focus is an optimized process with minimal processing time, parallel processing, and data aggregations to reduce the storage space and provide the user’s best dashboard performance. 

                          Data quality improvement 

                          Fractal helps transform the data cleansing operation with faster reduction of data defects, improved data quality, and tracking key KPIs like asset score, coverage, and conformance. 

                          How can Fractal help maintain data quality with Google Cloud?

                          Fractal leverages all the services provided by Google Cloud and supports integrations with the Google Cloud Platform. It also determines which Google Cloud services best meet Fractal’s data quality needs for each project.

                          Here are some ways Google Cloud can help maintain data quality.

                          • Data Governance: It helps automatically detect and protect sensitive information in data, ensuring data privacy and compliance. It also helps enable granular control over data access, preventing unauthorized modifications or deletions.
                          • Data Profiling & Cleansing: It offers a user-friendly interface for data cleaning and transformation, including outlier detection, data standardization, and data validation. It also provides AI-powered tools for data profiling and anomaly detection, helping identify data quality issues proactively.
                          • Data Monitoring & Improvement: It offers comprehensive dashboards and alerts for monitoring data quality metrics, allowing for proactive identification and resolution of issues. It also helps run large-scale data quality checks and analysis, providing deeper insights into data quality trends.
                          • Machine Learning & AI Integration: It provides tools for developing and deploying custom AI models for data quality tasks, such as entity extraction, classification, and matching. It also helps build serverless data quality functions that can be triggered on specific data events, ensuring real-time data quality checks.

                          Conclusion

                          In today’s data-driven world, maintaining high-quality data is no longer an option but a necessity. Poor data quality can lead to financial losses, operational inefficiencies, and inaccurate decision-making. By leveraging Fractal’s data quality methodology and Google Cloud’s powerful tools and services, organizations can effectively address data quality challenges and unlock their full potential.

                          Fractal empowers organizations to achieve data quality excellence. When combined with Google Cloud’s data quality capabilities, Fractal delivers a comprehensive solution for managing and improving data quality throughout the entire data lifecycle.

                          Are you seeking help to improve the data quality of your organization? Contact us to get started!

                           

                          Top 7 Announcements from AWS re:Invent 2023

                          Amazon Web Services recently concluded its highly anticipated re:Invent 2023 event, showcasing a resurgence of big community events in the tech industry after the pandemic-related hiatus. With a record-breaking attendance of around 50,000 participants, re:Invent 2023 was a milestone event with significant announcements and insights for the tech world. 

                          Here’s a summary of top announcements, followed by the keynote from Adam Selipsky, AWS CEO, and in-depth announcements.  

                          Generative AI was, unsurprisingly, the buzzword. AWS rolled out an exciting new AI chip, support for new foundation models, updates to its generative AI platform Amazon Bedrock, and support for vector databases and zero-ETL integrations. They also announced a new generative AI assistant dubbed Amazon Q (in reference to Star Trek’s “Q” character, seemingly). Q will probably have a broad set of use cases in enterprises. 

                          1. Amazon Q: A new generative AI assistant

                          A Gen AI-powered assistant designed to help get relevant answers, solve problems, and generate content using Retrieval Augmented Generation (RAG). It also integrates with other AWS services:  

                          • AWS Console: Amazon Q simplifies exploring AWS’s framework, best practices, and documentation. It is accessible in the AWS management console. For example, “How to build a web application on AWS?” yields a list of services like AWS Amplify, AWS Lambda, and Amazon EC2, along with their advantages and resources. 
                          • Connect: Cloud-based contact center service ensures scalable customer service operations. Amazon Q enhances customer interactions by understanding needs, minimizing wait times, and delivering real-time responses via API integration. 
                          • Amazon QuickSight: Business intelligence service offering interactive dashboards, paginated reports, and embedded analytics. Amazon Q within QuickSight enhances productivity for business analysts and users by swiftly creating visuals, summarizing insights, answering data queries, and constructing data stories using natural language.  

                          2. Expanded choice of models in Amazon Bedrock 

                            Bedrock is the foundation model library from AWS. It enables the integration of various models that underpin generative AI. AWS have introduced some new models to Amazon Bedrock: 

                            • Meta’s popular LLaMA-2  
                            • Anthropic Claude 2.1: An update to Claude 2, it now offers a 200k token context window (vs. 128k for GPT 4 Turbo), reduced rates of hallucination, and improved accuracy over long documents. 
                            • Amazon Titan Image Generator: Similarly to tools such as Mid-journey, Stable Diffusion, and Dall-E, Titan Image Generator lets you not only create but also improve images with natural language commands. Titan also supports enterprise needs for invisible watermarks on images. 
                            • Amazon Titan Multimodal Embeddings: Improve searches by understanding images and text. For instance, a stock photo company could use it to find specific images based on descriptions or other images, enhancing accuracy and speed. 

                            3. Four New Capabilities for AWS Supply Chain  

                            An application that unifies data and provides ML-powered actionable insights. It incorporates embedded contextual collaboration and demand planning features while seamlessly integrating with your client’s current enterprise resource planning (ERP) and supply chain management systems. Announced three new capabilities: 

                            • Supply planning (Preview): Plans purchases of raw materials, components, and finished goods. This capability considers economic factors, such as holding and liquidation costs.  
                            • Visibility and sustainability (Preview): Extends visibility and insights beyond your client’s organization to your external trading partners. This visibility lets you align and confirm orders with suppliers, improving the accuracy of planning and execution processes.  
                            • Amazon Q (Preview): As mentioned above, Amazon Q is now integrated with Supply Chain services to empower inventory managers and planners with intelligent insights and explore what-if scenarios. 

                            4. SageMaker capabilities (Preview) 

                              • HyperPod: accelerates model training by up to 40%, enabling parallel processing for improved performance.  
                              • Inference: reduces deployment costs and latency by deploying multiple models to the same AWS instance.  
                              • Clarify: supports responsible AI use by evaluating and comparing models based on chosen parameters.  
                              • Canvas: enhancements facilitate seamless integration of generative AI into workflows. 

                              5. Next-generation AWS-designed chips (Preview) 

                                Amazon jumped into the custom cloud-optimized chip fray and introduced two new chips: 

                                • Graviton4: A powerful and energy-efficient chip suitable for various cloud tasks.  
                                • Trainium2: A high-performance computing chip, it helps accelerate model training while making it more cost-effective. 

                                Notable AWS customers like Anthropic, Databricks, Datadog, Epic, Honeycomb, and SAP already use these chips. 

                                6. Amazon S3 Express One Zone  

                                A new purpose-built storage class for running applications that require extremely fast data access. 

                                7. New integrations for a zero-ETL future  

                                Aim to make data access and analysis faster and easier across data stores. Four new integrations are: 

                                To learn more about all those exciting news, please check the AWS Re:Invent keynote blog.  

                                Machine vision cameras

                                It is an exciting time in AI (Artificial Intelligence). ChatGPT is revolutionizing the world, and one of the best-known secrets of Microsoft’s platform is set to revolutionize the world of vision AI processing, through a deep technical collaboration with NVIDIA & Microsoft.

                                Machine vision cameras

                                Today, machine vision cameras are commonly deployed worldwide for visual inspections. These cameras enable people to see what they cannot see with the naked eye and have many use cases. They are frequently used in manufacturing settings for quality control on production lines and in detail-required industrial scenarios. Basler & Allied vision are leading producers of these cameras, which operate against the Gig-E vision protocol. NVIDIA announced compatibility and substantial investment in this protocol at GTC in March. Today, these cameras can be used to identify quality control issues and stitched together into any solution supporting and running with DeepStream.

                                Machine vision cameras

                                The importance of hardware acceleration

                                However, these cameras come with extremely specific requirements. Hardware that can process multiple streams of high-fidelity video coming from high frame rate cameras must be able to provide the right form factor, operating environment constraints, and processing requirements. Fortunately, NVIDIA’s GPUs (graphics processing units) can be leveraged for parallel processing at scale. This idea of massively parallelized hardware acceleration is in testing today for deployment on Azure Stack Edge and other Azure Certified Devices and will enable at-scale GPU deployment for hardware-accelerated machine vision running at the edge and connected to the cloud.

                                Edge and cloud

                                Finally, Azure’s Cognitive Services team’s unique and innovative Florence models (in preview today) enable customers to rapidly create and innovate upon cloud models – to accelerate edge-to-cloud hybrid AI processing workloads development. Azure’s Spatial Analysis cognitive service already supports edge deployments of certain industrial use cases on Azure Stack Edge. Leveraging these tools also allows for both edge deployments very rapidly and for hybrid deployments – leveraging custom models at the edge to identify and pass relevant frames to the cloud for deeper analysis.

                                Introducing Vision as a Service

                                Fractal has created Vison Analytics as a Service, an offering based on the repeated implementations of cloud-connected edge intelligence deployed by their Edge & IOT team. This stitches together the innovations offered by Microsoft and NVIDIA through a Managed Application, available to customers for a simple price. This simple managed offering allows for unified and integrated pricing as per the scale factors & standard requirements of a complex vision implementation, with an initial implementation of vision AI deployed in a 3-month timeline. It leverages standard templates and accelerators from Microsoft and NVIDIA, like the Azure DeepStream Accelerator, the NVIDIA GPU sharing work, MEC app solution accelerator, and the innovations mentioned above to enable a rapid deployment package for customers wanting to implement video AI solutions at scale.

                                This offering also provides several optional industrial extensions, including camera management, coordination with sophisticated AI engines, and integration with critical industrial equipment assets. One common use case is the deployment of deep reinforcement learning to fully automate command and control by leveraging a physics-based simulation on a manufacturing line, to drive activity after visual inspection. It can also be integrated with PLC control systems and other required industrial assets, as a direct extension, making it an incredibly useful solution for organizations wishing to upgrade their quality control processes with vision AI.

                                The goal with this offering is to enable as many customers as possible to unlock the value of vision AI by leveraging Fractals capabilities with the greatest ease possible.

                                We also offer a free consultation to scope a vision AI engagement and identify a hardware recommendation, edge & cloud strategy, audit camera requirements, scope the required skills, and any required DRL (Deep Reinforcement Learning) deployments. This standardized pricing is designed to offer greater scale, while Fractal leverages vision intellectual property development alongside Microsoft and NVIDIA’s offerings, enabled by this offer.

                                Transform Customer Digital Experience with AIDE

                                High digital abandonment rates are typical for brands across domains, driven mainly by the experiential issues site users face during their journey. Identifying these friction points can be burdensome for businesses as they need help digesting the new wealth of granular data generated along their customer’s digital journey.

                                Fractal’s Automated Insights for Digital Innovation (AIDE) is a smart digital solution for a broad range of industries, including retail, finance, insurance, and more, that uses a customizable open-source AI that works well with complex journeys, data security, and time-to-market needs for multiple industries. It helps make smarter decisions at every step and enables businesses to resolve issues quickly and increase potential revenue and leads.

                                AIDE helps:

                                • Analyze millions of digital consumer touchpoints to provide insights to increase revenue, engagement, and growth for the business.
                                • Identify the root cause of friction from call, chat, and website errors and use AI to parse out critical signals from all the unstructured data in the customer journey.
                                • Get the most comprehensive insights into the digital journey to drive data-driven hypotheses for supporting A/B tests to drive website design changes to improve the consumer experience.

                                 

                                What is AIDE?

                                AIDE is a digital optimization platform that helps detect and contextualize the issues faced by visitors on digital channels. It acts as an intelligent digital solution for various industries, including retail, finance and insurance, telecommunications, tech, media, and more. AIDE uses customizable, open-source AI that works well with complex journeys, data security, and time-to-market needs for multiple industries. It’s an insight-orchestrated platform supported by natural language-generated insights. AIDE:

                                • Selects the sales or service flow to influence particular focus points.
                                • Identifies the selected data domains to create a journey sensor.
                                • Helps detect the most important anomalies across key performance indicators.
                                • Finds the friction point on the website using various journey sensors.
                                • Helps analyze the customer voice to add context to insights.

                                Leveraging the power of the AWS cloud, AIDE is built on Amazon RDS, Redshift, EMR, LaMDA, E2, S3 and can be deployed in your AWS environment.

                                 

                                How can AIDE help my business?

                                AIDE product architecture

                                AIDE brings together data engineering, Natural Language Processing (NLP), machine learning, and UI capabilities to help clients:

                                • Easily develop new data features from raw data to power downstream data analytics use cases.
                                • Identify and locate precise points of friction on your company’s website, online events, or funnels.
                                • Deep dive into the context of customer dissonance using the voice of customer analytics.
                                • Prioritize the most critical areas based on value loss estimation.
                                • Analyze Omni-channel customer journey analysis.
                                • Provide user-friendly and intuitive UI for beginners and experts.
                                • Provide root cause analysis of customer pain points/dissonance during the digital journey.

                                 

                                What can I expect from AIDE?

                                With AIDE, you can capture every in-page interaction and micro-gesture to understand the site user’s journey and identify frustrations and errors impacting conversion and self-serve rate.

                                AIDE helps companies remove friction and errors that spoil the visitor’s experiences. It also helps leverage best-in-class AI/ML modules to identify struggle points and recommend changes to drive design improvements using multiple features such as:

                                • Sensorize: An automated AI/ML pipeline derives meaningful business indicators using the click activity across customer journeys.
                                • Detect: Deviations from expected behavior across digital journeys get captured by applying pattern recognition algorithms to the key digital indicators.
                                • Locate: A suite of supervised machine learning algorithms identify drivers of key customer journey outcomes (drop-off, clear cart, etc.) and measure relative impact at a page and click level of a customer’s experience.
                                • Reveal: NLP module performs sentiment analysis and entity extraction on the voice of customer data such as chat; feedback etc. to identify the root cause of the friction and generate actionable insights.
                                • Prioritize: Quantify the insights with respect to loss in revenue or incremental overhead costs to prioritize hypotheses for improving website design.

                                Overall, AIDE is adaptable and open source, making it a flexible solution for various needs and is effective at addressing the complex customer journeys of today’s digital world. It is secure and reliable, with a focus on data protection, and easy to use and deploy, with a quick time to market.

                                 

                                How do I get started?

                                AIDE is also available on AWS Marketplace. Contact us to learn how an AIDE solution can identify and reduce friction points, helping the business grow at scale.

                                Effective Enterprise Data Management using Federated DataOps

                                In the past, organizations have built monolithic analytics solutions with a high degree of centralization around a single data engineering team responsible for maintaining the platform and building data pipelines. This approach doesn’t work in large organizations as they operate in a decentralized way. That’s why new data management approaches such as data mesh have emerged to tackle conventional data management challenges such as data redundancy and dependency, privacy and security, limited data sharing and availability, and data inconsistency.

                                The underlying reasons for these challenges are that traditional strategies fail to address the following adequately:

                                Complexity: Data projects can be complex, involving multiple stakeholders, systems, and technologies. This complexity can make coordinating efforts difficult and ensure the project is delivered on time and within budget.

                                Lack of resources: Data projects often require specialized skills and expertise, which can be challenging to find and retain. The results can be delays and cost overruns as organizations need help to secure the necessary resources.

                                Data quality: Data experts are spread across the organization, departments, and business units. Centralized responsibility for managing data solutions leads to low quality. Good data quality can lead to correct insights and sound decision-making.

                                Lack of governance: With proper governance and controls, data projects can become more cohesive and organized, leading to delays and suboptimal outcomes.

                                Organizations must adopt modern data science solutions that can efficiently manage data’s ever-increasing analytics need, volume, and complexity.

                                Fractal recommends leveraging Federated DataOps with your organization’s data management strategies to provide accuracy, consistency, and faster time-to-value.

                                Federated DataOps

                                Data operations (DataOps) is evolving daily to meet automation needs for organizations ever-increasing data and analytics needs. DataOps aims to improve the quality, speed, and reliability of data processing and data-driven decision-making within an organization.

                                DataOps automates data flows between data managers and data consumers within an organization. By improving and automating these processes, DataOps can help organizations leverage their data assets faster and better and improve the quality of the insights. DataOps often uses a combination of tools and technologies, such as data integration platforms, data lakes, and data governance frameworks, to support the automation and integration of data flows across the organization.

                                Enterprise Federated DataOps architecture

                                Modern data management strategies like Data Mesh and Data Fabric architectures can use Federated DataOps Governance to help manage and automate the data flow within an organization. There are several reasons why organizations might choose to use Federated DataOps in conjunction with data fabric and data mesh architectures:

                                • Federated governance: Organizations are adopting self-serve distributed data platforms, which could lead to silos and duplicate DataOps practices by different teams. Federated DataOps ensures high standards, reusability, and governance across teams.
                                • Improved data quality: Continuous integration and delivery can enhance data quality by catching and fixing errors early in the data pipeline. It ensures data accuracy and reliability.
                                • Increased speed: Automating and streamlining data integration and management processes can help organizations get insights from their data faster and make quicker decisions based on those insights.
                                • Enhanced collaboration: It promotes collaboration and continuous improvement, fostering better communication and cooperation between data managers and data consumers within an organization. It leads to progress in the overall effectiveness of data analytics processes.
                                • Reduced complexity: It can help reduce the complexity of data management and analytics processes by automating and streamlining the data flow within an organization, enabling quick and informed decisions.

                                Get started with your data management journey with Fractal

                                Companies are looking for a faster and more efficient way of managing data and preserving and using their data as technology advances. A sound data strategy is key to data’s successful and practical usage to achieve business goals within the stipulated time. The process demands quickness, efficiency, and consistency; hence, DataOps is used with Data Mesh or Data Fabric to accelerate the implementation of data management strategies.

                                Fractal Analytics is a data & AI Microsoft Solutions Partner that catalyzes data-driven transformation initiatives for enterprises leveraging the Microsoft cloud. We assist enterprises in creating end-to-end data management solutions that encourage innovation and value generation in line with current and anticipated market trends and long-term corporate objectives.

                                Contact us to get started with a data management strategy using best DataOps practices.

                                Role of data & analytics in alleviating health inequities

                                The COVID-19 pandemic and the resulting global recession caused a significant increase in poverty around the world as many families lost their sources of income. The poorest families, especially those without access to social protection, were hit the hardest. As a result, the World Health Organization urged governments and health organizations to pay attention to the social determinants of health to take steps to optimize community well-being.

                                Social determinants of health (SDOH) are any of a variety of non-medical factors that influence health outcomes. They encompass the conditions in which people are born, grow, live, learn, work, play, worship, and age — shaping their overall health. These factors include attributes associated with one’s environment, patterns of social engagement, and one’s sense of security and well-being. Collectively, SDOH influences health, working life conditions, and quality of life.

                                In this article, we’ll discuss how social determinants of health impact overall wellness across populations, including disparities in healthcare, and the role data can play in alleviating these inequities and shaping healthcare outcomes.

                                Social determinant factors

                                Patients’ health depends not only on the care they receive in a provider’s office but also on social determinants of health such as individual lifestyle choices, living situation, and access to healthy food.

                                According to a study by the CDC, social determinants of health fall into five broad groups:

                                HIMSS Illustration

                                 

                                 

                                Key benefits of studying social determinants

                                Addressing social determinants is important for improving health and reducing health disparities. Understanding social factors gives valuable insight into possible negative health outcomes for patients in many domains. Below are the key benefits of studying social determinants of health.

                                Holistic healthcare: Incorporating social determinants into healthcare practices fosters a more holistic and comprehensive approach to patient care. For instance, factors such as a patient’s education, income level, and environment should be considered when providing treatment and care.

                                Addressing health inequities: Social determinants have an important influence on health inequities – the unfair and avoidable differences in health status seen within and between countries.

                                Resource development: Acknowledging social determinants can initiate the development of resources to solve inequality and strengthen the overall health of the community.

                                Influencing health outcomes: Research shows that social determinants can be more important than healthcare or lifestyle choices in influencing health outcomes.

                                The impact of social determinants of health

                                Social determinants of health have a significant impact on people’s health, well-being, and quality of life. A growing body of research indicates:

                                • Children born to parents who haven’t completed high school are more likely to live in environments that contain barriers to health.
                                • Poor individuals who are white are less likely to live in areas of concentrated poverty than poor racial and ethnic minorities.
                                • As income decreases, the likelihood of premature death increases.
                                • There is a direct link between the likelihood of smoking, shorter life expectancy, and lower income.
                                • The environment in which an individual lives may impact future generations.
                                • Stress related to disparities has a direct link to health and often results from overlapping factors.

                                Negative social determinants of health can lead to disparities in healthcare, which can be costly and inhibit the overall quality of care and population health. This can result in added healthcare expenses, loss of productivity, and premature death. According to the Kaiser Family Foundation, 30% of direct medical costs for black, Hispanic, and Asian Americans are unnecessary costs incurred due to inefficiencies, disparities, or inequities in the healthcare system. In addition, the US economy loses an estimated $309 billion annually due to the direct and indirect costs of disparities.

                                Role of data & analytics in alleviating healthcare inequities

                                Barriers like a lack of data standards and costly datasets can hinder an organization’s access to social determinants information. However, developing an approach for more holistic patient care will be necessary for organizations looking to improve patient and population health, whether the data is complete or not.

                                Data and analytics are vital in helping to end these disparities and ensuring that all populations have the same access to services and care, not only for COVID-19 but also for all diseases and disorders that threaten public health.

                                Through data analytics and population health management, providers can improve patient outcomes, enhance care management, and address social determinants of health. Nowadays, data analytics are helping providers replace the “one size fits all” care mentality to deliver value-based care. Providers can assess which processes are the most effective methods for wellness and prevention within value-based care models. With population health management, organizations can consider physical and social determinants of health that may impact individuals and focus on “well care” rather than waiting for a patient to become ill.

                                Building a better healthcare system with Fractal

                                Health disparities and inequities are shaped by a multitude of factors in an individual’s socio-economic and healthcare journey. The health outcomes of an individual are significantly influenced by social determinants of health of the community in which they reside.

                                Fractal’s RAISE, powered by AWS, is an AI-powered solution that helps organizations speed up population health and health equity journeys. RAISE combines a member’s community data with clinical and social needs, assisting organizations in crafting the right interventions. It also creates a multifaceted approach involving data, analytics, and AI to advance health equity. It helps identify the drivers of inequities and disparities that are specific to the members of the organization.

                                AWS equips Fractal’s AI/ML solutions with the scalability, reliability, and security needed to deliver its solutions to healthcare providers and patients worldwide. By leveraging AWS, Fractal is helping to make a real difference in the lives of people affected by health inequities by supporting community-based healthcare organizations, advocating for policies that address the social determinants of health, and promoting health education and literacy.

                                Fractal recently led an on-demand webinar in collaboration with the Healthcare Information and Management Systems Society (HIMSS) to discuss health inequities. It explained how Fractal has been solving various healthcare-related issues using advanced AI/ML solutions.

                                The webinar featured Fractal’s Chief Practice Officer, Matt Gennone, and Dr. David Nash, an advisor at Fractal. They discussed the economic impact of health inequities and delayed care and how healthcare can return to its “true north” of providing high-quality care.

                                To know more, watch this session.

                                Contact us to learn more.

                                  7 ways implementing MLOps can transform your business

                                  Machine learning operations (MLOps)

                                  As per Gartner, only 50% of machine learning models reach production, leading to an efficiency challenge for many organizations. To improve the rate of machine learning model deployment, many organizations have begun to adopt MLOps.

                                  MLOps is the concept of applying DevOps principles to ML (Machine Learning) workflows and helping increase the deployment rate of ML models to production by leveraging a set of practices, processes, and tools per the used case scenario.

                                  The MLOps process includes several stages: data preparation, model training, testing and validation, deployment, monitoring, and feedback loops. The main goal of MLOps is to increase the machine learning process’s efficiency, accuracy, and reliability while ensuring the models perform as expected in real-world environments.

                                  MLOps typically involves using various processes, tools, and technologies, such as version control systems, containerization platforms, and continuous integration and deployment pipelines, to automate and streamline the machine learning workflow. The tools and practices enabled by MLOps can help reduce the risk of errors, improve collaboration, and ensure that machine learning models are updated and optimized for better performance over time.

                                  Let us look into the 7 key benefits of implementing MLOps

                                  7 key benefits of implementing MLOps-illustration

                                  1. Increases productivity

                                  MLOps practitioners leverage various tools and practices designed to help streamline and automate machine learning development and deployment processes. It can include automating data preprocessing and feature engineering, managing model training and evaluation, and deploying and monitoring models in production.

                                  By implementing tools designed to automate and standardize development and deployment processes, organizations can reduce the time and effort required to develop and deploy machine learning models, allowing data scientists and engineers to focus on higher-level tasks.

                                  This results in the faster and more efficient delivery of high-quality machine learning models, ultimately driving business value and improving productivity.

                                  2. Faster deployment and easy monitoring

                                  MLOps methodologies can help organizations accelerate modeling processes. They can also help facilitate machine learning models’ seamless construction and deployment by helping leverage automated systems.

                                  Commonly used MLOps tools can help with automatic system monitoring systems, which can be used in the continuous monitoring of models in production, allowing for quick identification and resolution of any issues. These tools help organizations improve the speed and quality of their machine learning deployment, leading to increased productivity and better outcomes.

                                  3. Budget and cost management

                                  Implementing MLOps can help ensure efficient resource usage and cost control by using tooling designed to monitor usage patterns, identify bottlenecks, and scale resources based on demand. These tools can help estimate and track costs before, during, and after experimentation.

                                  4. Reproducibility and versioning

                                  Organizations can leverage MLOps policies to enable a structured approach for practitioners and data scientists to track and manage changes, enable versioning, and provide a history of edits and versions created.

                                  Versioning aids in deploying models in production and enhances the reliability and scalability of machine-learning models.

                                  5. Reliability

                                  Leveraging MLOps methods and tools can result in more reliable ML pipelines by minimizing the scope of human error and providing real-time data insights. MLOps can improve the dependability of machine learning models, ensuring their consistency and accuracy in production.

                                  By continuously monitoring model performance and dependencies, teams can promptly identify and address issues, increasing the models’ reliability.

                                  6. Collaboration

                                  MLOps best practices and policies are meant to break down silos between teams, allowing them to collaborate, share data more efficiently, and seamlessly integrate their workflows. This collaboration can lead to faster and more efficient model deployment and a more streamlined machine-learning process.

                                  With MLOps, organizations can ensure that their machine learning projects are connected and working together efficiently, leading to better outcomes and improved productivity.

                                  7. Monitorability

                                  Through MLOps, organizations can get insights into model performance and retrain the model continuously to ensure it gives the most accurate output. MLOps enables practitioners to do this by providing guidance on best practices for successfully implementing automated monitoring systems. These monitoring systems facilitate constant model monitoring and allow stakeholders to identify any issues or anomalies that may arise quickly.

                                  Identifying problems and irregularities can help to improve model performance and reduce downtime, leading to better outcomes and a more efficient deployment process.

                                  With MLOps, organizations can ensure that their machine learning models always perform optimally, improving productivity and better business results.

                                  MLOps with Fractal

                                  Fractal has a successful track record of delivering ML projects for clients leveraging our in-house MLOps methodology. Our secret lies in using a reliable methodology designed to remove uncertainty, foster consistency, and enable the quick realization of value, as well as continuous packaging, validation, and deployment of models to production—partnering with us for MLOps implementation grants you access to this proven methodology.

                                  Getting started with MLOps

                                  Organizations looking to implement MLOps can leverage the same proven methodology we use in-house to deliver projects for clients successfully.

                                  Contact us to get started.

                                  Scale AI model governance with the AI Factory framework

                                  AI and machine learning projects are on the rise. According to Gartner, 48% of CIOs and tech executives have deployed, or plan to deploy, an AI/ML project in 2022. It is also estimated that 50% of IT leaders will struggle to drive their AI initiatives from Proof of Concept (PoC) to production through 2023.

                                  Challenges moving AI/ML initiatives from PoC to production

                                  What causes the gap between PoC and AI/ML model implementation?

                                  IT and business leaders often cite challenges relating to security, privacy, integration, and data complexity as the key barriers to deploying AI/ML models in production. It is often due to governance frameworks not being shared across an organization to ensure compliance and maintainability – if a framework exists at all.

                                  “At some point, your proof-of-concept is likely to turn into an actual product, and then your governance efforts will be playing catch-up,” writes Mike Loukides in an O’Reilly report. “It is even more dangerous when you’re relying on AI applications in production. Without formalizing some kind of AI governance, you’re less likely to know when models are becoming stale, when results are biased, or when data has been collected improperly.”

                                  AI models require constant attention in production to achieve scalability, maintainability, and governance. To do that, organizations need a strong MLOps foundation.

                                  Leveraging MLOps at scale

                                  In one survey, Deloitte found that organizations that strongly followed an MLOps methodology were…

                                  • 3x more likely to achieve their goals
                                  • 4x more likely to feel prepared for AI-related risks
                                  • 3x more confident in their ability to ethically deploy AI initiatives

                                  AI Factory framework benefits

                                  Organizations following an MLOps methodology also gain a clear advantage in time to deployment. McKinsey found that companies without a formalized MLOps process often took 9 months to implement a model. In comparison, companies applying MLOps could deploy models in 2 to 12 weeks!

                                  The secret? By applying MLOps practices, these companies were able to create a “factory” approach for repeatable and scalable AI/ML model implementation. Their engineers weren’t building everything from scratch–they could pull from a library of reusable components, automate processes, and ensure compliance and governance throughout the organization.

                                  Luckily, you can also take this approach with our AI Factory Framework.

                                  Our AI Factory Framework

                                  The AI Factory Framework is a cloud-based MLOps framework that provides organizations with the foundation to deliver Data Science, Machine Learning, and AI projects at scale. It offers enterprise-level reusability, security, integration, and governance.

                                  Simply put, AI Factory helps customers scale MLOps, centralize governance, and accelerate time to deployment.

                                  Key benefits of the AI Factory

                                  By leveraging reusable and standardized artifacts, automated pipelines, and governance solutions, our AI Factory framework reduces duplicate effort and upskilling needs between teams and projects.

                                  AI Factory framework deliverablesAI Factory Framework benefits

                                  Customers leveraging the AI Factory Framework can take advantage of our AI engineering best practices to accelerate deployment and ensure model governance at scale.

                                  AI Factory also helps businesses:

                                  • Make the entire end-to-end lifecycle more repeatable, governable, safer, and faster
                                  • Shorten planning and development with accelerated time to deployment
                                  • Streamline operational, security, and governance processes
                                  • Reduce development risks & improve model quality
                                  • Reduce team’s upskilling needs
                                  • Achieve higher success rates & ROI

                                   

                                  Learn more

                                  Over the last decade, we have helped many customers build and execute their AI governance strategy. We distilled this experience and the derived best practices in this framework, to help deliver customers’ AI/ML initiatives at scale.

                                  Want to set up your own AI Factory Framework? Contact us to get in touch with one of our experts!

                                  Resources:

                                  Model development diagram

                                  Operationalizing and scaling machine learning to drive business value can be challenging. While many businesses have started diving into it, only 13% of data science projects actually make it to production. Moving from the academics of ML to real-world deployment is difficult, as the journey requires finetuning ML models to fit the practical needs of a business and ensuring the solution can be implemented at scale.

                                  Many organizations struggle with ML operationalization due to a lack of data science and machine learning capabilities, difficulty harnessing best practices, and insufficient collaboration between data science and IT operations.

                                  Common challenges with ML operationalization

                                  Many organizations get attracted to buzzwords like “machine learning” and “AI,” and spend their development budgets pursuing technologies rather than addressing a real problem. ML projects are an investment, and obstacles in operationalizing the solution make it even harder for the business to realize value from these solutions.

                                  Here are some common ML operationalization bottlenecks and the solutions to tackle them.

                                  • Lack of communication, collaboration, and coordination: Proper collaboration between the data scientist team and other teams, like business, engineering, and operations, is crucial. The ML project may not add real-world business value without proper alignment and feedback.
                                  • Lack of a framework or architecture: When ML models lack the proper framework and architecture to support model building, deployment, and monitoring, they fail.
                                  • Insufficient infrastructure: ML models use vast data to train the model. Most of the time is spent preparing data and dealing with quality issues without the proper infrastructure. Data security and governance are crucial factors that must be considered in the initial phase.
                                  • The trade-off between prediction accuracy and model interpretability: Complex models are generally harder to interpret but provide more accurate predictions. The business must decide what’s an acceptable tradeoff to get a “right-sized” solution.
                                  • Compliance, governance, and security: The data science team may not always consider other issues like legal, compliance, IT operations, and others that occur after the deployment of ML models. In production, setting up performance indicators and monitoring how the model can run smoothly is important. So, understanding how ML models run on production data is a crucial part of risk mitigation.

                                  Unfortunately, many ML projects fail at various stages without ever reaching production. However, with the correct approach and a mix of technical and business expertise, such as that provided by Fractal’s data science team, it is possible to avoid or quickly resolve many of these common pitfalls. Fractal can help organizations deploy more ML models to production and achieve a faster time to value for ML projects with the tools, practices, and processes of MLOps.

                                  Starting with the business objective

                                  Fractal’s proven MLOps methodology helps streamline and standardize each stage of the ML lifecycle from model development to operationalization. It allows collaboration between technical and non-technical users and empowers everyone to participate in the development process actively.

                                  We have helped many organizations leverage MLOps, allowing them to overcome their challenges. It includes a process for streamlining model training, packaging, validating, deployment, and monitoring to help ensure ML projects run consistently from end to end.

                                  Our successful 5-stage model

                                  Model development diagram

                                  1. Train: We create and train an initial model based on available data, business requirements, and desired outcomes.
                                  2. Package: Once the model is trained, we package up the model to make it easy to test, iterate, and deploy at scale.
                                  3. Validate: Later, we help validate models by measuring candidate models against predefined KPIs, deployment testing, and testing application integrations.
                                  4. Deploy: On validating, we deploy models by identifying the deployment target, planning the deployment, and then deploying the models to their production environment. We ensure that the services are implemented to support scalability, data pipelines are automated, and a model selection strategy is implemented.
                                  5. Monitor: Finally, we monitor models to track behavior, continuously validate KPIs, measure accuracy and response times, watch for drift in model performance, and more.

                                  Google Cloud services for ML model deployment

                                  We can help teams successfully deploy and integrate more ML models into production and achieve a faster time to value for ML projects with more efficient model training using Google Cloud services such as:

                                  • Google Cloud Storage: It enables organizations to store, access, and maintain data so that they do not need to own and operate their own data centers, moving expenses from a capital expenditure model to an operational expenditure.
                                  • Cloud Functions: It provides a simple way to run code responding to events with minimal configuration and maintenance. Cloud Functions are event-driven, meaning they can be triggered by changes in data, new messages, and user interactions.
                                  • Big Query: A fully managed enterprise data warehouse helps you manage and analyze your data with built-in features like machine learning, geospatial analysis, and business intelligence.
                                  • Kubernetes engine: A solution to help organizations achieve zero ops. Kubernetes is an open-source container orchestration system for automating software deployment, scaling, and management.
                                  • Data Proc: It is a fully managed and highly scalable service for running Apache Hadoop, Apache Spark, and 30+ open-source tools and frameworks.
                                  • Vertex AI: It is a machine learning platform that lets you train and deploy ML models and AI applications and customize large language models (LLMs) for use in your AI-powered applications frameworks.

                                  Leveraging both business and technical expertise

                                  MLDEVOps diagram

                                  Our ML model lifecycle unifies data collection, pre-processing, model training, evaluation, deployment, and retraining to a single process that teams maintain. It allows businesses to quickly tackle obstacles faced by the data scientists and IT operations team while providing a mix of technical and business expertise.

                                  How Fractal’s methodology benefits organizations

                                  • Eliminates guesswork
                                  • Supports consistency
                                  • Enables continuous packaging, validation, and deployment of models to production
                                  • Rapid time to value
                                  • Accelerate time-to-value and time-to-deployment
                                  • Efficiently manage data error and model performance
                                  • Increase model scalability during training and serving

                                  Conclusion

                                  As we continue through 2023, the MLOps market is surging rapidly. As ML applications become a key component for maintaining a competitive advantage, businesses realize they need a systematic and reproducible way to implement ML models. According to the analyst firm Cognilytica, MLOps is expected to be a $4 billion market by 2025. Fractal has deep expertise in MLOps and can help deliver solutions for unique business challenges across virtually all industries and sectors.

                                  Ready to begin leveraging MLOps in your organization? Contact us to get started.

                                  Benefits of DevOps

                                  Technology projects require maintenance throughout the entire lifecycle, from development to deployment and monitoring. Maintaining a project from version to version can become a manual and strenuous process. Developers must take special considerations at each stage to ensure smooth rollouts. Failure to do so can result in extended-release planning cycles ensuring the software is ready for use by end users.

                                  Development and IT Operations teams may end up spending unnecessary cycles supporting and fixing buggy software. And even worse, failed software releases can impact a company’s financial performance through operations inefficiencies, lost sales, and customer attrition. Failure to maintain working software can impact SLAs, regulatory compliance for some industries, resulting in fines or legal action.

                                  Successful organizations have adapted and created a set of best practices for governing projects, called DevOps.

                                  DevOps illustration

                                  What is DevOps?

                                  DevOps aims to create a common culture that brings together the people, processes, and technology to deliver value (i.e., working software) to end-users.

                                  It has also come up with procedures for automating many manual maintenance steps to reduce the time it takes to develop, test, and deploy software. Many companies are rushing to implement DevOps to avoid the high costs associated with manually maintaining projects.

                                  Benefits of DevOps on Google Cloud

                                  If you’re asking this question, keep reading. Outlined below are three key benefits of implementing DevOps in your organization.

                                  Benefits of DevOps

                                  1. Improved quality

                                  • Standardized tools and processes (i.e., Google Cloud DevOps and Agile Project Management) help keep quality consistent across projects and releases
                                  • Increased DevOps flow on Google Cloud helps improve software quality, delivery, and operations which leads to maintained security and faster deployment from the start
                                  • Quality control implemented through source control branching, code reviews, environment management, release management, etc.
                                  • Reduced fire drills and break-fix measures because of following DevOps best practices

                                  2. Reduced effort

                                  • Fewer manual processes and interventions through improved automation
                                  • Improved software stability leads to faster deployment with minimum manual intervention
                                  • Lower effort to support/maintain because solutions have gone through the appropriate governance and management processes
                                  • Leverage templates for common infrastructure builds/configurations to accelerate new projects going forward

                                  3. Increased collaboration

                                  • Agile project management structure encourages frequent collaboration among team
                                  • Google Cloud provides robust CI/CD capabilities, allowing teams to automate the entire software delivery pipeline. By integrating development, testing, and deployment processes, DevOps teams can collaborate more effectively and achieve faster and more reliable software releases
                                  • Improved communication channels enable the team to identify, track, and manage risks as they arise
                                  • Clear direction and prioritization through collaboration between stakeholders, developers, and end-users

                                  Hopefully, this helps you better understand some of the benefits that implementing DevOps using Google Cloud can bring to your business.

                                  Implementing DevOps is a journey and is not as easy as installing a package, flipping a switch, or buying new technology. Fractal specializes in helping our customers through the process of implementing DevOps, no matter what their current maturity level is. Fractal can provide strategic planning, technology assessments, proof of concepts, as well as technical resources to get you started on your DevOps journey.

                                  Interested in learning more about how Fractal can help you implement DevOps? Please contact us for additional information from our experts.

                                  Three pillars of the retail industry: Replenishment, allocation, and transportation 

                                  The retail industry is one of the most dynamic and fast-paced sectors, comprised of various stakeholders engaged in selling finished products to end-user consumers. In 2022, the U.S. retail sector was estimated at more than seven trillion dollars. The sector is projected to continue to grow, and by 2026 U.S. retail sales are expected to reach approximately USD 7.9 trillion. With the increased demand for consumer goods in different sectors and the ever-increasing choices of products at low costs, investments in the retail sector have also grown over the past few years.

                                  As there is always an element of change in this industry, new challenges are encountered daily. Take product stockouts, for example. Suppose a customer walks into a grocery store to purchase items of their favorite brand but discovers the product is unavailable. Frustrated by this, the customer chooses to either buy another brand or postpone the purchase; both scenarios are unfavorable to the business. The brand image and sales of the product are damaged because of this out-of-stock issue. The out-of-stock situation occurs when the inventory of a particular product is exhausted; this causes a problem for suppliers and retailers.

                                  Multiple reasons could cause the product stockout, such as inaccurate inventory data, lack of demand forecasting, or an unseasonal spike in purchasing. Many of these underlying causes of stockouts can be avoided if the business implements adequate processes to be carried out every month.

                                  To avoid situations like the stockout example above, retail companies need to develop a methodology for streamlining the following operations:

                                  3 pillars of retail industry

                                  1. Replenishment
                                  2. Allocation
                                  3. Transportation

                                  These three operations create the three pillars of the retail industry that help monitor real-time insight into customer behavior and understand their buying patterns, hence strengthening the retail foundation.

                                  1. Replenishment

                                  Replenishment refers to a situation where the amount of stock left in the store is counted so that the right products are available in an optimal quantity. It is considered an essential aspect of inventory management as it ensures that the right products are being reordered to meet the customer demand.

                                  In operational terms, the efficiency of store replenishment has a significant impact on profitability. The effectiveness and accuracy of store ordering affect sales through shelf availability and storage, handling, and wastage costs in stores and other parts of the supply chain. By optimizing demand forecasting, inventory management, and setting order cycles and order quantities by making them more systematic, the gains obtained are significant, often amounting to savings of several percent of total turnover.

                                  For companies that must manage many SKUs, one of the most effective ways of making store replenishment more accurate, efficient, and cost-effective is by using a replenishment system specifically tailored to business operations. When many different products need to be managed, manual ordering is highly labor-intensive and expensive; this results in companies using the replenishment system.

                                  An efficient replenishment system reduces process costs, improves inventory turnover, and provides higher service levels. The system constantly monitors the stock, sales, and demand while considering the forecast changes in demand and adjusting the replenishment orders. The company can control its inventory to ensure long-term profitability by recognizing the sales frequency, value, or profit margin. The replenishment system calculates the safety stock level for each SKU separately and sets them to meet the service level targets with efficiency, considering the predictability of demand.

                                  2. Allocation

                                  In allocation, the new stock of products is distributed to individual store units to maximize the product’s sales and prevent any stock-out situation in the future. This process enables assigning supplies to support the organization’s strategic goals. Having sufficient stock levels is an essential component for any retail business; with the changing consumer habits, it becomes crucial for the stock to be available in the right place at the right time.

                                  To meet new and increasing demands, retailers need an efficient process to gather, interpret, and analyze data from customer behaviors and habits, which would help get a more localized and specific idea of what is sold at a larger quantity in different locations. Items that are high sellers in one particular area may not sell well in others, so recognizing and monitoring this can ensure that the stock is allocated to the most needed location. Due to this, an opportunity is provided for the retailers to encourage sales by pushing stock of a similar type that a customer may favor at a particular location.

                                  3. Transportation

                                  Transportation plays a significant role in delivering the right stock of products at the right point of delivery. It connects the business to its supply chain partners and influences customer satisfaction with the organization. With the ever-changing customer preferences and as their expectations continue to evolve, the transportation industry is undergoing a dramatic transformation to meet these demands.

                                  Today, data plays a vital role in shaping the industry’s progress amidst tough competition. Due to the maturation of automation technologies, AI will help the transportation industry to manage drivers and fleet managers. By employing the techniques of AI, fleet and truck adjustments will offer data in real-time, eventually improving the industry’s standard. The safety and retention of the drivers will also increase from these newly acquired standards, and with enhanced access to data, there will be transparent communication between drivers and carriers.

                                  The average time between goods purchasing and delivery decreases by using real-time findings, making retailers focus on transportation to improve their business performance. The ability to automate insights, alerts, and data exchange more quickly will be the game-changer for this industry.

                                  These three pillars of retail can be strengthened by adopting in-house solutions and capabilities like StockView for retail, Edge video analytics, and Dynamics 365 customer insights.

                                  How could these solutions help the retail industry?

                                  StockView for retail helps retailers reduce lost sales and improve customer experience by automatically detecting out of stock items on shelves. It uses computer vision technology running at the edge to detect gaps on store shelves automatically. It also provides retailers with powerful insights and analytics into stock-out activities at both single-store and multi-store levels.

                                  Powered by Microsoft Azure Stack Edge, it offers a scalable, flexible, and cost-effective solution that brings the power of the Azure cloud platform down to the individual store, eliminating the need for costly and unreliable data transfers while offering a predictable and consistent TCO (Total Cost of Ownership).

                                  Edge Video Analytics solutions for retail typically leverage Azure Stack Edge devices, IoT devices, vision AI (Artificial Intelligence), and other technologies in conjunction with in-store cameras or other video sources.

                                  Insights gained from Edge Video Analytics can allow retailers to swiftly react to in-store customer behavior and product stock-outs while improving security and reducing shrinkage.

                                  Also, organizations can track store foot traffic, automatically notify employees to open more checkouts due to lengthy queues, and automatically detect product stock-outs. These insights can be used to help improve demand forecasting, optimize supply chains, and detect variances in population preferences down to the individual store.

                                  Dynamics 365 customer insights give you access to Microsoft’s extensive library of pre-built data connectors to help businesses gain faster time to value. Plus, you can power up your customer profiles with AI by leveraging Azure Machine Learning and Azure Cognitive Services, and Fractal’s library of custom AI models. So, we leverage this solution to maximize the potential of customer data.

                                  All these solutions and capabilities help understand the customer motivations, preferences, and desires to meet their demands and increase sales effectively, strengthening the retail industry’s pillars.

                                  Conclusion

                                  To meet these growing customer expectations, retailers should prioritize collecting customer data and analyzing it to support business decisions throughout their value chain. The inventory stocking patterns and shipping routes will shift in relation to patterns informed by this data. Retailers should make a concentrated effort to leverage the data while making critical business decisions, and to remain efficient; they must remain flexible and transform their operations as they capture more insights from their data.

                                  Over the past 20+ years, Fractal has helped many retail companies make their replenishment, allocation, and transportation operations more efficient by leveraging AI, engineering, and design. If you would like to learn more about optimizing these aspects of your retail business, please contact us to speak with one of our experts.

                                  Find how Fractal helped a global CPG company to operationalize an analytics engine designed and provide store recommendations for maximizing investments in their Azure environment. Read the full case study here: https://fractal.ai/casestudies/store-level-insights/

                                  GenAI for field technicians and engineers

                                  Field technicians and engineers are the backbone of various industries, including manufacturing, energy, telecom, and more. They play a crucial role in installing, repairing, maintaining, inspecting, and troubleshooting equipment and systems — often in challenging environments.

                                  One of the most significant pain points they encounter is the difficulty of accessing essential information from thick manuals and field service guides while working in the field. Advances in Generative AI (or GenAI), however, can revolutionize the way field technicians and engineers access information and streamline their daily tasks.

                                  Common challenges faced by field technicians

                                  Field technicians and engineers encounter a lot of challenges daily in their line of work. These challenges vary depending on the industry and environment they operate in, but they include:

                                  • Complex tasks: Field technicians often deal with intricate machinery and systems that require specialized knowledge and expertise to operate, maintain, and repair.
                                  • Constantly changing environments: Whether it’s working in remote locations, extreme weather conditions, or high-pressure situations, field technicians must adapt to diverse and unpredictable work environments.
                                  • Information overload: Technicians are frequently required to consult technical manuals, documentation, and safety guidelines to perform their tasks accurately. The sheer volume of information can be overwhelming.  
                                  • Time constraints: Efficiency is crucial, and technicians must perform tasks swiftly without compromising safety or quality.
                                  • Customer expectations: Meeting customer expectations for prompt service and problem resolution is paramount for customer satisfaction and loyalty.
                                  • Wide range of field technicians’ experience levels: While experienced technicians may be able to quickly find the relevant information, more junior ones can struggle to effectively find it.

                                  How GenAI can help

                                  Recent advances in GenAI offer natural language understanding and generation capabilities, making it an ideal solution for field technicians and engineers. It enables them to use voice and text to ask questions in plain English and receive well-articulated answers grounded in products, systems, and processes documentation.

                                  GenAI for field technicians and engineers

                                  Here are some examples of questions that GenAI could reply to:

                                  • How do I replace the filter of machine model XYZ?
                                  • What is the safety process for operation X?
                                  • What are the potential causes of error code 1234?
                                  • How do I calibrate sensor model XYZ?

                                  Benefits of using GenAI for field service work

                                  GenAI can help both transform the worker experience and deliver immense value for the customer. Here we discuss a few benefits Gen AI delivers:

                                  benefits of using GenAI

                                  • Efficiency: GenAI reduces the need to search and browse through lengthy and complex manuals on small screens, allowing technicians to find information more quickly and efficiently.
                                  • Accuracy: GenAI provides instant and relevant answers based on the latest approved documentation, minimizing the risk of errors and accidents.
                                  • Convenience: With hands-free and voice-based interaction, technicians can access information without the need to type or tap on mobile devices, enhancing convenience especially in scenarios where typing is not the best option: gloves, greasy hands, etc.
                                  • Customer satisfaction: Faster and more reliable service delivery translates to improved customer satisfaction, leading to increased trust and loyalty.

                                  FractalGPT for field service work

                                  FractalGPT is a Google Cloud Vertex AI-GenAI so called “LLM”-powered AI chatbot solution that enables field service technicians to quickly find accurate information from company products and process documentation using a natural language interface. It helps enhance the productivity and efficiency of field technicians and engineers.

                                  Some key features and benefits are:

                                  key features and benefits

                                  • Device compatibility: FractalGPT is accessible on any device, whether it’s a phone, tablet, or laptop, via its responsive web app running on your private cloud tenant.
                                  • Customization: Tailor FractalGPT to your company’s machines, processes, and policies, to ensure it meets your specific needs.
                                  • Content breath: FractalGPT can access and analyze any type of documentation, including manuals, guides, datasheets, and more.
                                  • Contextual answers: It generates precise answers that are tailored to the context and user preferences, improving the user experience.
                                  • Complex query handling: FractalGPT can handle complex, multi-step queries that require reasoning and inference, ensuring that technicians get the information they need to complete tasks successfully.

                                    Get started with FractalGPT

                                    In today’s fast-paced world, field technicians and engineers need efficient solutions to overcome the challenges they face daily. FractalGPT offers a transformative approach to addressing these challenges.

                                    It is designed to be secure, scalable, and easy to use, making it ideal for businesses of all sizes and across industries. With FractalGPT, your teams can improve productivity while ensuring your data remains secure.

                                    Leveraging our partnership with Google Cloud and decades of AI expertise, Fractal can swiftly and efficiently assist you in implementing your Generative AI solution using FractalGPT.

                                    Take the first step in improving your field service operations by deploying your custom-branded FractalGPT in a week.

                                    Contact us to learn more about FractalGPT.

                                    Benefits of DevOps

                                    Technology projects require maintenance throughout the entire lifecycle, from development to deployment and monitoring. If maintaining a project from version to version becomes a manual process, it can become quite strenuous. Developers must take special considerations at each stage to ensure smooth rollouts. Failure to do so can result in extended-release planning cycles until the software is ready for use by end users.

                                    Development and IT Operations teams may spend unnecessary cycles supporting and fixing buggy software. And even worse, failed software releases can impact a company’s financial performance through operational inefficiencies, lost sales, and customer attrition. Failure to maintain working software can impact SLAs, and regulatory compliance for some industries, resulting in fines or legal action.

                                    Successful organizations have adapted and created a set of best practices for governing projects called DevOps.

                                    What is DevOps?

                                    DevOps aims to create a common culture that brings together the people, processes, and technology to deliver value (i.e., working software) to end-users.

                                    DevOps illustration

                                    By automating many manual maintenance steps, a successful DevOps plan can also reduce the time it takes to develop, test, and deploy software. Many companies are rushing to implement DevOps to avoid the high costs of manually maintaining projects.

                                    What are the benefits of DevOps?

                                    If you’re asking this question, keep reading. Outlined below are three key benefits of implementing DevOps in your organization.

                                    Benefits of DevOps

                                    1. Improved quality

                                    • Standardized tools and processes (i.e., Azure DevOps and Agile Project Management) help keep quality consistent across projects and releases.
                                    • Quality control implemented through source control branching, code reviews, environment management, release management, etc.
                                    • Reduced fire drills and break-fix measures as a result of following DevOps best practices.

                                    2. Reduced effort

                                    • Fewer manual processes and interventions through improved automation.
                                    • Lower effort to support/maintain because solutions have gone through the appropriate governance and management processes.
                                    • Leverage templates for common infrastructure builds/configurations to accelerate new projects forward in the future.

                                    3. Increased collaboration

                                    • Agile project management structure encourages frequent collaboration among teams.
                                    • Improved communication channels enable the team to identify, track, and manage risks as they arise.
                                    • Clear direction and prioritization through collaboration between stakeholders, developers, and end-users.

                                     

                                    Hopefully, this helps you better understand some of the benefits that implementing DevOps can bring to your business. Implementing DevOps is a journey and is more challenging than installing a package, flipping a switch, or buying new technology.

                                    At Fractal, we specialize in assisting customers at any maturity level in their DevOps journey. We offer strategic planning, technology assessments, proof of concepts, and the technical resources necessary to kickstart your Azure DevOps implementation. Azure for DevOps is a comprehensive platform that provides a wide range of DevOps tools and services and seamlessly integrates with other Azure services, allowing for a unified and scalable solution.

                                    If you’re interested in learning more about how Fractal can help you implement DevOps in your Azure environment, please get in touch with our experts for additional information.