[Day 101] Doing the Google Cloud Digital Leader Learning Path

 Hello :)
Today is Day 101!


A quick summary of today:


After yesterday's attempt and failure to deploy an LLM to production, I wanted to learn a bit more about the cloud platforms used. The top 3 are GCP, Azure and AWS. From past experiences (in my previous job at Lloyds Banking Group) I heard many people using GCP (some Azure), but also that AWS is very popular. I decided to check out what kind of learning paths GCP and AWS have, and I found the GCP's Digital leader learning path and AWS' Cloud Practitioner: Foundamental and both have certification. 

GCP's Digital Leader Learning Path covers the below topics:

  • Digital Transformation with Google Cloud
  • Exploring Data Transformation with Google Cloud
  • Innovating with Google Cloud Artificial Intelligence
  • Modernize Infrastructure and Applications with Google Cloud
  • Trust and Security with Google Cloud
  • Scaling with Google Cloud Operations

and below are my notes.


The cloud

The cloud refers to a network of data centers that store and process information accessible via the internet. It replaces the traditional setup of managing IT infrastructure by combining software, servers, networks, and security into one accessible entity. Cloud technology supports digital transformation by offering scalable, on-demand computing resources through various implementations, including on-premises, private cloud, public cloud, hybrid cloud, and multicloud. On-premises infrastructure involves hosting hardware and software on-site while private cloud dedicates infrastructure to a single organization. Public cloud services are managed by third-party providers and shared among multiple organizations. Hybrid cloud combines different environments, like public and private clouds, while multicloud involves using multiple public cloud providers simultaneously. Most organizations today adopt a multicloud strategy, combining public and private clouds for flexibility and efficiency.

Its benefits

  • scalability
  • flexibility
  • agility
  • strategic value
  • security
Challenges that lead to a digital transformation

  • data management
  • technology infrastructure
  • hybrid workplace
  • security
  • sustainability
Strategies for going on the cloud

  • Private Cloud: Involves virtualized servers in an organization's data centers or those of a private cloud provider, offering benefits like self-service, scalability, and elasticity. It's preferred when significant infrastructure investments have been made or for regulatory compliance
  • Hybrid Cloud: Combines private and public cloud environments, allowing applications to run across different environments. It's a common setup today, enabling organizations to leverage both on-premises servers and public cloud services
  • Multicloud: Involves using multiple public cloud providers, leveraging the strengths of each. Organizations choose multicloud to access the latest technologies, modernize at their own pace, improve ROI, enhance flexibility, and ensure regulatory compliance
Abstractions
  • On-premises: It's like owning a car. You're responsible for its usage and maintenance, and upgrading means buying a new car, which can be time-consuming and costly
  • IaaS (Infrastructure as a Service): It's like leasing a car. You choose a car and drive it wherever you want, but the car isn't yours. Upgrading is easier as you can simply lease a new car
  • PaaS (Platform as a Service): It's like taking a taxi. You provide specific directions, like the code, but the driver (platform provider) does the actual driving (manages the underlying infrastructure)
  • SaaS (Software as a Service): It's like going by bus. You still get access to transport (software), but it's less customizable. Buses have designated routes, and you share the space with other passengers (users)
Responsibilities depending on model

Data Value Chain
The data value chain is like a assembly line, where raw data undergoes various stages to eventually generate actionable insights.
  1. data genesis
  2. data collection
  3. data processing
  4. data storage
  5. data analysis
  6. data activation
Data governance

Data governance involves setting internal standards and policies for data collection, storage, processing, and disposal, while also adhering to external standards and regulations. It aims to make data available across the organization's lifecycle, ensuring accessibility, security, and compliance. Implementing data governance brings several benefits, including increased data value, improved decision-making, cost control, regulatory compliance, enhanced trust, risk management, and efficient data access. Organizations without effective data governance may face compliance violations, poor data quality, delayed analysis, missed opportunities, and reduced accuracy in AI models. As data becomes increasingly crucial, every organization needs robust data governance to navigate digital transformations successfully.

Data storage options

Looker
It is a Google Cloud business intelligence platform, enables individuals and teams to analyze, visualize, and share data through interactive dashboards and reports. It supports integration with BigQuery and over 60 SQL databases, ensuring rich visualizations without compromising performance or security. Looker's web-based interface facilitates easy integration into existing workflows and collaboration across teams.

PUB/SUB and DataFlow
The data pipeline begins with Data Ingestion, where diverse data streams from sources like IoT devices are received, often through Pub/Sub. Dataflow, a serverless and fully managed service by Google, then processes this data for analysis. It handles both streaming and batch data, simplifying ETL processes with Apache Beam's unified programming model. Dataflow's auto-scaling and management features free users from infrastructure concerns, allowing more focus on analyzing insights from datasets.

ML options on GCP


Other full AI solutions

  • Contact centre AI
  • Document AI
  • Discovery AI for retail
  • Cloud talent solution

Cloud migration terms

  • Workload: A specific application, service, or capability that can be run in the cloud or on premises. Examples include containers, databases, and virtual machines
  • Retire: Removing a workload from a platform due to reasons like it being unnecessary, not cost-effective, secure, or compatible with a specific platform
  • Retain: Intentionally keeping a workload, often on premises or in a hybrid cloud environment, under the business's management without full cloud provider control
  • Rehost: Migrating a workload to the cloud without altering its code or architecture, also known as 'lift and shift'
  • Replatform: Migrating a workload to the cloud while making some changes to its code or architecture, often referred to as 'move and improve'
  • Refactor: Changing the code of a workload, possibly to utilize cloud-based microservices or serverless architecture, aiming for efficiency, scalability, or security improvements
  • Reimagine: Rethinking how an organization uses technology, particularly cloud computing, to achieve its business goals, potentially involving reconsideration of cloud strategy and adoption of technologies like AI and ML for efficiency, cost reduction, and agility improvements
Benefits of running compute workloads on the Cloud

  • Total Cost of Ownership (TCO) Reduction: Cloud computing saves money by eliminating the need for purchasing and maintaining physical infrastructure. Pay-as-you-go models and long-term commitment discounts further reduce costs
  • Scalability: Cloud offers the ability to easily scale resources up or down to meet changing demand without significant upfront investments
  • Reliability: Cloud providers offer high reliability and uptime through redundant data centers and automated problem detection and resolution
  • Security: Cloud providers ensure data security through features like encryption, identity and access management, network security, and real-time threat detection and response
  • Flexibility: Organizations can choose and adapt cloud services according to their needs, such as increasing storage space or adding new services
  • Abstraction: Cloud providers handle infrastructure management, allowing customers to focus on their applications without worrying about underlying infrastructure details. This abstraction also enables access to a wide range of services and technologies
Google’s Compute Engine

Google Compute Engine is an Infrastructure as a Service (IaaS) offering from Google Cloud Platform (GCP). It allows users to create and manage virtual machines (VMs) on Google's infrastructure. Compute Engine provides scalable, high-performance virtual machines that can run various workloads, from simple web applications to complex data processing tasks

Essential Cloud Security Terms and Concepts

  • Privileged Access Security Model: Grants specific users broader access to resources, but misuse can pose risks. Requires careful management and monitoring
  • Least Privileged Security Principle: Advocates granting users only the access needed for their job responsibilities, reducing the risk of unauthorized access to sensitive data
  • Zero-Trust Architecture: Assumes no user or device can be trusted by default. Requires authentication and authorization before accessing resources, ensuring robust security
  • Security by Default: Integrates security measures into systems and applications from the initial stages of development, establishing a strong security foundation in Cloud environments
  • Security Posture: Evaluates an organization's overall security status in its Cloud environment by assessing security controls, policies, and practices
  • Cyber Resilience: The ability to withstand and recover quickly from cyber attacks by identifying, assessing, and mitigating risks, responding to incidents effectively, and recovering from disruptions quickly

Data sovereignity and Data residency
  • Data Sovereignty: Refers to the legal concept that data is subject to the laws and regulations of the country where it resides. For example, GDPR in the EU mandates compliance with data protection laws for EU citizens' personal data
  • Data Residency: Refers to the physical location where data is stored or processed. Some countries require data to be stored within their borders to ensure compliance with local laws
GCP access levels



Operational excellence and reliability at scale


Service level indicators, objectives and agreements



That is all ^^


I will also look at AWS' for-beginners course and certification and try to make an informed decision into which exam to invest.

That is all for today!

See you tomorrow :)

Popular posts from this blog

[Day 198] Transactions Data Streaming Pipeline Porject [v1 completed]

[Day 107] Transforming natural language to charts

[Day 54] I became a backprop ninja! (woohoo)