status-updates

Week 2 Status Updates

Monday: Fresh Start

Beginning the week with new challenges and opportunities.

Python Debugging Adventures 🐞

Started the week by diving back into the Udemy course that was handed to us last week. Today’s focus was particularly interesting - Python debugging! I learned how to add breakpoints to evaluate how code works, which is like having X-ray vision into your program’s execution.

To put this into practice, I created a simple sum function and used VS Code’s debugging tools to understand its inner workings. It’s fascinating how you can pause the code execution at any point and inspect variables!

Here’s what the the process looked like:

Debugging in Action

This hands-on experience with debugging tools has already made me feel more confident about troubleshooting code issues. It’s amazing how these tools can make the debugging process so much more intuitive!

Diving into Google Cloud Platform ☁️

After getting comfortable with debugging, I decided to explore Google Cloud Platform (GCP). As someone new to cloud computing, I found this really exciting! I discovered a super helpful YouTube video that breaks down GCP for beginners in a really clear way.

Check out the beginner-friendly guide here: GCP Tutorial

GCP Learning

The way GCP organizes its services and projects is fascinating, I also learnt about

Cloud Services Deep Dive 🌐

To broaden my understanding, I watched an excellent video by Neetcode that explains the generic concepts behind all cloud services. It was eye-opening to learn about:

This comprehensive overview really helped connect the dots!

Check out Neetcode’s explanation here: Cloud Services Overview

Cloud Concepts

Understanding these fundamental concepts has given me a solid foundation for working with any cloud platform.

Data Storage Architecture 📊

Next, I delved into understanding the key differences between data lakes and data warehouses. This exploration was fascinating as it helped me understand how organizations handle different types of data at various scales.

A data lake is like a vast reservoir that can store any type of data in its raw form, while a data warehouse is more like a refined storage system with structured, processed data ready for specific business needs.

Here’s a visual comparison that helped me understand the differences:

Data Lake vs Data Warehouse

This knowledge is crucial for understanding modern data architectures and making informed decisions about data storage solutions.

Cloud Providers Comparison 🌩️

After understanding data storage concepts, I dove into exploring the “Big Three” cloud providers: AWS, Azure, and GCP. Each platform has its unique strengths and specialties, making this comparison really interesting!

Amazon Web Services (AWS)

The pioneer in cloud computing, AWS offers the broadest set of services.

Microsoft Azure

Azure’s integration with Microsoft’s ecosystem is impressive. Got to explore their interface:

Google Cloud Platform (GCP)

GCP’s strength in data analytics and machine learning is notable. Here’s their dashboard:

AWS Console

AWS

AWS Console


Tuesday: Deep Dive

Reverse Knowledge Transfer Presentation: Rising to the Challenge

In a pivotal moment of professional growth, I delivered a reverse knowledge transfer presentation to senior leadership. The atmosphere was electric as I shared my journey of adaptation and triumph. With gratitude in my heart, I began by acknowledging the precious time our seniors invested in hearing our voices.

The Ubuntu Chronicles: Conquering New Territories

From the familiar shores of Windows to the uncharted territories of Ubuntu, my journey was nothing short of transformative. The initial days were a dance with uncertainty, but as they say, “smooth seas never made a skilled sailor.” I emerged victorious, mastering:

Each challenge was met with unwavering determination, turning obstacles into stepping stones.

presentation

Feedback: Pearls of Wisdom

The presentation sparked a cascade of invaluable feedback,(which ofcourse I cant mention in depth here xD) illuminating the path forward. Key focus areas emerged like constellations guiding a night traveler:

AWS Expedition: Venturing into the Cloud ☁️

The day’s journey didn’t end with the presentation. Like a determined explorer, I ventured into the vast expanse of AWS, discovering:

Core Services Discovered

  1. EC2 (Elastic Compute Cloud)
    • The beating heart of cloud computing
    • Virtual servers in the cloud
  2. S3 (Simple Storage Service)
    • The fortress of data storage
    • Buckets of infinite possibilities
  3. IAM (Identity and Access Management)
    • Crafted a new user account
    • Bestowed administrative powers while maintaining security
    • A strategic move away from root access

Key Achievements 🏆

Looking Ahead 🔭

Armed with feedback and newfound knowledge, the path ahead is clear. Each step forward is a step toward mastery in this ever-evolving technological landscape.


Wednesday: AWS Adventures & Big Data Discoveries! 🚀

Today was an absolute rollercoaster of cloud computing and big data exploration!

Morning AWS Shenanigans ☁️

Started my day by diving deep into AWS fundamentals with this awesome tutorial: AWS Full Course for Beginners

Virtualization: The Matrix of Cloud Computing 🤯

Learned about virtualization - basically

EC2: My Virtual Computer in the Sky! 💻

The excitement was real when I launched my first EC2 instance! Learned about:

Security Groups: The Bouncers of the Cloud 🚧

Got to know about security groups in network settings - they’re like the cool bouncers who decide who gets in and who doesn’t!

SSH Adventures: Local CLI meets Cloud CLI 🤝

Successfully connected my local command line to the cloud instance using SSH - felt like a proper cloud ninja!

S3 Bucket Brigade! 🪣

Quick Visits to Other AWS Services 🎯

Big Data Finale: A Journey Through Time 📚

Ended the day with some mind-bending big data history:


Thursday: AI/ML Cloud Adventures! 🤖

Diving into the fascinating world of AWS’s AI and ML services and completing an assignment

Amazon SageMaker: Where ML Dreams Come True ✨

Got my hands on SageMaker today - AWS’s powerhouse for machine learning! It’s like having a full-stack ML workshop in the cloud:

 SageMaker Image

Space for SageMaker Image

Amazon Bedrock: The LLM Playground 🎮

Explored Bedrock - AWS’s newest addition to the AI family! It’s like having a VIP pass to the world of Large Language Models:

Bedrock Image

Bedrock

Bedrock

Boto3: The Python Whisperer 🐍

Got introduced to Boto3 - the Swiss Army knife for AWS automation in Python! This SDK is amazing:

First Assignment: Time to Put Knowledge to Work! 💪

After exploring these awesome AWS services, it was time to tackle my first assignment from my senior !

Space for Assignment Image

The requirements were straightforward so I began to work , I explored various repo’s to see how production grade code is maintained as I didnt want to submit just anyother assignment I wanted to submit THE ASSIGNMENT , I always make it dramatic XD

code-repo

code-repo

After a bit of research I got the idea on how to properly structure my code and present it in a professional manner , so I got to work and ended up with this github-repo and this web-page

web-app

with this webapp-readme.Pretty exciting stuff for me.

During the assignment learnt about:

Closing out with Intro to Distributed Computing and PySpark 🌟

Ended the day by diving into the fascinating world of distributed computing! Here’s what caught my attention:

The Evolution: From Hadoop to Spark 🚀

Discovered how the big data landscape evolved:

distributed

pandas vs dask

MapReduce: The OG of Distributed Processing 🗺️

Got my mind blown learning about MapReduce:

pandas vs dask

Framework Face-off: Dask vs Pandas vs PySpark 🥊

Compared different distributed computing frameworks:

pandas vs dask

Watched some tutorials to prepare for tomorrow’s deep dive into PySpark. Can’t wait to get my hands dirty with some real distributed computing! 🔥


Friday: Deep Dive into Spark’s Universe! 🌟

A day of unraveling Spark’s architecture and getting hands-on with distributed computing

Spark Infrastructure: The Master Plan 🏗️

Discovered the fascinating architecture of Spark:

Spark Architecture Image

Spark Architecture Image

Spark Architecture Image

RDDs: The Building Blocks of Spark Magic ✨

Got mind-blown understanding Resilient Distributed Datasets (RDDs):

rdd Image

Hadoop vs Spark: The Speed Revolution 🚀

Understood the key differences:

hadoop Image

Hands-on Adventures 🛠️

Successfully set up my local Spark playground:

hands-on

hand2-

hand2-

hand2-

The Art of RDD Operations 🎨

Mastered the two types of RDD operations:

  1. Transformations:
    • Lazy evaluated
    • Create a DAG workflow
    • Don’t execute until an action is called
  2. Actions:
    • Instant execution
    • Trigger the actual computation
    • Return results to driver program

Space for RDD Operations DAG Image

Space for RDD Operations DAG Image

Space for RDD Operations DAG Image

Key Takeaways 🎯

Looking Forward 🔭

Today was heavy on theory, but it laid a solid foundation! Can’t wait to dive deeper into practical implementations and see these concepts in action!


Saturday: Hands-on with Spark & Modern Data Tools! 🛠️

A day of practical exploration and clearing up common confusions

Morning CLI Adventures ⌨️

Started the day by getting familiar with Spark on the command line:

Space for CLI Setup Image

The Evolution: From RDDs to DataFrames 📈

Learned about how Spark data handling evolved:

Space for RDD vs DataFrame Image

Space for RDD vs DataFrame Image

Databricks Playground Time! 🎮

Had fun exploring Databricks:

Space for Databricks Interface Image

Diving into Data Lakes & Delta Lake 🌊

Discovered the world of modern data storage:

Space for Delta Lake Architecture Image

Clearing Up the PySpark vs Hadoop Confusion 🤔

Finally understood the big picture:

Hands-on with PySpark 🐍

Completed some tutorials to understand:

Key Learnings 🎯


Sunday: The Great AWS Billing Adventure! 💰

What started as a normal coding day turned into an AWS billing investigation!

Morning Spark Session ⚡

Started the day productively with local Spark experiments:

The Plot Twist: AWS Bill Discovery! 😱

While preparing to start an assignment, discovered an unexpected AWS bill from my practice sessions.

The Investigation Journey 🔍

Dug deep into AWS console to find the issue:

  1. SageMaker Services:
    • Found active domain services
    • Promptly deleted them
  2. IAM Cleanup:
    • Discovered ~30 active roles
    • Cleaned them all up
  3. Network Resources:
    • Found an active NAT Gateway (sneaky!)
    • Spotted and removed an Elastic IP
    • Terminated all unnecessary services

The Plot Thickens! 📈

Silver Linings ✨

Even though the original assignment had to wait, learned valuable lessons:

Action Items for Next Week 📝

Final Thoughts 💭

What started as a billing surprise turned into an intensive learning experience about AWS resource management. Looking forward to Week 3 with excitement 🚀