Found 187 repositories(showing 30)
DataWithBaraa
End-to-end Data Lakehouse project built on Databricks, following the Medallion Architecture (Bronze, Silver, Gold). Covers real-world data engineering and analytics workflows using Spark, PySpark, SQL, Delta Lake, and Unity Catalog. Designed for learning, portfolio building, and job interviews.
mesbahiba
Gain the job-ready skills for an entry-level data analyst role through this eight-course Professional Certificate from IBM and position yourself competitively in the thriving job market for data analysts, which will see a 20% growth until 2028 (U.S. Bureau of Labor Statistics). Power your data analyst career by learning the core principles of data analysis and gaining hands-on skills practice. You’ll work with a variety of data sources, project scenarios, and data analysis tools, including Excel, SQL, Python, Jupyter Notebooks, and Cognos Analytics, gaining practical experience with data manipulation and applying analytical techniques.
This program includes over 180 hours of instruction and hundreds of practice-based assessments, which will help you simulate real-world data analytics scenarios that are critical for success in the workplace. The content is highly interactive and exclusively developed by Google employees with decades of experience in data analytics. Through a mix of videos, assessments, and hands-on labs, you’ll get introduced to analysis tools and platforms and key analytical skills required for an entry-level job. Skills you’ll gain will include: Data cleaning, problem solving, critical thinking, data ethics, and data visualization Platforms and tools you will learn include: Presentations, Spreadsheets, SQL, Tableau and R Programming In addition to expert training and hands-on projects, you'll complete a case study that you can share with potential employers to showcase your new skill set. Learn concrete skills that top employers are hiring for right now.
khushal2405
We Build an ETL pipeline using Airflow that accomplishes the following: Downloads data from an AWS S3 bucket, Runs a Spark/Spark SQL job on the downloaded data producing a cleaned-up dataset of delivery deadline missing orders and then Upload the cleaned-up dataset back to the same S3 bucket in a folder primed for higher level analytics
shbnq422
بسم الله الرحمن الرحيم C.V. NAME: SHA'BAN YOUSEF MOHAMMAD ELQASASS TELE PHONE NO. : 0799455189 NATIONALITY: JORDANIAN ADDRESS: JORDAN-AMMAN E-MAIL: SHBNQ422@YAHOO.COM WWW.SHAABANALQASSAS.COM CERTIFICATES: • HIGHER DIPLOMA IN INFORMATION TECHNOLOGY ( THE UNIVERSITY OF JORDAN) • B.CS IN CHEMISTRY AND COMPUTER FROM YARMOUK UNIVERSITY JORDAN • MCAD.NET (MICROSOFT CERTIFIED APPLICATION DEVELOPER 2003) • ADO.NET; ASP.NET ; MVC :LINQ :VB.NET : C# • MS SQL SERVER DATABASE 2005-2008-2012-2014 • ACCESSING DATA WITH MICROSOFT .NET FRAMEWORK 4 CERTIFICATE -MICROSOFT • JOOMLA WEBSITES DEVELOPER • PHP; MYSQL DATABASE; • MICROSOFT CERTIFIED TECHNOLOGY SPECIALIST (MCTS 2013) • CERTIFICATE DEVELOPING DATA ACCESS SOLUTIONS- MICROSOFT • CERTIFICATE DEVELOPING WEB APPLICATIONS – MICROSOFT • CERTIFICATE ADMIN TEAM FOUNDATION SERVER – MICROSOFT • CERTIFICATE DEVELOPING SILVERLIGHT – MICROSOFT • OFFICIAL TRANING IN ARCGIS 10.1 (I) – BY ESRI • OFFICIAL TRANING IN ARCGIS 10.1 ( II) – BY ESRI • OFFICIAL TRANING IN ARCGIS 10.1 ( III) performing analysis – BY ESRI • OFFICIAL TRANING IN ARCGIS 10.1 ( BUILDING GEODATABASE) – BY ESRI • OFFICIAL TRANING IN IMPLEMENTING VERSIONED WORKFLOWS IN A MULTIUSER GEODATABASE – BY ESRI • OFFICIAL TRANING IN ARCGIS 10.1 ( BUILDING WEB APPLICATION USING THE ARCGIS API ) – BY ESRI • OFFICIAL TRANING IN ARCGIS 10.1 ( INTRODUCTION TO ARCGIS SERVER) – BY ESRI • OFFICIAL TRANING IN ARCGIS 10.1 (CONFIGURING & MANAGING THE MULTIUSER GEODATABASE (SDE)) – BY ESRI • CERTIFICATE MAINTAINING MICROSOFT SQL SERVER DATABASE-MICROSOFT • CERTIFICATE IMPLEMENTING MICROSOFT SQL SERVER DATABASE REPORTING SERVICES- MICROSOFT • CERTIFICATE SHAREPOINT SERVER ADMINISTRATION- MICROSOFT • CERTIFICATE SHAREPOINT SERVER DEVELOPMENT- MICROSOFT • BUSINESS INTELLIGENCE TOOLS :DATA ANALYSIS SERVICES, DATA TRANSFORMATION SERVICES (DTS), DATA MIGRATION SERVICES, ETL (EXTRACT, TRANSFORM, AND LOAD) SERVICES, • ONLINE ANALYTICAL PROCESSING (OLAP) AND DATA-MINING TOOLS • CERTIFICATE WRITING QUERIES USING MICROSOFT SQL SERVER TRANSACT SQL –MICROSOFT • CERTIFICATE IMPLEMENTING AND MAINTAINING MICROSOFT SQL SERVER ANALYSIS SERVICES –MICROSOFT • CERTIFICATE IMPLEMENTING AND MAINTAINING MICROSOFT SQL SERVER INTEGRATION SERVICES –MICROSOFT • CERTIFICATE ADMINISTERING TEAM FOUNDATION SERVER • SYSTEM ADMINISTRATOR • Adobe program (graphics & design) • MS - SQL SERVER REPORTING SERVICE-OLAP-BI • SPSS ADVANCED TRAINING • ICDL CERTIFICATE • GEOMEDIA PROFESSIONAL GIS • GEOMEDIA WEBMAP GIS • GEOMEDIA TRANSPORTAION GIS • INTEL CERTIFICATE FUTURE FOR LEARNING • A+ COMPUTER MAINTENANCE (HARDWARE & SOFTWARE) • NETWORK + • SPECIALIZED COURSE IN METHODS OF CALCULATING THE RATE OF POPULATION GROWTH / DEPARTMENT OF STATISTICS • COURSE COMPLETION DEMOGRAPHIC / DEPARTMENT OF STATISTICS • SPECIALIZED COURSE IN THE MEASUREMENT OF IMMIGRATION / DEPARTMENT OF STATISTICS. • Develop GIS websites using GIS SILVERLIGHT asp.net technology. • Develop GIS websites using GIS flex technology. • Certificate projection methods and simulation models-UNESCO • Specialized course in fertility / Department of Statistics. • Specialized course in the measurement of levels and patterns of DEATH / DEPARTMENT OF STATISTICS • COURSE CODE POPULATION PROJECTIONS / DEPARTMENT OF STATISTICS • SPECIALIZED COURSE IN THE PREPARATION OF PROJECTIONS IN CRISES AND STRATEGIC PLANNING - UNESCO • WEB DESIGN USING DREAM WAVER • ASSOCIATE REGIONAL WORKSHOP ON EDUCATION STATISTICS - UNESCO INSTITUTE FOR STATISTICS • USING SPSS PROGRAM • ADVANCED COURSE IN STRATEGIC PLANNING AND THE PREPARATION OF INDICATORS • ADVANCED COURSE IN-DEPTH DEMOGRAPHIC ANALYSIS (U.S.BUREAU OF CENSUS) • ADVANCED COURSE IN POPULATION PROJECTIONS (U.S.BUREAU OF CENSUS) • EMIS EDUCATIONAL MANAGEMENT INFORMATION SYSTEM EXPERIENCES: • HEAD OF EMIS AND GIS DIVISION ) EDUCATIONAL MANAGEMENT INFORMATION SYSTEM( • HEAD OF PLANNING AND STATISTICS DIVISION • SUPERVISOR IN THE STATISTICS OF THE POPULATION IN THE KINGDOM IN 2004 AND DESIGNED STATISTICAL SOFTWARE.PARTICIPANT IN THE DEVELOPMENT OF DATABASES AND SOFTWARE EDUCATIONAL INDICATORS SIDE WITH GIS MAPS OF SCHOOL • TEAM LEADER IN DEVELOPMENT OF INFORMATION SYSTEM PROJECT ERSP • TEAM LEADER IN DEVELOPMENT OF INFORMATION SYSTEM PROJECT OPENEMIS • PARTICIPANT IN THE DEVELOPMENT OF DATABASES AND SOFTWARE STATISTICAL INFORMATION AND SURVEY FOR SCHOOL FOOD.. UNITED NATIONS WORLD FOOD PROGRAMME (WFP) • PARTICIPANT IN THE DEVELOPMENT OF DATABASES AND SOFTWARE STATISTICAL INFORMATION EMIS-EDUWAVE. • DEVELOP AND MANAGE IMPLEMENTATION OF PROJECT COMMUNICATION STRATEGIES AND COORDINATE CHANGE MANAGEMENT. • NATIONAL COORDINATOR OF THE EDUCATION INDICATORS WEI "WORLD EDUCATION INDICATOR" -UNESCO • DATABASE SPECIALISTS OF PARLIAMENTARY ELECTIONS • DATABASE ADMINISTRATION • DATABASE DESIGN, IMPLEMENTATION, TRANSACTION PROCESSING, REFERENTIAL INTEGRITY, PERFORMANCE TUNING AND MONITORING WITH AN EMPHASIS IN MS SQL. • MONITOR, REVIEW, AND VERIFY BACKUPS, RECOVERY, AND DISASTER RECOVERY PROCESSES. • CONTROL AND MONITOR USER ACCESS TO DATABASES AND MAINTAIN DATABASE SECURITY. • TROUBLESHOOT JOB FAILURES, PERFORMANCE ISSUES, DATABASE INTERGRITY, ETC. • MONITOR ALERTS AND TAKE CORRECTIVE ACTION OR ESCALATE ISSUES. • RESOLVE DATABASE REQUESTS AND ISSUES. • PERFORM DATABASE TUNING FOR OPTIMAL PERFORMANCE. • WORK CLOSELY WITH DEVELOPMENT AND NETWORK TEAMS ON RESOLVING ISSUES. • MAINTAIN DOCUMENTATION OF SYSTEMS AND STANDARD OPERATING POLICIES AND PROCEDURES. • ASSOCIATE IN THE DEVELOPMENT OF PARLIAMENTARY ELECTIONS SCHEDULES FOLLOW-UP AND THE RESULTS. • PARTICIPANTS IN THE INFORMATION SYSTEMS INTEGRATION PROJECT IN THE MINISTRY OF EDUCATION. • PARTICIPANT IN THE DEVELOPMENT OF DATABASES AND APPLICATION FOR SYRIAN REFUGEES CRISIS INFORMATION SYSTEM • PARTICIPANT IN THE LAMP (LITERACY ASSESSMENT AND MONITORING PROGRAM) –UNESCO • PARTICIPANT IN THE OUT OF SCHOOL CHILDREN STUDY – WITH UNICEF AND UNESCO • PARTICIPANT IN THE DEVELOPMENT OF THE DIRECTORATE AND THE SCHOOL (SDDP), • PARTICIPANT IN WORKSHOP FOR CHANGE MANAGEMENT –USAID • PARTICIPANT IN WORKSHOP FOR STATISTICAL INFORMATION –UNESCO &ALECSO- TUNIS • PARTICIPANT IN WORKSHOP FOR MONITORING AND EVALUATION -USAID • DESIGN AND CONSTRUCTION OF APPROPRIATE EDUCATIONAL POLICIES. • ADVANCED COURSE IN LEADERSHIP IN STRATEGIC COMMUNICATION • ADVANCED COURSE IN USING SPSS FOR DATA ANALYSIS • ADVANCED COURSE IN USING SPSS FOR DEMOGRAPHIC HEALTH SURVEYS • DESIGNER STATISTICAL SOFTWARE ESPECIALLY IN THE CAMPAIGN PROVIDES YALA MAY IN AMMAN IN 2009. • WEB SITE DEVELOPER AND DESIGN • PREPARE AND PARTICIPATE IN WORKSHOPS, INTERNATIONAL INDICATORS AND STRATEGIC PLANNING AND THE USE OF SOFTWARE AND DEVELOPMENT • PROGRAMMER IN PLANNING DIVISION SINCE 1/2004 • TEACHER FOR COMPUTER AND CHEMISTRY FROM 8/1995 • TRAINER FOR VISUAL BASIC.NET ADO.NET ASP.NET AND ICDL CERTIFICATE الاسم: شعبان يوسف محمد القصاص هاتف: 0799455189 فاكس :5666492 الجنسية: أردني العنوان: الأردن -عمان البريد الالكتروني: SHBNQ422@YAHOO.COM WWW.SHAABANALQASSAS.COM المؤهل العلمي: دبلوم عالي في تكنولوجيا المعلومات الجامعة الأردنية بكالوريوس كيمياء و حاسوب من جامعة اليرموك الأردنية عام 1993 شهادة مطور برمجيات ميكروسوفت من شركة ميكروسوفت العالمية ( MCAD .NET ) MS SQL server DATABASE 2005-2008-2012-2014 /ASP.NET/ ADO.NET/ VB.net C#: MVC : LINQ PHP ; MYSQL DATABASE دورة "أخصائي تكنولوجيا معتمد من مايكروسوفتMCTS MS - SQL SERVER REPORTING SERVICE –OLAP-BI MS ANALYSIS SERVICE System administrator Joomla website developer شهادةNETWORK + الدولية في شبكات الحاسوب شهادة التعليم للمستقبل INTEL لحوسبة التعليم شهادةA+ الدولية في صيانة الحاسوب شهادة نظم إدارة المعلومات التربوية EMIS وشهادة الخرائط الجغرافية GIS تدريب متخصص ArcGIS10.1 (I) شركة ESRI العالمية تدريب متخصص ArcGIS10.1 (II) شركة ESRI العالمية تدريب متخصص ArcGIS10.1 (III) performing analysis شركة ESRI العالمية تدريب متخصص ArcGIS10.1 (building geodatabase) شركة ESRI العالمية تدريب متخصص SharePoint إدارة وتطوير GEOMEDIA PROFESSIONAL GIS GEOMEDIA WEBMAP GIS GEOMEDIA TRANSPORTAION GIS دورة استخدام برمجيات المؤشرات والتحليل الاحصائي devinfo spectrum -- دورة متخصصة في اساليب حساب معدل النمو السكاني / دائرة الاحصاءات العامة دورة متخصصة في الاستكمال الديمغرافي / دائرة الاحصاءات العامة دورة متخصصة في قياس الهجرة / دائرة الاحصاءات العامة دورة متخصصة في الخصوبة / دائرة الاحصاءات العامة دورة متخصصة في قياس مستويات وانماط الوفاة / دائرة الاحصاءات العامة دورة متخصصة في برمجية الاسقاطات السكانية / دائرة الاحصاءات العامة Adobe program (graphics & design) تصميم مواقع الانترنت باستخدام Dream Waver استخدام برمجية SPSS في التحليل الاحصائي دورة متقدمة في التخطيط الاستراتيجي واعداد المؤشرات دورة متخصصة في اعداد الاسقاطات في الازمات والتخطيط الاستراتيجي - يونيسكو دورة متقدمة في التحليل الديمغرافي المعمق U.S.BUREAU OF CENSUS دورة متقدمة في الاسقاطات السكانية U.S.BUREAU OF CENSUS الخبرات: • رئيس قسم نظام إدارة المعلومات التربوية EMIS والخرائط GIS في وزارة التربية • رئيس قسم التخطيط والاحصاء التربوي في وزارة التربية والتعليم من تاريخ 2006 . • اعداد والمشاركة في ورش المؤشرات الدولية والتخطيط الاستراتيجي واستخدام برمجياتها وتطويرها • مشرف في الاحصاءات العامة للسكان في المملكة للعام 2004 ومصمم برمجيات احصائية. • تطوير مواقع انترنت الجغرافية باستخدام تقنية GIS SILVERLIGHT asp.net • تطوير مواقع انترنت الجغرافية باستخدام تقنية GIS flex • اخصائي قواعد بيانات في الهيئة المستقلة للانتخاب –الأردن • مدير فريق تطوير نظام إدارة المعلومات التربوي والمؤشرات التربية ERSP • مدير فريق تطوير نظام إدارة المعلومات التربوي والمؤشرات التربية OpenEMIS • مشارك في تطوير قواعد بيانات و معلومات الطلبة اللاجئين السوريين • مشارك في ورشة عمل في إدارة التغيير – USIAD • مشارك في ورشة عمل في المتابعة والتقييم - USIAD • مشارك في تطوير قواعد بيانات و برمجيات المؤشرات التربوية باستخدام الخرائط المدرسية GIS • مشارك في الانتخابات النيابية تطوير جداول المتابعة واصدار النتائج. • مشارك في تطوير قواعد بيانات وبرمجيات المعلومات الاحصائية EMIS-Eduwave • مشارك في مشروع تطوير المديرية والمدرسة (SDDP) ورسم وبناء السياسات التربوية . • مشارك في مشروع تكامل انظمة المعلومات في وزارة التربية والتعليم • مشارك في مشروع احصائيات معرفة القراءة والكتابة برنامج تقييم وتتبع مستويات القرائية LAMP • دورة متقدمة في القيادة في استراتيجية الاتصال . جامعة ميتشغان الامريكية • المنسق الوطني لمؤشرات التعليم WEI "World Education Indicator " -UNESCO • مشارك في ورشة عمل الاقليمية حول احصاءات التعليم – معهد اليونسكو للاحصاء • مصمم برمجيات احصائية خاصة في حملة يالا نوفر مي في عمان للعام 2009 • مطور ومصمم مواقع انترنت • مبرمج في قسم التخطيط في مديرية التربية والتعليم .تطوير برامج الحاسوب لحوسبة أعمال القسم من تاريخ 2004 • مدرس لمادتي الكيمياء والحاسوب للمرحلة الثانوية والأساسية من تاريخ 1995 • مدرب لبرمجية visual basic .net asp.net ado.net من تاريخ 2004 • مدرب لشهادة الحاسوب الدولية ICDL)) من تاريخ 2001
Naseer5196
Indeed Home - For employers Dashboard Find resumes Analytics Need Help? Start of main content Jobs Candidates Messages Search candidates Search candidates Data Engineer -Immediate Joiner (Work From Office) Vedhas Technology Solutions Pvt Ltd – Hyderabad, Telangana Clicks Your job 17/09/21 18/09/21 19/09/21 20/09/21 21/09/21 22/09/21 23/09/21 0 15 30 Clicks this week 63 Candidates Awaiting Review 6 Total (excluding rejected) 6 0 Rejected Discover your top applicants faster by sending a free assessment Get a more complete picture of each candidate by being able to view and compare their assessment score results when you turn on the assessment of your choice. Job description Required Data Engineer - Work From Office Location: Himayatnagar, Hyderabad Experience: 2 - 4 yrs. Job Description: · Experience in Big Data components such as Spark, Kafka, Scala/PySpark, SQL, Data frames, Airflow etc. implemented using Data Bricks would be preferred. · Databricks integration with other cloud services like (Azure - Data Lake, Data Factory, Synapse, Azure DevOps, etc.) or (AWS S3, GLUE, Athena, Redshift, Lambda, CloudWatch etc.) · Reading, processing, and writing data in various file formats using Spark & Databricks. · Knowledge of best Databricks Job Optimization process and standards. Good to Have: · Databricks Delta Table & ML-Flow knowledge will be a plus. · AWS/Azure/Databricks Certifications will be a plus · Strong Data Warehousing experience · Good understanding of Database schema, design, optimization, scalability. · Ability to learn new technologies quickly. · Great communication skills, strong work ethic. Role Data Engineer Industry Type IT Services & Consulting Functional Area IT Software - Application Programming, Maintenance Employment Type Full Time, Permanent Education: UG- B. Tech/B.E. in Any Specialization Key Skills: Data Bricks, Data Lake, Kafka, Azure DevOps, SQL. Remuneration: No Bar for Right Candidate. Work Shift: Day Working Days: 5 per week Location: Vedhas Technology Solutions Pvt Ltd 1st Floor City Centre Himayatnagar, Hyderabad -500029. Email ID: HR@TECHVEDHAS.COM Contact HR: 040-23224181.
ikebude
Data Analytics Job simulation using Excel, Power BI, SQL, python and R programming language.
ETL-monitoring-team
Enterprise-grade ETL Monitoring Dashboard for job scheduling, log tracking, staging analysis, sales analytics, and fact exploration. Built with ASP.NET Core MVC, Entity Framework, and a SQL-based data warehouse.
We Build an ETL pipeline using Airflow that accomplishes the following: Downloads data from an AWS S3 bucket, Runs a Spark/Spark SQL job on the downloaded data producing a cleaned-up dataset of delivery deadline missing orders and then Upload the cleaned-up dataset back to the same S3 bucket in a folder primed for higher level analytics
Madzzzzz
Important Note: This position is for full-time and please apply only if you have a Green card or Citizenship in the United States. The Client will not provide sponsorship. Opportunity: One of our major clients in the United States is looking for a “Software Development Engineer” for a Fulltime opportunity for completely remote. Job description: Position: Data Engineer Position Type: Full-time Location: Glendale, CA We are looking for a Data Engineer to work with one of our reputed client. Our client is a leading Data Analytics and AI Consulting. They enable business transformation through the intelligent application of data, analytics and AI in a multi cloud environment, empowering businesses with data-driven decision making to improve financial performance making to drive transformational outcomes Data engineering role requires creating and managing technological infrastructure of a data platform, be in-charge / involved in architecting, building, and managing data flows / pipelines and construct data storages (noSQL, SQL), tools to work with big data (Hadoop, Kafka), and integration tools to connect sources or other databases. Requirement: • Experience in developing SQL stored Procedures for extracting, cleaning, transforming, and loading data • Experience in writing optimized SQL queries for integration using python • Experience with database objects tables, external tables, views, triggers, functions • Experience in working with database security a plus • Experience in Snowflake, AWS Redshift, SAP Hana is a plus • Experience in design, implement, and monitor regular backups, disaster recovery plans is a plus Key skills: Python, SQL / PL SQL
emedinak
Interactive data analytics dashboard analyzing Data & BI job demand in Spain using Adzuna API, Python, SQL, and Streamlit.
jeremyrieunier
Anonymized data analytics and analytics engineering case studies from job interviews. Features SQL, Python, data modeling, forecasting, A/B testing, and attribution analysis solutions for real business problems.
Ritikpal07
📊 Data Analytics Project: Analyzed 1200+ job listings to uncover hiring trends, salary patterns & skill demands in India's data job market. Built with Python, SQL, and interactive dashboard.
kreeshvar
End-to-end exploratory data analytics project analyzing job market trends (2023–2024). Built using Python (Pandas ETL), SQL Server (stored procedures, views, data modeling) and Power BI for market insights on high-paying job postings.
saideeppallela
📊 120-day public worklog of real Data Analytics practice using SQL, Excel, Power BI & Python. 🔍 Shows daily analyst workflows: data exploration, cleaning, analysis, visualization & insights. 🚀 Built to demonstrate consistency, problem-solving & job-ready analytics skills.
Ozee08
This project analyzes job postings data using **SQL** to uncover insights about in-demand skills, salaries, and top companies. It demonstrates how SQL can be applied to **real-world data** for decision-making and is part of my data analytics portfolio.
ckotkar8860
🔹 Use Case (Pharma Client) Simulate a pharma analytics pipeline using 3 source tables: Source Tables patients prescriptions drug_master We build: Bronze: Raw ingestion Silver: Cleaned & conformed data Gold: Business-ready analytics tables 🔹 Tech Stack PySpark Databricks Delta Lake Databricks Jobs SQL Medallion Architecture
lukejones3
A data analytics project transforming unstructured job postings into structured datasets using SQL and Python, with Power BI visualizations to analyze skill requirements, salary ranges, and labor market trends.
shivalikasingh19
An analytical study of global Data Science job salaries using Excel, SQL, and Power BI. Project includes data cleaning, salary comparison by job role and experience level, top-paying countries, and interactive visual dashboards.
hariommmofficial
Fake Job Posting Analysis is a data analytics project that identifies patterns in fraudulent job listings using Python, SQL, and data visualization. It analyzes features like missing information, email domains, keywords, and employment type to detect scam indicators and highlight risk factors that distinguish fake job postings from legitimate ones.
Vidhya-bharathi-raj
In this job simulation, I completed tasks in data analysis and visualization using Excel, Power BI, SQL, and Power Point, showcasing analytical skills and problem-solving.
karkakasadara-tharavu
💼 Complete career transformation path: BE Graduate → Data Engineer ($65K-$130K). Master SQL Server administration, T-SQL programming, SSIS ETL pipelines, Power BI analytics. 593KB content, 75+ files, AdventureWorks databases. Learn database design, normalization, backup strategies, security, CDC, dimensional modeling. Job-ready in 5 months.
Streaming is data processing for unbounded datasets (infinite datasets). Data is in motion. Cloud Pub/Sub connect applications and service through a messaging infrastructure. Pub/Sub is a global messaging queue, essentially a message bus (buffer). Message bus is reliable, has high throughput, and low latency. Pub/Sub is about capturing data and distributing data. It is serverless and global. PubSub can be the source and BigQuery for the sink when streaming events. DataFlow does batch and streaming as long as code does not change (can control late-arriving data and out-of-order data). It does continuous computations, continuous queries. Dataflow does autoscaling and rebalancing. Stream processing is best used with DataFlow. Dataflow resources are deployed on demand, per job, and work is constantly rebalanced across resources. BigQuery does analytics on both historical data and streaming data. BigQuery can query data as it arrives from streaming pipelines. BigQuery is SQL and the latency is on the order of seconds. BigQuery is good for ad hoc. Bigtable is big, fast, and autoscaling NoSQL. Bigtable uses clusters but those clusters only contain pointers to the data but do not contain the actual data. Data is in Cloud Storage. Nodes read continuous rows of data. Bigtable supports the HBase API. The latency in Bigtable is on the order of milliseconds. BigQuery and Bigtable are about user generated queries, ad hoc queries, queries that you do only once in a long while. Apache Beam is a programming model for both batch and streaming. It supports multiple runtimes. Beam supports time-based shuffle to put data in correct window. Windowing is about event time and not processing time. Beam lets you choose between high and low latency. Beam handles structured data, semi-structured, object data, and let’s you run queries. Beam offers a single pipeline--a unified model for processing batch and stream data.
corvidfox
This is my personal portfolio site built with Quarto and hosted via GitHub Pages. It showcases my work across data analytics, dashboard development, and applied SQL storytelling—built to support job searches, recruiter walkthroughs, and professional demos.
vijayalakshmipkagi-dotcom
HR Analytics project analyzing IBM employee attrition using data visualization, SQL, and ML. Identifies key factors like overtime, salary, job satisfaction, and work-life balance. Provides insights to reduce attrition and improve retention through better engagement, pay, and career growth strategies.
Sainiaal
10-Day Data Dash is a hands-on challenge designed to boost data analytics skills through real-world projects using Excel, SQL, and Power BI. Perfect for freshers aiming to build a strong portfolio and stand out in a competitive job market by demonstrating practical experience in data cleaning, analysis, and visualization.
Youssef-68
This project transforms raw multi-source job market datasets into a fully normalized relational SQL data warehouse designed for analytical reporting and dashboarding. The system processes over 6 million job records and structures them into optimized relational tables to enable efficient querying and business intelligence reporting.
Strano1
This is a Data Cleaning project carried out on the Data Science Job Postings dataset. This project makes use of SQL queries to remove unnecessary values and unused columns from the messy dataset. It aims to make the dataset easier to use for analytical purposes.
Data analytics isn’t just about the future, it is being put to use at this very moment in all businesses. It forms an integral part of the company and the professionals are paid highly for their part. Here are reasons why joining data analytics training in Gurgaon is a viable option After the completion of Data Analytics Course, you will be able to: Understand Scala & Apache Spark implementation Spark operations on Spark Shell Spark Driver & its related Worker Nodes Spark + Flume Integration Setting up Data Pipeline using Apache Flume, Apache Kafka & Spark Streaming Spark RDDs and Spark Streaming Spark MLib : Creating Classifiers & Recommendations systems using MLib Spark Core concepts: Creating of RDDs: Parrallel RDDs, MappedRDD, HadoopRDD, JdbcRDD. Spark Architecture & Components Spark SQL experience with CSV, XML & JSON Reading data from different Spark sources Spark SQL & Dataframes Develop and Implement various Machine Learning Algorithms in daily practices & Live Environment Building Recommendation systems and Classifiers Perform various type of Analysis (Prediction & Regression) Implement plotting & graphs using various Machine Learning Libraries Import data from HDFS & Implement various Machine Learning Models Building different Neural networks using NumPy and TensorFlow Power BI Visualization Power BI Components Power BI Transformations Dax functions Data Exploration and Mapping Designing Dashboards Time Series, Aggregation & Filters Placement Gyansetu is providing complimentary placement service to all students. Gyansetu Placement Team consistently works on industry collaboration and associations which help our students to find their dream job right after the completion of training. Why Choose us? Gyansetu trainers are well known in Industry; who are highly qualified and currently working in top MNCs. We provide interaction with faculty before the course starts. Our experts help students in learning Technology from basics, even if you are not good at basic programming skills, don’t worry! We will help you. Faculties will help you in preparing project reports & presentations. Students will be provided Mentoring sessions by Experts.
DESCRIÇÃO DA VAGA Buscamos engenheiros de dados que se motivem com tecnologia de ponta e um ambiente com com muita autonomia para testar coisas novas. Somos um time que está sempre se reinventando para arquitetar soluções para para processar, armazenar e prover dados cada vez mais relevantes para todos os nossos produtos e também para nossos clientes. Você participará de um time que estará arquitetando sistemas distribuídos, criando pipelines escaláveis e confiáveis, combinando múltiplas fontes de dados e pensando em arquiteturas de dados escaláveis e otimizando recursos pensando na eficiência da nossa infra. Nossa tecnologia suporta as maiores marcas e varejos do mercado a tomarem decisões estratégicas sobre suas vendas no canal digital - e-commerce - e ajudam a aproximá-los das dos shoppers em campanhas em redes sociais. Responsabilidades: Conhecer e interagir com as diferentes áreas da Lett com o objetivo de ter um conhecimento amplo do negócio e das bases de dados; Desenvolver e implantar arquiteturas e processos que suportem as soluções das demais equipes de forma escalável; Governar, documentar e prover acesso a metadados a todas as equipes; Modelar Data Lakes e Data Warehouse; Pesquisar e trazer abordagens e tecnologias modernas para as soluções de Big Data da empresa; Criar e gerenciar data flows, clusters de processamento e armazenamento de dados em nossa cloud; Propor melhorias, otimizações de baixo nível e novas arquiteturas para os outros times; Democratizar o acesso a dados utilizando ferramentas e desenvolvimento de interfaces (como APIs, ETLs, SQL); e trabalhar diretamente com equipes de produto. REQUISITOS DA VAGA Requisitos: Experiência com Python; Experiência com Docker e docker-compose; Estar muito confortável com ambiente Spark (Pyspark no serviço EMR da AWS ou em Kubernetes); AWS (Elastic Beanstalk, SQS, RDS, Lambda Functions, EC2, EMR, S3, SNS); Vasta experiência com Data Lakes em storage de objetos (AWS S3); Vasta experiência com Google Big Query (modelagem de dados, ELTs, manutenção e governança); Experiência com o conceito de ELT; Experiência com governança e catalogação de dados; Apache Airflow (Implementação de DAGs e deploys do Airflow em formato de workers distribuídos); Data warehouses, data lakes, suas interfaces (engines SQL, processos de ETL, acesso direto a objetos) e sua organização ( particionamento, formas de orientar dados, custo e performance); Estar confortável com SQL, bancos de dados relacionais, bancos orientados a documento e armazenamento de arquivo; Arquiteturas diversas de processamento (filas, jobs, workers, functions... etc); Ferramentas de deploy, versionamento de código e infraestrutura na nuvem; Ser capaz de criar fluxos de execução paralela ou concorrente de processamento, assim como execução distribuída quando aplicável; Saber interagir de forma técnica e não-técnica com outros membros da equipe e com pessoas de outras áreas. Diferenciais: Dremio/Athena/AWS Glue Catalog; Ferramenta DBT (Data Build Tool da Fishtown Analytics); Amundsen; Processos de dados envolvendo CDC (Change Data Capture); Terraform; Prometheus/Grafana; Kubernetes/AWS ECS/AWS EKS; PostgreSQL, MongoDB, ElasticSearch, DynamoDB; Java/Scala; Jenkins; Inglês. BENEFÍCIOS DA VAGA Plano de Saúde; Plano Odontológico; VR; VA; Home Office.