Junior Data Analyst
Job in
City of Westminster, Central London, Greater London, England, UK
Listed on 2026-01-13
Listing for:
Information Tech Consultants
Full Time
position Listed on 2026-01-13
Job specializations:
-
IT/Tech
Data Engineer, Data Analyst, Data Science Manager, Data Warehousing
Job Description & How to Apply Below
Job Overview
We’re looking for an enthusiastic and detail-oriented Junior Big Data Developer to join our data engineering team. This role is ideal for an early-career professional with foundational knowledge in data processing, strong proficiency in Python, and expert skills in SQL. You’ll focus on building, testing, and maintaining data pipelines and ensuring data quality across our scalable Big Data platforms.
Responsibilities- Data Pipeline Development:
Assist in the design, construction, and maintenance of robust ETL/ELT pipelines to integrate data from various sources into our data warehouse or data lake. - Data Transformation with Python:
Write, optimize, and maintain production-grade Python scripts to clean, transform, aggregate, and process large volumes of data. - Database Interaction (SQL):
Develop complex, high-performance SQL queries (DDL/DML) for data extraction, manipulation, and validation within relational and data warehousing environments. - Quality Assurance:
Implement data quality checks and monitoring across pipelines, identifying discrepancies and ensuring the accuracy and reliability of data. - Collaboration:
Work closely with Data Scientists, Data Analysts, and other Engineers to understand data requirements and translate business needs into technical data solutions. - Tooling & Automation:
Utilize version control tools like Git and contribute to the automation of data workflows and recurring processes. - Documentation:
Create and maintain technical documentation for data mappings, processes, and pipelines.
- Strong proficiency in Python for data manipulation and scripting.
- Familiarity with standard Python data libraries (e.g., Pandas, Num Py).
- Expert-level proficiency in SQL (Structured Query Language).
- Experience writing complex joins, stored procedures, and performing performance tuning.
- Foundational understanding of Big Data architecture (Data Lakes, Data Warehouses) and distributed processing concepts (e.g., Map Reduce).
- Basic knowledge of ETL principles and data modeling (star schema, snowflake schema).
- Practical experience with Git (branching, merging, pull requests).
- Experience with a distributed computing framework like Apache Spark (using PySpark).
- Familiarity with cloud data services (AWS S3/Redshift, Azure Data Lake/Synapse, or Google Big Query/Cloud Storage).
- Exposure to workflow orchestration tools (Apache Airflow, Prefect, or Dagster).
- Bachelor's degree in Computer Science, Engineering, Information Technology, or a related field.
We are an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national origin, disability, or protected veteran status.
#J-18808-LjbffrNote that applications are not being accepted from your jurisdiction for this job currently via this jobsite. Candidate preferences are the decision of the Employer or Recruiting Agent, and are controlled by them alone.
To Search, View & Apply for jobs on this site that accept applications from your location or country, tap here to make a Search:
To Search, View & Apply for jobs on this site that accept applications from your location or country, tap here to make a Search:
Search for further Jobs Here:
×