Data Engineer
- Posted 22 March 2023
- LocationSingapore
- Job type Permanent
- Reference23WD68046
Company's Benefits
-
Flexible Working Arrangements
-
Equal Pay Initiatives
-
Mentorship Program
-
Leadership Development Program
-
Paid Parental Leave
-
Return to Work Policy
-
Breastfeeding Rooms
-
Sponsorship Program
-
Coaching Program
-
Raise Numbers Of Women In Leadership
-
Internal Women's Networking Group
Job Description
Position Overview
Autodesk is looking for a Data Engineer to join the Data Ingestion team within the Analytics Data organization. The Data Ingestion team processes around 100 billion events and 250TB of data per day from all Autodesk products. The team is responsible for building and maintaining streaming real-time enterprise level data pipelines and developing simple, repeatable ingestion patterns for Autodesk’s internal data platform.
As a Data Engineer, you will be contributing to rapidly improve critical data processing & analytics pipelines. You will be involved in delivering innovative solutions to sophisticated and modern engineering problems. As part of the team, you will learn, grow, and help bring data closer to our users. You will make critical choices, tackle hard problems and improve the platform’s reliability, resiliency, and scalability.
We are looking for someone who is enthusiastic about working in a team, can own and deliver long-term projects to completion. You are detail and quality oriented, and excited about the prospects of having a big impact with data at Autodesk.
Responsibilities
Contribute to the team’s deliverables and innovate on scalability and latency challenges
You will need a product-focused mindset. It is essential for you to understand business requirements and deliver systems that will scale and extend to accommodate those needs
Diagnose and solve complex problems in distributed systems, develop and document technical solutions and sequence work to make fast, iterative deliveries and improvements
Build and maintain high-performance, fault-tolerant and scalable distributed systems that can handle our massive scale
Ideate and deliver innovative projects that will improve user experience
Efficiently deliver thorough data-driven approach, robust systems designs, and effective software implementation
Participate in, or spearhead design reviews with peers and stakeholders to adopt what’s best suited amongst available technologies
Review code developed by other developers and provide feedback to ensure best practices (e.g., style guidelines, checking code in, accuracy, testability, and efficiency)
Automate cloud infrastructure, services, and observability
Develop CI/CD pipelines and testing automation
Establish and uphold best engineering practices through thorough code and design reviews and improved processes and tools
Drive a culture of trust, respect and inclusion within your team
Minimum Qualifications
3+ years of relevant industry experience in streaming data systems
Solid experience with streaming technologies (i.e. Kafka, Spark Streaming, Flink) and streaming tables (i.e. Hudi, Iceberg)
Solid Proficiency with Amazon Web Services
Proficient overall programming skills, able to write modular, maintainable code, preferably Python & SQL
Understanding of SQL, dimensional modeling, and at least one relational database
Experience with automation frameworks/tools like Git, Jenkins, Ansible, and Cloudformation (or Terraform)
Familiarity with containers and infrastructure-as-code fundamentals
Problem solver with excellent written and interpersonal skills; ability to make sound, complex recommendations in a fast-paced, technical environment
Humble, collaborative, team player, willing to step up and support your colleagues
Effective communication, problem solving and interpersonal skills
Commit to grow deeper in the knowledge and understanding of how to improve our existing applications
Enthusiasm for cutting edge technologies, complex problems, and building things
Familiar with non-functional testing such as load, performance and resiliency testing
Good command of English (Speaking, Writing and Reading)
Working in an agile environment using test driven methodologies
Bachelor’s degree in Computer Science, Engineering or related field, or equivalent training, fellowship or work experience
Preferred Qualifications
Experience with Spark
Experience with Hive and/or Snowflake
Strong knowledge and experience in Hadoop 2.0 and its ecosystem.
Experience with Airflow
Experience with Looker