Data Engineer – Big Data Pipeline Development & Cloud Infrastructure (All Levels)
Posted 2026-05-06Join arenaflex: Building the Future of Data Engineering
Are you ready to transform your career and become part of a revolutionary data engineering team? At arenaflex, we're not just moving packages – we're moving data that powers global logistics and shapes the future of connected commerce. As a Data Engineer with arenaflex, you'll be at the heart of our technological evolution, designing and building the data pipelines that keep the world moving.
We're currently seeking talented Data Engineers across all experience levels to join our Dataworks division in Washington, USA. This is a unique opportunity to work with cutting-edge big data technologies, collaborate across multidisciplinary teams, and make a tangible impact on how data drives business decisions at one of the world's most iconic logistics companies.
About arenaflex and Dataworks
At arenaflex, we believe that data is the lifeblood of modern logistics. Our Dataworks team serves as the backbone of our information infrastructure, focused on creating and driving engineering innovation while delivering key business initiatives. We operate at massive scale, processing millions of data points daily to ensure smooth operations across our global network.
As a Data Engineer at arenaflex, you'll work within an ecosystem that values collaboration, innovation, and continuous learning. You'll serve as a "universal translator" – bridging the gap between IT professionals, business stakeholders, programmers, and data scientists to deliver transformative solutions that create real business value.
What You'll Do
As a vital member of our data engineering team, you'll be responsible for designing, building, testing, and maintaining robust data pipelines at unprecedented scale. Your daily work will involve:
- Pipeline Architecture: Design and implement scalable data pipelines that efficiently process data from multiple internal and external sources, ensuring data integrity and optimal performance.
- Data Processing: Develop and optimize batch processing systems that handle massive volumes of collected data, transforming and matching information to stored datasets for analysis.
- Environment Optimization: Maintain and continuously improve our data environment and pipeline infrastructure, investigating performance bottlenecks and resolving complex data-related issues with L3-level support.
- Data Transformation: Implement parsers, validators, transformers, and correlators to reformat, update, and enhance data quality, ensuring consistency across all data assets.
- Technical Leadership: Provide guidance and mentorship to team members in less senior positions, sharing best practices and fostering professional growth.
- Innovation: Recommend solutions for highly complex technical challenges, pushing the boundaries of what's possible with our data infrastructure.
Core Responsibilities
- Develop an in-depth understanding of both business and technical challenges that Dataworks aims to solve, translating business requirements into elegant technical solutions.
- Build tools, platforms, and pipelines that enable teams to analyze data clearly and efficiently, construct models, and drive data-informed decisions.
- Scale solutions from "PC scale" to "cluster scale" challenges, addressing both infrastructure and problem structure considerations.
- Collaborate across functional teams to generate data-driven operational insights that result in high-value optimization solutions.
- Deliver tangible value rapidly while working collaboratively with diverse teams of varying backgrounds and disciplines.
- Establish best practices for future reuse by creating open, reusable templates, patterns, and codebases.
- Partner with senior technologists across the broader enterprise and external partner ecosystems to ensure smooth integration with downstream operational systems.
- Participate actively in code reviews and contribute to large-scale architectural decisions.
What We're Looking For
Required Qualifications
- Bachelor's degree in Computer Science, Engineering, Information Technology, or a related technical field.
- Strong foundation in software development principles, database systems, and distributed computing architectures.
- Experience with programming languages such as Python, Scala, or Java.
- Proficiency in SQL and familiarity with at least one scripting language.
- Understanding of data pipeline architecture patterns and best practices.
- Experience with version control systems and collaborative development workflows.
- Strong problem-solving skills and the ability to troubleshoot complex technical issues.
- Excellent communication skills with the ability to translate technical concepts for non-technical stakeholders.
Preferred Skills and Experience
- Deep understanding of the transportation and logistics industry, with the ability to identify high-opportunity areas and design approaches that generate and capture value.
- Hands-on experience with Microsoft Azure cloud platform and services.
- Proven track record building and maintaining CI/CD pipelines using modern DevOps tools and methodologies.
- Experience developing and operationalizing capabilities in near-real-time, high-volume streaming scenarios.
- Strong debugging skills with the ability to work at the code level to resolve difficult issues.
- Demonstrated ability to deliver technical projects under tight timelines while maintaining quality standards.
- An "engineering" mindset – comfortable making rapid, pragmatic decisions to improve performance and accelerate progress.
- Experience working with distributed teams on code-based deliverables, utilizing Git workflows and participating in code review processes.
- DevOps expertise including coding, scripting, database management, YAML, JSON, and Infrastructure as Code (Terraform or similar).
Technical Tool Proficiency
Strong familiarity with the following technologies is highly desirable:
- Big Data Processing: Apache Spark (Scala and PySpark), Apache HDFS, Apache Kafka, and other high-volume data processing frameworks.
- Database Technologies: SQL and NoSQL storage systems including MySQL, PostgreSQL, Cassandra, MongoDB, and ElasticSearch.
- Python Data Ecosystem: Pandas, Scikit-Learn, Matplotlib, TensorFlow, Jupyter Notebooks, and related data science tools.
- Cloud Platforms: Microsoft Azure (preferred), AWS, or Google Cloud Platform.
- Containerization: Docker, Kubernetes, and related orchestration technologies.
Why Join arenaflex?
At arenaflex, we understand that our greatest asset is our people. That's why we offer a comprehensive compensation package designed to attract, retain, and grow top talent:
- Competitive Compensation: Earn $28 per hour with opportunities for overtime and performance-based bonuses.
- Comprehensive Benefits: Full-time employees enjoy health, dental, and vision insurance, along with life insurance and disability coverage.
- Retirement Savings: 401(k) plan with company matching to help you secure your financial future.
- Paid Time Off: Generous vacation, sick leave, and holiday pay to support work-life balance.
- Professional Development: Access to continuous learning opportunities, certifications, and training programs to enhance your skills.
- Career Growth: Clear advancement pathways within our data engineering organization and across the broader arenaflex enterprise.
- Innovation Culture: Work on challenging problems with the latest technologies in a culture that celebrates innovation and creative thinking.
- Collaborative Environment: Join teams that value diverse perspectives and foster inclusive collaboration.
Work Environment
As a Data Engineer at arenaflex, you'll enjoy a modern, flexible work environment that supports both collaborative in-office work and remote flexibility. Our Washington-based team operates with an 8-hour standard workday, though we understand that great engineering sometimes requires flexible scheduling to meet project demands.
You'll have access to state-of-the-art development tools, cloud environments, and the freedom to explore new technologies that can improve our data infrastructure. Our offices feature collaborative spaces designed for team interaction, as well as quiet areas focused work when you need to dive deep into complex problems.
Ready to Make an Impact?
If you're passionate about data engineering, excited about working at scale, and want to be part of a team that's transforming how a global enterprise leverages data, arenaflex is the place for you. We hire problem-solvers, innovators, and collaborative team players who are ready to take their careers to the next level.
Don't miss this opportunity to join one of the world's most respected companies and help shape the future of data engineering in the logistics industry. Apply today and become part of the arenaflex family!
arenaflex is an equal opportunity employer. We value diversity and are committed to creating an inclusive environment for all employees.