Remote data engineer jobs at top U.S. companies

We, at Turing, are looking for talented data engineers who can build data processing and reporting infrastructure that powers company-wide insights and technical intelligence. Join forces with the top 1% of data engineers and grow with the best minds.

Find remote software jobs with hundreds of Turing clients

Job description

Job responsibilities

  • Design database and data pipeline/ETL using emerging technologies and tools
  • Drive the team to develop operationally efficient analytic solutions
  • Define standards and methodologies for the data warehousing environment
  • Design and build highly scalable data pipelines using new-generation tools and technologies like AWS, Snowflake, Spark, and Kafka to induct data from various systems
  • Translate complex business requirements into scalable technical solutions that meet data warehousing design standards
  • Create scalable data pipelines and ETL applications that support business operations in advertising, content, and finance/accounting
  • Assist with deciphering data migration issues and improving system performance
  • Collaborate efficiently with product management, technical program management, operations, and other engineers

Minimum requirements

  • BS, MS, or Ph.D. in Computer Science or a relevant technical field
  • Extensive experience building scalable data systems and data-driven products, as well as working with cross-functional teams
  • 2+ years of related software/data engineering experience with proficiency in Python
  • Ability to create data pipelines and ETL applications with large datasets
  • Proficiency in building REST APIs for back-end services
  • Exposure to implementing, testing, debugging and deploying data pipelines using any of the following tools: Prefect, Airflow, Glue, Kafka, Serverless (Lambda, Kinesis, SQS, SNS), Fivetran, or Stitch Data/Singer
  • Experience with any of the cloud data warehousing technologies: Redshift, BigQuery, Spark, Snowflake, Presto, Athena, or S3
  • Experience with SQL DB administration (PostgreSQL, MS SQL, etc.)
  • Fluent in English to communicate effectively
  • Ability to work full-time (40 hours/week) with a 4-hour overlap with U.S. time zones

Preferred skills

  • Understanding of complex, distributed, microservice web architectures
  • Experience with Python back-end and ETL to move from one database to another
  • Solid understanding of analytics and a drive to build solutions to improve efficiency

Interested in this job?

Apply to Turing today.

Apply now

Why join Turing?

Elite US Jobs

1Elite US Jobs

Turing’s developers earn better than market pay in most countries, working with top US companies.
Career Growth

2Career Growth

Grow rapidly by working on challenging technical and business problems on the latest technologies.
Developer success support

3Developer success support

While matched, enjoy 24/7 developer success support.

Developers Turing

Read Turing.com reviews from developers across the world and learn what it’s like working with top U.S. companies.
4.65OUT OF 5
based on developer reviews as of June 2024
View all reviews

How to become a Turing developer?

Work with the best software companies in just 4 easy steps
  1. Create your profile

    Fill in your basic details - Name, location, skills, salary, & experience.

  2. Take our tests and interviews

    Solve questions and appear for technical interview.

  3. Receive job offers

    Get matched with the best US and Silicon Valley companies.

  4. Start working on your dream job

    Once you join Turing, you’ll never have to apply for another job.

cover

How to become a data engineer

The core focus of data engineering is to build dependable infrastructures that enable constant data flow in a data-driven environment. Data engineers function as facilitators of clean and raw data from multiple sources so that businesses can use it to make data-driven choices.

Data engineering is the process of developing and constructing large-scale data collection, storage, and analysis systems. Data engineers design systems that gather, handle and turn raw data into usable information for data scientists and business analysts to comprehend. The ultimate objective is to make data more available so that businesses may assess and improve their performance.

What is the scope of a data engineering job?

The demand for data-related jobs is growing by the day as more firms leverage big data to gain meaningful insights. Companies are always on the search for competent data engineers. Since the work calls for significant big data experience, the earning potential of data engineers has also increased.

What are the roles and responsibilities of data engineers?

The main responsibility is to conceptualize and build a dependable infrastructure for translating data into meaningful insights that data scientists can understand. Remote data engineers must be able to identify trends in massive datasets in addition to designing scalable processes to transform semi-structured and unstructured data into usable representations. Let's take a look at some of the other responsibilities of remote data engineer jobs.

  • Develop, construct, test, and maintain data architectures
  • Assemble complicated datasets that align with business requirements
  • Deploy sophisticated analytics programs, machine learning, and statistical methods
  • Ensure data security and governance with modern-day security controls
  • Translate complex functional and technical needs into elaborate designs
  • Implement data storage with technologies like Hadoop, NoSQL, etc.
  • Integrate data management processes into the organization’s current structure
  • Help in seamless third-party integration
  • Create high-performance and scalable web services to track data

How to become a data engineer?

Data engineers usually have a background in Computer Science and Engineering, Applied Mathematics, or a related IT profession. However, coming from a non-tech background, it is still possible to become a Data Engineer.

The Data Engineer job necessitates a high technical understanding of data structuring and storage. If you're still pursuing studies, you can opt for technical degrees like Computer Science, Data Engineering, or Machine Learning. A bachelor's degree in Computer Science or a similar subject is common among data engineers. By acquiring a degree, you can establish the foundation of knowledge in this rapidly changing sector. You can also pursue a master's degree to enhance your career and gain access to possibly higher-paying opportunities.

You'll require programming skills in a variety of languages, including Python and Java, as well as an understanding of SQL database architecture. A boot camp or certification can help design a CV for remote data engineering jobs if you already have a background in IT or a related area like mathematics or analytics.

If you don't have a background in technology or IT, you can choose the self-learning or mentorship programs. Mentorship programs are online courses available on different professional education platforms that provide guided learning courses (one on one). Self-learning is the path many choose because of the vast number of technical resources available on the Internet today, but this is unguided learning, so it may take more time/resources. Keep checking details about master's degrees in data analytics and data engineering if you have an undergraduate degree, but it isn't in a relevant discipline.

Take some time to look through job postings to discover what employers are looking for, and you'll better understand how your experience fits into that role.

Interested in remote data engineer jobs?

Become a Turing developer!

Apply now

Skills required to become a data engineer

1. Hadoop and Spark

The Apache Hadoop software library is a framework that uses basic programming principles to enable the distributed processing of massive data volumes across clusters of machines. It's built to expand from a single server to thousands of devices, each with its computing and storage capabilities.

Python, Scala, Java, and R are among the programming languages supported. While Hadoop is the most powerful tool for large data, it has several limitations, including slow processing and a high level of coding.

Apache Spark is a data processing engine that allows stream processing, which involves continuous data input and output. It is similar to Hadoop in that it performs many of the same activities.

2. C++

C++ is a very basic yet powerful programming language for swiftly calculating massive datasets. It's the only language that can handle over 1GB of data in a single second. The data can be retrained and predictive analytics can be used in real-time while keeping the system of record consistent.

3. Data warehousing

A data warehouse is a relational database to query and analyze data. It's intended to provide a long-term picture of data across time. A database, on the other hand, continuously updates real-time data. Knowledge of systems like Amazon Web Services and Amazon Redshift is needed for data engineering. In fact, AWS is a prerequisite in various on-site and remote data engineer jobs.

4. Azure

Azure is a cloud platform from Microsoft that allows data engineers to create large-scale data analytics solutions. With an easy-to-deploy bundled analytics solution, it simplifies the deployment and support of servers and applications.

The package includes pre-built services for everything, from data storage to powerful machine learning. Azure is so popular that some data engineers even specialize in it.

5. SQL and NoSQL

Database management systems (DBMS) is a software application that offers an interface to databases for information storage and retrieval - required knowledge for data engineers.

The SQL programming language is the industry standard for creating and maintaining relational database systems. On the other hand, non-tabular NoSQL databases come in several shapes and sizes depending on their data models, such as a graph or a text.

6. Machine learning

Data scientists use machine learning algorithms to create various predictive models based on current and past data. Data engineers, however, only require rudimentary knowledge of machine learning to better understand the needs of data scientists (and by extension, the needs of the company), and construct more accurate data pipelines.

7. Data APIs

An API is a data access interface for software applications. It enables two apps or devices to interact with one another to complete a certain job. Web applications, for example, employ API to interact between the user-facing front-end and the back-end functionality and data.

An API allows an application to read a database, get information from relevant tables in the database, process the request, and deliver an HTTP-based response to the web template, which is then shown in the web browser. Data engineers provide APIs in databases for data scientists and business intelligence analysts to query the data.

8. Extract, Transfer, Load (ETL)

ETL (Extract, Transfer, Load) is the process of extracting data from a source, converting it into a format that can be analyzed, and storing it in a data warehouse.

The ETL collects data from a variety of sources, applies business rules to the data, and then loads the transformed data into a database or business intelligence platform where it can be accessed and utilized by everyone in the company.

Interested in remote Data Engineer jobs?

Become a Turing developer!

Apply now

How to get remote data engineer jobs

Becoming a data engineer is highly rewarding. However, you need to have a thorough understanding of programming. Practicing programming is important, along with having a vision of the product. Good communication skills are also helpful in collaborating with team members and prioritizing work.

To simplify your search for data engineer remote jobs, Turing has made things a little easier. We offer the best opportunities to suit your career trajectory. Join a network of the world's top developers and get full-time, long-term remote data engineer jobs with better compensation and career growth prospects.

Why become a Data engineer at Turing?

Elite U.S. jobs

Long-term opportunities to work for amazing, mission-driven U.S. companies with great compensation.

Career growth

Work on challenging technical and business problems using cutting-edge technology to accelerate your career growth.

Exclusive developer community

Join a worldwide community of elite software developers.

Once you join Turing, you’ll never have to apply for another job.

Turing's commitments are long-term and full-time. As one project draws to a close, our team gets to work identifying the next one for you in a matter of weeks.

Work from the comfort of your home

Turing allows you to work according to your convenience. We have flexible working hours and you can work for top U.S. firms from the comfort of your home.

Great compensation

Working with top U.S. corporations, Turing developers make more than the standard market pay in most nations.

How much does Turing pay their Data engineer?

Turing helps you suggest a salary range that allows you to settle for a fruitful and long-term opportunity. Most of our recommendations are an assessment of market conditions and the demand set by our clients. However, at Turing, we believe in flexibility. Thus, every Data engineer is eligible to fix their salary range as per their skills and expertise.

Frequently Asked Questions

Turing is an AGI infrastructure company specializing in post-training large language models (LLMs) to enhance advanced reasoning, problem-solving, and cognitive tasks. Founded in 2018, Turing leverages the expertise of its globally distributed technical, business, and research experts to help Fortune 500 companies deploy customized AI solutions that transform operations and accelerate growth. As a leader in the AGI ecosystem, Turing partners with top AI labs and enterprises to deliver cutting-edge innovations in generative AI, making it a critical player in shaping the future of artificial intelligence.

After uploading your resume, you will have to go through the three tests -- seniority assessment, tech stack test, and live coding challenge. Once you clear these tests, you are eligible to apply to a wide range of jobs available based on your skills.

No, you don't need to pay any taxes in the U.S. However, you might need to pay taxes according to your country’s tax laws. Also, your bank might charge you a small amount as a transaction fee.

We, at Turing, hire remote developers for over 100 skills like React/Node, Python, Angular, Swift, React Native, Android, Java, Rails, Golang, PHP, Vue, among several others. We also hire engineers based on tech roles and seniority.

Communication is crucial for success while working with American clients. We prefer candidates with a B1 level of English i.e. those who have the necessary fluency to communicate without effort with our clients and native speakers.

Currently, we have openings only for the developers because of the volume of job demands from our clients. But in the future, we might expand to other roles too. Do check out our careers page periodically to see if we could offer a position that suits your skills and experience.

Our unique differentiation lies in the combination of our core business model and values. To advance AGI, Turing offers temporary contract opportunities. Most AI Consultant contracts last up to 3 months, with the possibility of monthly extensions—subject to your interest, availability, and client demand—up to a maximum of 10 continuous months. For our Turing Intelligence business, we provide full-time, long-term project engagements.

No, the service is absolutely free for software developers who sign up.

Ideally, a remote developer needs to have at least 3 years of relevant experience to get hired by Turing, but at the same time, we don't say no to exceptional developers. Take our test to find out if we could offer something exciting for you.

View more FAQs

Latest posts from Turing

Turing Blog: Articles, Insights, Company News and Updates

Explore insights on AI and AGI at Turing's blog. Get expert insights on leveraging AI-powered solutions to drive ...

Read more
The Fifteen Best Front-End Frameworks

The Fifteen Best Front-End Frameworks

In this post, we’ve listed front-end frameworks that remote software developers can use that easily link with...

Read more

Django vs. Flask: Which Framework to Choose?

Both are Python-based, free, open-source frameworks. But who takes the lead in this Django vs Flask battle?

Read more
Here Are the Five Fastest Growing IT Jobs

Here Are the Five Fastest Growing IT Jobs

If you’re contemplating changing careers in the same domain or embarking on a new career path entirely, you’ve co...

Read more
10 PHP Frameworks Developers Should Look Out for

10 PHP Frameworks Developers Should Look Out for

This post shares a list of the ten best PHP frameworks and explains how software developers can leverage each...

Read more

Turing Blog: Articles, Insights, Company News and Updates

Explore insights on AI and AGI at Turing's blog. Get expert insights on leveraging AI-powered solutions to drive ...

Read more
What Are the Best Programming Languages for AI Development?

What Are the Best Programming Languages for AI Development?

Enterprises worldwide have reported plans to expand their AI strategies. This post lists the ten best...

Read more
Six Common Interview Mistakes Remote Developers Make

Six Common Interview Mistakes Remote Developers Make

What shouldn’t you do when interviewing? Here are the most common interview mistakes a candidate must...

Read more

Leadership

In a nutshell, Turing aims to make the world flat for opportunity. Turing is the brainchild of serial A.I. entrepreneurs Jonathan and Vijay, whose previous successfully-acquired AI firm was powered by exceptional remote talent. Also part of Turing’s band of innovators are high-profile investors, such as Facebook's first CTO (Adam D'Angelo), executives from Google, Amazon, Twitter, and Foundation Capital.

Equal Opportunity Policy

Turing is an equal opportunity employer. Turing prohibits discrimination and harassment of any type and affords equal employment opportunities to employees and applicants without regard to race, color, religion, sex, sexual orientation, gender identity or expression, age, disability status, protected veteran status, or any other characteristic protected by law.

Explore remote developer jobs

briefcase
Python Automation and Task Creator

About Turing:

Based in San Francisco, California, Turing is the world’s leading research accelerator for frontier AI labs and a trusted partner for global enterprises deploying advanced AI systems. Turing supports customers in two ways: first, by accelerating frontier research with high-quality data, advanced training pipelines, plus top AI researchers who specialize in coding, reasoning, STEM, multilinguality, multimodality, and agents; and second, by applying that expertise to help enterprises transform AI from proof of concept into proprietary intelligence with systems that perform reliably, deliver measurable impact, and drive lasting results on the P&L.


Role Overview

We are seeking a detail-oriented Computer-Using Agent (CUA) to perform structured automation tasks within Ubuntu-based virtual desktop environments. In this role, you will interact with real desktop applications using Python-based GUI automation tools, execute workflows with high accuracy, and document every step taken.

This is a hands-on execution role ideal for candidates who are comfortable working with Linux systems, virtualization tools, and repeatable task workflows in a controlled environment.


What Does the Day-to-Day Look Like?

  • Set up and operate Ubuntu virtual machines using VMware or VirtualBox
  • Automate mouse and keyboard interactions using Python-based GUI automation (e.g., PyAutoGUI)
  • Execute predefined workflows across various Ubuntu desktop applications
  • Ensure tasks are completed accurately and can be reproduced consistently
  • Capture and document all actions, steps, and outcomes in a structured format
  • Collaborate with the delivery team to refine automation scenarios and workflows

Required Skills & Qualifications

  • Hands-on experience with Ubuntu/Linux desktop environments
  • Working knowledge of PyAutoGUI or similar GUI automation frameworks
  • Basic Python scripting and debugging skills
  • Familiarity with VMware or VirtualBox
  • Strong attention to detail and ability to follow step-by-step instructions
  • Clear documentation and reporting skills

Application Domains

You will be expected to perform automation tasks across the following Ubuntu-based environments:

  • os – Core Ubuntu desktop environment
  • chrome – Ubuntu with Google Chrome
  • gimp – Ubuntu with GIMP
  • libreoffice_calc – LibreOffice Calc
  • libreoffice_writer – LibreOffice Writer
  • libreoffice_impress – LibreOffice Impress
  • thunderbird – Thunderbird email client
  • vlc – VLC media player
  • vs_code – Visual Studio Code

Perks of Freelancing With Turing

  • Fully remote work.
  • Opportunity to work on cutting-edge AI projects with leading LLM companies.

Offer Details:

  • Commitments Required: 40 hours per week with 4 hours of overlap with PST. 
  • Engagement  type  : Contractor assignment (no medical/paid leave)
  • Duration of contract : 2 month
Holding Companies & Conglomerates
10K+ employees
Python
briefcase
Knowledge Graph Expert (Knowledge Graph / SQL / LLM)
About the Client

Our mission is to bring community and belonging to everyone in the world. We are a community of communities where people can dive into anything through experiences built around their interests, hobbies, and passions. With more than 50 million people visiting 100,000+ communities daily, it is home to the most open and authentic conversations on the internet.

About the Team

The Ads Content Understanding team’s mission is to build the foundational engine for interpretable and frictionless understanding of all organic and paid content on our platform. Leverage state-of-the-art applied ML and a robust Knowledge Graph (KG) to extract high-quality, monetization-focused signals from raw content — powering better ads, marketplace performance, and actionable business insights at scale.

We are seeking a Knowledge Graph Expert to help us grow and curate our KG of entities and relationships, bringing it to the next level.


About the Role


We are looking for a detail-oriented and strategic Knowledge Graph Curator. In this role, you will sit at the intersection of AI automation and human judgment. You will not only manage incoming requests from partner teams but also proactively shape the growth of our Knowledge Graph (KG) to ensure high fidelity, relevance, and connectivity. You will serve as the expert human-in-the-loop, validating LLM-generated entities and ensuring our graph represents the "ground truth" for the business.

 

Key Responsibilities


  • Onboarding of new entities to the Knowledge Graph maintained by the Ads team
  •  Data entry, data labeling for automation of content understanding capabilities
  • LLM Prompt tuning for content understanding automation

What You'll Do


1. Pipeline Management & Prioritization

  • Manage Inbound Requests: Act as the primary point of contact for partner teams (Product, Engineering, Analytics) requesting new entities or schema changes.
  • Strategic Prioritization: Triage the backlog of requests by assessing business impact, urgency, and technical feasibility.

2. AI-Assisted Curation & Human-in-the-Loop

  • Oversee Automation: Interact with internal tooling to review entities generated by Large Language Models (LLMs). You will approve high-confidence data, edit near-misses, and reject hallucinations.
  • Quality Validation: Perform rigorous QA on batches of generated entities to ensure they adhere to the strict ontological standards and factual accuracy required by the KG.
  • Model Feedback Loops: Participate in ad-hoc labeling exercises (creation of Golden Sets) to measure current model quality and provide training data to fine-tune classifiers and extraction algorithms.

3. Data Integrity & Stakeholder Management

  • Manual Curation & Debugging: Investigate bug reports from downstream users or automated anomaly detection systems. You will manually fix data errors, merge duplicate entities, and resolve conflicting relationships.
  • Feedback & Reporting: Close the loop with partner teams. You will report on the status of their requests, explain why certain modeling decisions were made, and educate stakeholders on how to best query the new data.


Qualifications for this role:

  • Knowledge Graph Fundamentals: Understanding of graph concepts (Nodes, Edges, Properties)
  • Taxonomy & Ontology: Experience categorizing data, managing hierarchies, and understanding semantic relationships between entities.
  • Data Literacy: Proficiency in navigating complex datasets. Experience with SQL, SPARQL, or Cypher is a strong plus.
  • AI/LLM Familiarity: Understanding of how Generative AI works, common failure modes (hallucinations), and the importance of ground-truth data in training.

Operational & Soft Skills

  • Analytical Prioritization: Ability to look at a list of 50 tasks and determine the 5 that will drive the most business value.
  • Attention to Detail: An "eagle eye" for spotting inconsistencies, typos, and logical fallacies in data.
  • Stakeholder Communication: Ability to translate complex data modeling concepts into clear language for non-technical product managers and business stakeholders.
  • Tool Proficiency: Comfort learning proprietary internal tools, ticketing systems (e.g., Jira), and spreadsheet manipulation (Excel/Google Sheets).


Offer Details


  • Full-time contractor or full-time employment, depending on the country
  • Remote only, full-time dedication (40 hours/week)
  • 8 hours of overlap with Netherlands
  • Competitive compensation package.
  • Opportunities for professional growth and career development.
  • Dynamic and inclusive work environment focused on innovation and teamwork
Media & Internet
251-10K employees
LLMSQL
sample card

Apply for the best jobs

View more openings
Turing books $87M at a $1.1B valuation to help source, hire and manage engineers remotely
Turing named one of America's Best Startup Employers for 2022 by Forbes
Ranked no. 1 in The Information’s "50 Most Promising Startups of 2021" in the B2B category
Turing named to Fast Company's World's Most Innovative Companies 2021 for placing remote devs at top firms via AI-powered vetting
Turing helps entrepreneurs tap into the global talent pool to hire elite, pre-vetted remote engineers at the push of a button

Work with the world's top companies

Create your profile, pass Turing Tests and get job offers as early as 2 weeks.