Python Data Scraping Engineer - Freelance
Mindrift is looking for highly skilled Python Data Scraping Engineers to join the Tendem project and drive specialized data scraping workflows within our hybrid AI + human system. As an AI Pilot - our internal title for this role - you will collaborate with Tendem Agents, providing critical thinking, domain expertise, and quality control to deliver accurate, actionable results. This part time remote opportunity is ideal for technical professionals with experience in web scraping, data extraction, and processing. From coding and automation to fine tuning AI outputs, you'll play a key role in advancing AI capabilities and real world applications.
Responsibilities
- Own end to end data extraction workflows across complex websites, ensuring complete coverage, accuracy, and reliable delivery of structured datasets.
- Leverage internal tools (Apify, OpenRouter) alongside custom workflows to accelerate data collection, validation, and task execution while meeting defined requirements.
- Ensure reliable extraction from dynamic and interactive web sources, adapting approaches as needed to handle JavaScript rendered content and changing site behavior.
- Enforce data quality standards through validation checks, cross source consistency controls, adherence to formatting specifications, and systematic verification prior to delivery.
- Scale scraping operations for large datasets using efficient batching or parallelization, monitor failures, and maintain stability against minor site structure changes.
Compensation
Contributors can earn up to $32 per hour equivalent, depending on level and pace of contribution. Compensation varies by project scope, complexity, and required expertise.
Qualifications
- At least 3 years of relevant experience in data engineering, web scraping, automation, or software development.
- Bachelor's or Master's Degree in Engineering, Applied Mathematics, Computer Science, or related technical fields is a plus.
- Strong experience in Python web scraping (BeautifulSoup, Selenium or similar), including dynamic content (JS, AJAX, infinite scroll) and APIs via proxies.
- Proven ability to extract data from complex structures (hierarchies, archived pages, inconsistent HTML).
- Solid background in data cleaning, normalization, and validation, delivering structured datasets (CSV, JSON, Google Sheets).
- Hands on experience with LLMs and AI frameworks to enhance automation and problem solving.
- Strong attention to detail and commitment to data accuracy.
- Self directed work ethic with ability to troubleshoot independently.
- English proficiency: Upper intermediate (B2) or above.
Benefits
- Fully remote work with flexible schedule; just a laptop and stable internet connection.
- Opportunity to gain hands on experience in a hybrid environment where human expertise and AI agents collaborate.
- Performance based bonus programs rewarding high quality work and consistent delivery.