Experience: 3+ Years
Location: Delhi/Pune
Required skills:
- Proficiency in Python programming language with strong knowledge of its ecosystem and libraries
- Deep understanding of web technologies, including HTML, CSS, JavaScript, and DOM manipulation
- Experience with web scraping frameworks and libraries such as Scrapy or Selenium
- Experience with dealing with proxy issues:IP Rotation, residential Proxies, and Middleware
- Familiarity with HTTP protocols, web servers, and client-server communication.
- Strong problem-solving skills and ability to analyze and understand complex web structures
- Knowledge of database systems and SQL for storing and querying extracted data
- Excellent communication and teamwork skills to collaborate effectively with cross-functional teams
- Familiarity with cloud computing platforms and services, such as AWS
- Excellent English communication skills
Scope of work:
- Develop, and maintain web crawlers, spiders, and scraping scripts using Python
- Optimize crawling performance and scalability by implementing efficient data retrieval strategies and handling various edge cases
- Work closely with other team members, such as data scientists and backend developers, QA to integrate the extracted data into downstream systems or applications
- Troubleshoot issues related to web crawling, data extraction, and parsing on the Production environment, by reading logs, creating SQL queries in DB’s, and debugging.