- Experienced professional with expertise in web scraping and Python development, with a focus on building efficient spiders using Scrapy. Proficient in regex, CSS, HTML, and JMESPATH, enabling accurate data extraction from diverse web sources. Skilled in understanding and reverse engineering APIs for scraping purposes to enhance data retrieval capabilities.
- Experience of SQL for proficient data handling and storage. Expertise in using web drivers such as Playwright and Selenium to navigate complex, JavaScript-heavy websites and ensure comprehensive data extraction.
- A collaborative mindset encourages teamwork in building the scraping infrastructure on Scrapy, actively contributing to its development and improvement. Commitment to continuous learning and professional growth, staying abreast of the latest web scraping technologies and methodologies.
- Committed to delivering tangible results with a strong work ethic, attention to detail, and exceptional problem-solving skills. Proactively seek professional development opportunities to ensure proficiency in a constantly evolving field.
- Languages: Python, JavaScript(basic), C(basic)
- Frontend Technologies: HTML, CSS, Jinja2, Angular(basic)
- Backend Technologies: Flask, FastAPI, Django(basic)
- Scrapyng Technologies: Scrapy, BeautifulSoup, Parcel, Playwright, Selenium, JMESPATH, XPATH
- Databases: MySQL, PostgreSQL, SQLite
- Tools & Platforms: Git, Docker, Nginx, Google Cloud Platform, Regex
- Played a substantial role in developing web scraping architecture from inception.
- Created more than 2500 spiders for a leading European rental aggregator.
- Engineered validation and extraction methods to ensure data accuracy and reliability
- Support and updates for written spiders
- Scrapy, Playwright, JavaScript, Git, PostgreSQL, SQLAlchemy
- Developed a full-stack website using Angular and Flask, with Websocket communication.
Utilized PostgreSQL for data storage and implemented user authentication using JWT tokens.
- Script developed to automate web scraping spiders using Scrapy.
- Utilized SQLite database for storing and tracking changes.
- Spiders scheduled for execution at specific times and intervals.
- Reliance on settings from other modules for database and Discord integration.
- Ukrainian (Fluent)
- English (Intermediate)
- Russian (Fluent)
- Email: monteship@gmail.com
- LinkedIn: linkedin.com/in/misha-ship
- Google Skill Boost: cloudskillsboost.google/monteship
Available upon request.