facebook Created with Sketch. instagram Created with Sketch. linkedin Created with Sketch. twitter Created with Sketch.

Technology

Data Scraping Engineer (Intern)

Location: Bengaluru | Full Time

Who We Are:

Stylumia believes in augmenting human intelligence in fashion and lifestyle retail for a better world. With relevant data and the right technology, people and businesses in fashion and lifestyle retail can solve challenging problems and change the world for the better. In 2015-16, when we looked at the available technology and solutions, we saw fashion forecasting and demand planning are either subjective in nature. We saw these methods do not solve the fundamental challenge in fashion trend spotting, trend forecast, and the business metrics of full-price sell-through continue to remain around the 50% mark creating over USD 750 billion/annum of wastage globally. The biggest wastage in the fashion industry is the one caused by ill-informed decision-making. We saw a need for a different kind of technology and approach, and we knew it would need a different kind of company to build it. That is why we formed Stylumia, a Deep Tech company solving fundamental retail challenges. Stylumia is a global innovation-first Retail AI organisation. We have products that impact the daily lives of varying multi-disciplinary personas across the value chain as we strive to help them eliminate wastage in the industry, by augmenting their decisions with consumer-backed insights.  

Skills & Responsibilities:

  • Strong programming skills with proficiency in Python and Node.js.
  • Solid understanding of web scraping techniques and experience with relevant libraries and
    frameworks (e.g., BeautifulSoup, Scrapy, Puppeteer, etc.).
  • Proficiency in SQL and experience working with relational databases for data storage and retrieval.
  • Familiarity with version control systems, particularly Git.
  • Knowledge of key-value stores, such as Redis, for caching and data storage.
  • Experience with Elasticsearch or other search engines for data indexing and searching.
  • Strong problem-solving skills and the ability to handle complex data extraction scenarios.
  • Excellent attention to detail and the ability to ensure data accuracy and quality.
  • Strong communication and collaboration skills to work effectively within a team environment.
  • Familiarity with data privacy and security best practices.
  • Bachelor’s degree in Computer Science, Data Science, or a related field (or equivalent experience).

Responsibility:

  • Develop and maintain web scraping scripts to extract data from a variety of sources.
  • Collaborate with the data team to understand data requirements and implement effective scraping
    strategies.
  • Conduct data quality assessments to ensure the accuracy and reliability of scraped data.
  • Optimize scraping processes for efficiency and performance.
  • Troubleshoot and resolve issues related to data extraction and scraping.
  • Implement data storage and management solutions using SQL databases, Git, Redis, and
    Elasticsearch.
  • Collaborate with cross-functional teams to integrate scraped data into downstream applications
    and systems.
  • Stay up to date with the latest web scraping techniques, tools, and technologies.

Preferred Qualifications:

  • Experience with distributed data scraping and parallel processing techniques.
  • Knowledge of data cleaning and preprocessing techniques.
  • Understanding of cloud platforms and their services (e.g., AWS, GCP, Azure).
  • Experience with containerization technologies like Docker.
  • Familiarity with data visualization tools and libraries.
  •  

Enter multiple links in comma seperated value

Upload a file or drag & drop here